Redirecting to original paper in 30 seconds...
Click below to go immediately or wait for automatic redirect
📄 Abstract
Abstract: Contrastive learning (CL) is a prevalent technique for training embedding
models, which pulls semantically similar examples (positives) closer in the
representation space while pushing dissimilar ones (negatives) further apart. A
key source of negatives are 'in-batch' examples, i.e., positives from other
examples in the batch. Effectiveness of such models is hence strongly
influenced by the size and quality of training batches. In this work, we
propose 'Breaking the Batch Barrier' (B3), a novel batch construction strategy
designed to curate high-quality batches for CL. Our approach begins by using a
pretrained teacher embedding model to rank all examples in the dataset, from
which a sparse similarity graph is constructed. A community detection algorithm
is then applied to this graph to identify clusters of examples that serve as
strong negatives for one another. The clusters are then used to construct
batches that are rich in in-batch negatives. Empirical results on the MMEB
multimodal embedding benchmark (36 tasks) demonstrate that our method sets a
new state of the art, outperforming previous best methods by +1.3 and +2.9
points at the 7B and 2B model scales, respectively. Notably, models trained
with B3 surpass existing state-of-the-art results even with a batch size as
small as 64, which is 4-16x smaller than that required by other methods.
Moreover, experiments show that B3 generalizes well across domains and tasks,
maintaining strong performance even when trained with considerably weaker
teachers.
Authors (10)
Raghuveer Thirukovalluru
Rui Meng
Ye Liu
Karthikeyan K
Mingyi Su
Ping Nie
+4 more
Key Contributions
Introduces 'Breaking the Batch Barrier' (B3), a novel batch construction strategy for contrastive learning that leverages a teacher model and community detection to create batches rich in high-quality negatives. This method aims to improve the effectiveness of embedding models by optimizing batch composition.
Business Value
Leads to more robust and accurate embedding models, which are foundational for many AI applications like search, recommendation, and classification, potentially reducing training costs and improving performance.