Papers
Topics
Authors
Recent
Search
2000 character limit reached

MTEB: Massive Text Embedding Benchmark

Published 13 Oct 2022 in cs.CL, cs.IR, and cs.LG | (2210.07316v3)

Abstract: Text embeddings are commonly evaluated on a small set of datasets from a single task not covering their possible applications to other tasks. It is unclear whether state-of-the-art embeddings on semantic textual similarity (STS) can be equally well applied to other tasks like clustering or reranking. This makes progress in the field difficult to track, as various models are constantly being proposed without proper evaluation. To solve this problem, we introduce the Massive Text Embedding Benchmark (MTEB). MTEB spans 8 embedding tasks covering a total of 58 datasets and 112 languages. Through the benchmarking of 33 models on MTEB, we establish the most comprehensive benchmark of text embeddings to date. We find that no particular text embedding method dominates across all tasks. This suggests that the field has yet to converge on a universal text embedding method and scale it up sufficiently to provide state-of-the-art results on all embedding tasks. MTEB comes with open-source code and a public leaderboard at https://github.com/embeddings-benchmark/mteb.

Citations (258)

Summary

  • The paper introduces MTEB, a benchmark that evaluates text embeddings across 8 varied tasks using 58 datasets and 112 languages.
  • It benchmarks 33 models, revealing no single model excels in every task and highlighting differences between supervised and self-supervised methods.
  • The analysis underscores the trade-off between model size and performance while offering practical guidelines for selecting embeddings in NLP applications.

MTEB: Massive Text Embedding Benchmark

The paper "MTEB: Massive Text Embedding Benchmark" addresses the prevalent challenge in text embeddings evaluation. Typically, text embeddings are assessed on a limited set of datasets, often within a single task domain, such as Semantic Textual Similarity (STS). This narrow focus hinders a comprehensive understanding of these models' generalizability across various tasks. To address this issue, the authors introduce the Massive Text Embedding Benchmark (MTEB), an extensive benchmark comprising eight embedding tasks across 58 datasets and 112 languages, establishing the broadest evaluation framework for text embeddings yet.

Major Contributions

Diverse Task Evaluation

MTEB encompasses a diverse set of embedding tasks beyond the commonly evaluated STS and classification, including:

  1. Bitext Mining: Identification of parallel sentence pairs across different languages.
  2. Classification: Training classifiers on embedded text representations to categorize text.
  3. Clustering: Grouping text inputs into meaningful clusters via embedding similarities.
  4. Pair Classification: Determining binary labels (e.g., duplicates or paraphrases) for text pairs.
  5. Reranking: Ranking a list of reference texts against a given query.
  6. Retrieval: Identifying and ranking relevant documents from a corpus given a query.
  7. STS: Evaluating sentence pair similarities.
  8. Summarization: Assessing machine-generated summaries' quality against human-written references.

Comprehensive Dataset Coverage

MTEB's 58 datasets span various languages and text types (sentences, paragraphs, and documents), facilitating a robust assessment of text embeddings' applicability across multiple contexts. The datasets are meticulously curated to include both short and long texts, providing insights into model performance over different text lengths. This extensive dataset pool enables a clear comparison of embeddings across multilingual and monolingual scenarios.

Extensive Model Benchmarking

The authors benchmark 33 models, ranging from traditional word embeddings like GloVe and Komninos to state-of-the-art transformer-based models such as SimCSE, BERT, MiniLM, and SGPT. This wide array includes models both open-sourced and accessible via APIs, delivering a comprehensive performance map. Key findings indicate no single model consistently outperforms others across all tasks, underscoring the field’s progression toward a universal text embedding method.

Detailed Analysis

The analysis reveals significant variability in the models' performance across different tasks. Self-supervised methods trail behind supervised variants, necessitating fine-tuning for competitive performance. They also identify a strong correlation between model size and performance, with larger models generally excelling, albeit at a higher computational cost. Additionally, a notable distinction is observed between models optimized for STS tasks and those geared towards retrieval and search tasks, illustrating the "asymmetric" versus "symmetric" embedding dichotomy.

Multilingual Capabilities

MTEB includes 10 multilingual datasets, offering insight into how models handle cross-lingual tasks. LaBSE and LASER2 exhibit robust performance in bitext mining across multiple languages, although LASER2 shows high performance variance. SGPT-BLOOM-7.1B-msmarco, while tailored for multilingual retrieval, excels in languages seen in pre-training, yet struggles with those it has not encountered.

Practical Implications

Practically, MTEB aids in selecting the most suitable text embeddings for specific applications by providing clear performance metrics across diverse tasks. This allows researchers and practitioners to make informed decisions about model deployment based on their specific requirements, balancing between model performance and computational efficiency.

Future Directions

Looking forward, the authors see potential for expanding MTEB to include longer document datasets, additional multilingual datasets for underrepresented tasks like retrieval, and modalities beyond text, such as image content. They encourage contributions from the research community to further enhance MTEB’s scope and utility.

Conclusion

MTEB stands out as a significant step towards standardizing text embedding evaluations, offering a robust, multifaceted framework that moves beyond the traditional, narrow evaluation regimes. By open-sourcing MTEB and maintaining a public leaderboard, the authors have laid a foundation to fuel advancements in the development and application of text embeddings.

Researchers and practitioners in NLP will find MTEB to be a vital resource in benchmarking and selecting text embeddings suitable for various embedding tasks, contributing to more versatile and effective NLP systems.

Paper to Video (Beta)

No one has generated a video about this paper yet.

Whiteboard

No one has generated a whiteboard explanation for this paper yet.

Open Problems

We haven't generated a list of open problems mentioned in this paper yet.

Collections

Sign up for free to add this paper to one or more collections.

Tweets

Sign up for free to view the 3 tweets with 243 likes about this paper.

HackerNews

  1. Text Embedding Benchmark (2022) (2 points, 0 comments)