Arctic-Embed: Scalable, Efficient, and Accurate Text Embedding Models
Abstract
The arctic-embed text embedding models achieved state-of-the-art retrieval accuracy for their size, surpassing closed-source models through detailed training methods and ablation studies.
This report describes the training dataset creation and recipe behind the family of arctic-embed text embedding models (a set of five models ranging from 22 to 334 million parameters with weights open-sourced under an Apache-2 license). At the time of their release, each model achieved state-of-the-art retrieval accuracy for models of their size on the MTEB Retrieval leaderboard, with the largest model, arctic-embed-l outperforming closed source embedding models such as Cohere's embed-v3 and Open AI's text-embed-3-large. In addition to the details of our training recipe, we have provided several informative ablation studies, which we believe are the cause of our model performance.
Get this paper in your agent:
hf papers read 2405.05374 Don't have the latest CLI?
curl -LsSf https://hf.co/cli/install.sh | bash Models citing this paper 8
Browse 8 models citing this paperDatasets citing this paper 0
No dataset linking this paper