I'm looking at the MTEB leaderboard this AM. Amazingly, mxbai-embed-large-v1 ranks at 12 despite its small size relative to the other Billion-sized parameters models. In addition to strong performance for a relatively small size, it comes with MRL and BQL flexibility, which can slash cost of storing and searching the embedding representations (with a slight degradation in accuracy) https://lnkd.in/d5jq9a2R
Giovanni Rizzi this is the model I was showing you
mixedbread.ai models are great! 👏❤️
Good stuff. 👀 Also, I haven't ran a T-test on this particular set of data, but I would guess that these 11 LLMs have scores that are not statistically significant from each other, at the 95% or even 90% 2-talied confidence level. ❓ What do you think, Jo Kristian Bergum
🤗 Sentence Transformers, SetFit & NLTK maintainer, MLE @ Hugging Face
4moLuckily, you can now filter by model size 😄