Replies: 3 comments 5 replies
-
I just checked the models in descending order for three things:
and the winner seems to be multilingual-e5-small. Even though the multilingual fine tuning data is very little in comparison to EN (and ZH), it seems to outperform our current multilingual model distiluse-base-multilingual-cased-v2. Apart from that, it's smaller and has less dimensions. Also, the issue with dense layers in SentenceTransformers (Python, 512 dimensions vs. without dense layer 768 dimensions) shouldn't apply (still need to check though). |
Beta Was this translation helpful? Give feedback.
-
Referencing #36 for completeness; default model is now gte-small. |
Beta Was this translation helpful? Give feedback.
-
SemanticFinder dropdown model selector@MentalGear as you were asking about the model dropdown: well it grew organically. It started with a hand full of models, then I automated the model mining process from HF with likes and downloads but quickly noticed, that it was annoying me that new but high-scoring models were at the bottom, so I switched the order a little with my personal most-used models on top. So pretty subjective to my personal preferences. I was thinking about providing only a text input and leaving the model choice entirely to the user but it might be frustrating when there are typos or similar. Also, atm I test all models I integrate so at least I can make sure all models do work too. I might add some kind of model selection help in the future but I don't want to reproduce MTEB. Instead, maybe a short (and subjective) guideline would be more appropriate... Any ideas welcome! |
Beta Was this translation helpful? Give feedback.
-
The Massive Text Embedding Benchmark (MTEB) Leaderboard shows that there might be some interesting models to replace all-MiniLM-L6-v2 as default model.
E.g. bge-small-en with only 134Mb (unquantized!) looks promising. It has only 384 dimensions which is important to lower memory usage.
But there are other candidates too with even smaller sizes:
E.g. gte-small with only 67Mb (unquantized).
Beta Was this translation helpful? Give feedback.
All reactions