-
Notifications
You must be signed in to change notification settings - Fork 289
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Integrate ModernBERT #1624
Comments
Thanks @Samoed! ModernBERT is the base model, so if you use it out of the box it will be pretty bad, just like an un-finetuned BERT or RoBERTa. @NohTow and @bclavie did some fine tuning on MS MARCO but I don’t think they’ve uploaded the models anywhere. They did put their fine tuning scripts here: https://github.com/AnswerDotAI/ModernBERT/blob/main/examples/train_st.py (and similar for ColBERT in the repo). I expect others will replace BERT with it in their pipelines and we will see more retrieval models with it soon! |
Hello, We decided not to chase the top of BEIR leaderboard because fine tuning to this extend is a whole project in itself and takes a lot of work if you do not have the data for it available. Also, to some extend, the leaderboard is a bit gamed and even if we put the time and energy to grind the leaderboard, we might have come a bit short or end up with a model that is not performing as we believe a model should. Thus, to avoid wasting time and get people only comparing the BEIR scores, we preferred to compare the models in a simple setup to get a signal comparing the actual potential of the base models and let the people that already have extensive pipelines available take the model and do a proper fine-tuning. These actors have seen the model and we have good reasons to believe that they will indeed do this fine-tuning in the future! Besides, I am also doing some experiments on my own, which might end up with a model that is not as strong as the top models, but way better than what we trained in the paper! Edit: I also have the checkpoints of the models we trained for the experiments, but again, not sure reporting these one on MTEB is worth it. |
so some thought. It can be reasonable (as a reference) to benchmark models like BERT, ModernBERT etc. These are fairly easy to benchmark (it can be run from the CLI). However I expect that we will see competitive finetunes due to:
|
Arxiv: https://arxiv.org/abs/2412.13663
Model: https://huggingface.co/answerdotai/ModernBERT-base
ModernBERT was evaluated on BEIR, and I think it could be integrated into MTEB with a specific configuration. I tried adding it using SentenceTransformers with different pooling methods, but my results were much lower than those reported.
@orionw, since you’re one of the co-authors (congrats, by the way!), do you have scripts to reproduce the results?
The text was updated successfully, but these errors were encountered: