-
Notifications
You must be signed in to change notification settings - Fork 289
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
MIRACLRetrieval results are missing for most models. #1550
Comments
Do they need to be run or missing from the leaderboard? |
They are missing, but they were present in the old version of the leaderboard |
Hmm so we need to check the results repo and if they are not there probably rerun them |
It seems strange. In the new leaderboard, only Results for these datasets are available in the results repository, such as for mteb/mteb/benchmarks/benchmarks.py Lines 245 to 248 in d713525
|
Sounds like we need to either rerun MICRACL or rewrite the scores.
@Samoed isn't it because the main score has changed? |
No, |
I have a suspicion, that this might be because of duplicate results, where the incorrect one gets used in the leaderboard. Can't confirm yet though, I'll have a look |
Wait a sec, bge-m3 shouldn't even show up in the leaderboard, since it has no metadata, and it doesn't for me. Are you sure it's bge-m3 we're looking at? @Samoed |
I've added bge-m3 just to show that in result repository some scores for tasks are presented |
If I load the results then I only get MIRACL scores for one model, even when I don't filter models out that don't have metadata. import mteb
all_results = mteb.load_results()
miracl = all_results.filter_tasks(["MIRACLRetrieval"])
print(len(miracl.model_results))
# 1
print(miracl[0])
# model_name='jinaai/jina-embeddings-v3' model_revision='215a6e121fa0183376388ac6b1ae230326bfeaed' task_results=[TaskResult(task_name=MIRACLRetrieval, scores=...)] |
so something is either wrong with the results or with the result loading script |
Also, when I load results, I get these warnings for almost all models:
|
Maybe the loading function should be a bit more graceful, when stuff is missing |
Raising warnings instead of errors when loading results and languages are missing partly fixes the issue. |
I will submit a PR |
I don't think this is solved, just circumvented |
right.. I mean we are still missing a lot of results on MIRACL, just not the Russian ones |
Models results for
MIRACLRetrieval
are missing forMTEB(rus)
The text was updated successfully, but these errors were encountered: