Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Models missing after leaderboard code refresh #2096

Closed
sfc-gh-lmerrick opened this issue Feb 18, 2025 · 5 comments
Closed

Models missing after leaderboard code refresh #2096

sfc-gh-lmerrick opened this issue Feb 18, 2025 · 5 comments
Labels
more information needed More information is needed before this issue can be resolved

Comments

@sfc-gh-lmerrick
Copy link

For some reason the new leaderboard only displays Snowflake's Arctic Embed v1.5 model, dropping both our v1 and v2 models from the listings. The actual MTEB eval results are still present in all v1 and v2 models' model cards, so it seems that something went awry in the leaderboard code changes themselves.

@Samoed
Copy link
Collaborator

Samoed commented Feb 18, 2025

Can you please provide more information about the leaderboard you are referring to? I am able to see all the models on the MTEB (Multilingual, v1) leaderboard.
Image

@isaac-chung isaac-chung changed the title Models missing after leaderbaord code refresh Models missing after leaderboard code refresh Feb 18, 2025
@isaac-chung isaac-chung added the more information needed More information is needed before this issue can be resolved label Feb 19, 2025
@sfc-gh-lmerrick
Copy link
Author

Apologies for the delayed response, I didn't see your reply until today!

MTEB(eng, v1)

The original English benchmark by Muennighoff et al., (2023). This page is an adaptation of the old MTEB leaderboard.

This page is missing our models, despite them showing up on the old MTEB leaderboard.

Results for our models also appear to be missing from the BEIR page, despite MTEB Retrieval being adapted from BEIR and the datasets overlapping (as seen in the figure from the MTEB paper)

Image

@KennethEnevoldsen
Copy link
Contributor

They do appear if you press "Allow All":

Image

They are simply being filtered now due to not being zero-shot on the leaderboard.

As an important provider in the embedding models space we would love to hear your opinion on how we should set the default over on #2119 (which is currently ongoing, but we will probably make a decision soon-ish)

@KennethEnevoldsen
Copy link
Contributor

I believe this resolved this issue, but do add to the discussion over on #2119

@sfc-gh-lmerrick
Copy link
Author

sfc-gh-lmerrick commented Feb 21, 2025

Ah, I think I have found the source of the confusion:

MTEB(eng, v1)

The original English benchmark by Muennighoff et al., (2023). This page is an adaptation of the old MTEB leaderboard.

This description is misleading, because the page has changed which models are displayed. I will chime in separately on #2119, but the choice of showing all models which do not report their training data but disqualifying those which do report it gives a very misleading view!

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
more information needed More information is needed before this issue can be resolved
Projects
None yet
Development

No branches or pull requests

4 participants