-
Notifications
You must be signed in to change notification settings - Fork 554
Add new dataset: LIMIT #3093
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Add new dataset: LIMIT #3093
Conversation
isaac-chung
left a comment
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Congrats on the paper!
|
Thanks @isaac-chung! It's still failing the tests which is a bit weird. @Samoed is this cuz of my PR or is this good to merge? I see you've been doing a lot of work to fix the branches CI, thank you! |
KennethEnevoldsen
left a comment
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Once tests pass this is good to merge - congrats on the paper, very happy to see it!
This adds a new dataset, LIMIT, that is available today from my internship with GDM:
Paper: https://arxiv.org/abs/2508.21038
Code: https://github.com/google-deepmind/limit
I have run the following models on the task (adding the results to the pr). These can be run using the
mteb run -m {model_name} -t {task_name}command.sentence-transformers/paraphrase-multilingual-MiniLM-L12-v2intfloat/multilingual-e5-smallI have checked that the performance is neither trivial (both models gain close to perfect scores) nor random (both models gain close to random scores).
I have considered the size of the dataset and reduced it if it is too big (2048 examples is typically large enough for most tasks)