diff --git a/mteb/models/linq_models.py b/mteb/models/linq_models.py index ec9e0ee9fe..2c0eafa591 100644 --- a/mteb/models/linq_models.py +++ b/mteb/models/linq_models.py @@ -6,6 +6,7 @@ from mteb.encoder_interface import PromptType from mteb.model_meta import ModelMeta +from mteb.models.e5_instruct import E5_MISTRAL_TRAINING_DATA from mteb.models.instruct_wrapper import instruct_wrapper @@ -42,5 +43,6 @@ def instruction_template( use_instructions=True, public_training_code=None, public_training_data=None, - training_datasets=None, + adapted_from="intfloat/e5-mistral-7b-instruct", + training_datasets=E5_MISTRAL_TRAINING_DATA, ) diff --git a/mteb/models/misc_models.py b/mteb/models/misc_models.py index 610d622ed4..8d5ad3fcaa 100644 --- a/mteb/models/misc_models.py +++ b/mteb/models/misc_models.py @@ -6,7 +6,7 @@ from mteb.model_meta import ModelMeta, sentence_transformers_loader from mteb.models.bge_models import bge_m3_training_data, bge_training_data -from mteb.models.e5_models import E5_TRAINING_DATA +from mteb.models.e5_instruct import E5_MISTRAL_TRAINING_DATA from mteb.models.sentence_transformers_models import sent_trf_training_dataset Haon_Chen__speed_embedding_7b_instruct = ModelMeta( @@ -203,9 +203,9 @@ reference="https://huggingface.co/BeastyZ/e5-R-mistral-7b", similarity_fn_name="cosine", use_instructions=None, - training_datasets=E5_TRAINING_DATA, + training_datasets=E5_MISTRAL_TRAINING_DATA, # not MTEB: {"BeastyZ/E5-R": ["train"]}, - adapted_from="/ConRetriever/public_weight_mistral", + adapted_from="intfloat/e5-mistral-7b-instruct", superseded_by=None, ) Lajavaness__bilingual_embedding_base = ModelMeta(