Skip to content
Merged
Show file tree
Hide file tree
Changes from all commits
Commits
File filter

Filter by extension

Filter by extension

Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
Original file line number Diff line number Diff line change
@@ -0,0 +1,26 @@
{
"dataset_revision": "b44c3b011063adb25877c13823db83bb193913c4",
"evaluation_time": 19.167288064956665,
"kg_co2_emissions": null,
"mteb_version": "1.14.5",
"scores": {
"validation": [
{
"cosine_pearson": 0.4651400159752993,
"cosine_spearman": 0.4878186009760077,
"euclidean_pearson": 0.4752496114658185,
"euclidean_spearman": 0.48774220462716233,
"hf_subset": "default",
"languages": [
"cmn-Hans"
],
"main_score": 0.4878186009760077,
"manhattan_pearson": 0.47372608661084276,
"manhattan_spearman": 0.48644628191666606,
"pearson": 0.4651400159752993,
"spearman": 0.4878186009760077
}
]
},
"task_name": "AFQMC"
}
Original file line number Diff line number Diff line change
@@ -0,0 +1,43 @@
{
"dataset_revision": "0f319b1142f28d00e055a6770f3f726ae9b7d865",
"evaluation_time": 38.02136039733887,
"kg_co2_emissions": null,
"mteb_version": "1.14.5",
"scores": {
"test": [
{
"cosine_pearson": 0.49166213462586844,
"cosine_spearman": 0.524470470140877,
"euclidean_pearson": 0.544684886804392,
"euclidean_spearman": 0.5244557775519562,
"hf_subset": "default",
"languages": [
"cmn-Hans"
],
"main_score": 0.524470470140877,
"manhattan_pearson": 0.5438184473109456,
"manhattan_spearman": 0.5235316311551168,
"pearson": 0.49166213462586844,
"spearman": 0.524470470140877
}
],
"validation": [
{
"cosine_pearson": 0.4936732652974832,
"cosine_spearman": 0.5300368024345922,
"euclidean_pearson": 0.5439133111794459,
"euclidean_spearman": 0.5300415812979856,
"hf_subset": "default",
"languages": [
"cmn-Hans"
],
"main_score": 0.5300368024345922,
"manhattan_pearson": 0.543055260482846,
"manhattan_spearman": 0.5293283235846927,
"pearson": 0.4936732652974832,
"spearman": 0.5300368024345922
}
]
},
"task_name": "ATEC"
}
Original file line number Diff line number Diff line change
@@ -0,0 +1,179 @@
{
"dataset_revision": "e8379541af4e31359cca9fbcf4b00f2671dba205",
"evaluation_time": 37.91304087638855,
"kg_co2_emissions": null,
"mteb_version": "1.14.5",
"scores": {
"test": [
{
"accuracy": 0.9621439280359819,
"ap": 0.7088112929976212,
"ap_weighted": 0.7088112929976212,
"f1": 0.9059858298034283,
"f1_weighted": 0.9635045229491273,
"hf_subset": "en-ext",
"languages": [
"eng-Latn"
],
"main_score": 0.9621439280359819,
"scores_per_experiment": [
{
"accuracy": 0.9632683658170914,
"ap": 0.7147810986593035,
"ap_weighted": 0.7147810986593035,
"f1": 0.9082539963254439,
"f1_weighted": 0.9644932743343304
},
{
"accuracy": 0.9640179910044977,
"ap": 0.7191602040706266,
"ap_weighted": 0.7191602040706266,
"f1": 0.9098648648648648,
"f1_weighted": 0.9651701851776815
},
{
"accuracy": 0.9625187406296851,
"ap": 0.7104557246027204,
"ap_weighted": 0.7104557246027204,
"f1": 0.9066521863261547,
"f1_weighted": 0.9638179628227905
},
{
"accuracy": 0.9587706146926537,
"ap": 0.6935395026261338,
"ap_weighted": 0.6935395026261338,
"f1": 0.8999081889207355,
"f1_weighted": 0.9606694026207802
},
{
"accuracy": 0.9617691154422788,
"ap": 0.7061830990102422,
"ap_weighted": 0.7061830990102422,
"f1": 0.9050593431436167,
"f1_weighted": 0.9631442311624211
},
{
"accuracy": 0.9632683658170914,
"ap": 0.7147810986593035,
"ap_weighted": 0.7147810986593035,
"f1": 0.9082539963254439,
"f1_weighted": 0.9644932743343304
},
{
"accuracy": 0.9625187406296851,
"ap": 0.7104557246027204,
"ap_weighted": 0.7104557246027204,
"f1": 0.9066521863261547,
"f1_weighted": 0.9638179628227905
},
{
"accuracy": 0.9632683658170914,
"ap": 0.7147810986593035,
"ap_weighted": 0.7147810986593035,
"f1": 0.9082539963254439,
"f1_weighted": 0.9644932743343304
},
{
"accuracy": 0.9617691154422788,
"ap": 0.7061830990102422,
"ap_weighted": 0.7061830990102422,
"f1": 0.9050593431436167,
"f1_weighted": 0.9631442311624211
},
{
"accuracy": 0.9602698650674663,
"ap": 0.6977922800756162,
"ap_weighted": 0.6977922800756162,
"f1": 0.9019001963328084,
"f1_weighted": 0.9618014307193959
}
]
},
{
"accuracy": 0.9474626865671644,
"ap": 0.7828146988474132,
"ap_weighted": 0.7828146988474132,
"f1": 0.9216277577137338,
"f1_weighted": 0.9490271713794746,
"hf_subset": "en",
"languages": [
"eng-Latn"
],
"main_score": 0.9474626865671644,
"scores_per_experiment": [
{
"accuracy": 0.9492537313432836,
"ap": 0.7885291399610241,
"ap_weighted": 0.7885291399610241,
"f1": 0.9240423602219376,
"f1_weighted": 0.9506906488683344
},
{
"accuracy": 0.9447761194029851,
"ap": 0.7735373338825544,
"ap_weighted": 0.7735373338825544,
"f1": 0.9179712187842268,
"f1_weighted": 0.9465257865452016
},
{
"accuracy": 0.9492537313432836,
"ap": 0.7877593227935444,
"ap_weighted": 0.7877593227935444,
"f1": 0.9236482591266809,
"f1_weighted": 0.9505736010451703
},
{
"accuracy": 0.9462686567164179,
"ap": 0.7784701216890185,
"ap_weighted": 0.7784701216890185,
"f1": 0.919985138201457,
"f1_weighted": 0.947911376623603
},
{
"accuracy": 0.9492537313432836,
"ap": 0.7885291399610241,
"ap_weighted": 0.7885291399610241,
"f1": 0.9240423602219376,
"f1_weighted": 0.9506906488683344
},
{
"accuracy": 0.9567164179104478,
"ap": 0.8148660965853629,
"ap_weighted": 0.8148660965853629,
"f1": 0.9343623213374816,
"f1_weighted": 0.9576883351527506
},
{
"accuracy": 0.9492537313432836,
"ap": 0.7877593227935444,
"ap_weighted": 0.7877593227935444,
"f1": 0.9236482591266809,
"f1_weighted": 0.9505736010451703
},
{
"accuracy": 0.9507462686567164,
"ap": 0.7929413998708746,
"ap_weighted": 0.7929413998708746,
"f1": 0.9257004022461263,
"f1_weighted": 0.9519695886099329
},
{
"accuracy": 0.9388059701492537,
"ap": 0.7555542896205993,
"ap_weighted": 0.7555542896205993,
"f1": 0.9104543780213905,
"f1_weighted": 0.9411372681756828
},
{
"accuracy": 0.9402985074626866,
"ap": 0.7602008213165835,
"ap_weighted": 0.7602008213165835,
"f1": 0.9124228798494196,
"f1_weighted": 0.9425108588605648
}
]
}
]
},
"task_name": "AmazonCounterfactualClassification"
}
Original file line number Diff line number Diff line change
@@ -0,0 +1,95 @@
{
"dataset_revision": "e2d317d38cd51312af73b3d32a06d1a08b442046",
"evaluation_time": 1152.2043359279633,
"kg_co2_emissions": null,
"mteb_version": "1.14.5",
"scores": {
"test": [
{
"accuracy": 0.9702932500000001,
"ap": 0.9565838143923754,
"ap_weighted": 0.9565838143923754,
"f1": 0.9702911531044303,
"f1_weighted": 0.9702911531044303,
"hf_subset": "default",
"languages": [
"eng-Latn"
],
"main_score": 0.9702932500000001,
"scores_per_experiment": [
{
"accuracy": 0.97144,
"ap": 0.9570174556171775,
"ap_weighted": 0.9570174556171775,
"f1": 0.9714398661322265,
"f1_weighted": 0.9714398661322265
},
{
"accuracy": 0.970885,
"ap": 0.9547192832525713,
"ap_weighted": 0.9547192832525713,
"f1": 0.970884086924966,
"f1_weighted": 0.970884086924966
},
{
"accuracy": 0.970705,
"ap": 0.9598885543850582,
"ap_weighted": 0.9598885543850582,
"f1": 0.9707037161079362,
"f1_weighted": 0.970703716107936
},
{
"accuracy": 0.969065,
"ap": 0.9597640806862735,
"ap_weighted": 0.9597640806862735,
"f1": 0.9690608619142589,
"f1_weighted": 0.9690608619142589
},
{
"accuracy": 0.9682375,
"ap": 0.9596866467115923,
"ap_weighted": 0.9596866467115923,
"f1": 0.9682312622033697,
"f1_weighted": 0.9682312622033697
},
{
"accuracy": 0.9709575,
"ap": 0.9551717623527256,
"ap_weighted": 0.9551717623527256,
"f1": 0.9709568315422532,
"f1_weighted": 0.9709568315422532
},
{
"accuracy": 0.971335,
"ap": 0.9590662794386791,
"ap_weighted": 0.9590662794386791,
"f1": 0.9713347784637019,
"f1_weighted": 0.971334778463702
},
{
"accuracy": 0.9715375,
"ap": 0.959086680313462,
"ap_weighted": 0.959086680313462,
"f1": 0.9715373656633037,
"f1_weighted": 0.9715373656633036
},
{
"accuracy": 0.97065,
"ap": 0.954157907709634,
"ap_weighted": 0.954157907709634,
"f1": 0.9706489006721853,
"f1_weighted": 0.9706489006721855
},
{
"accuracy": 0.96812,
"ap": 0.9472794934565799,
"ap_weighted": 0.9472794934565799,
"f1": 0.9681138614201025,
"f1_weighted": 0.9681138614201025
}
]
}
]
},
"task_name": "AmazonPolarityClassification"
}
Loading
Loading