Skip to content
Closed
Show file tree
Hide file tree
Changes from all commits
Commits
File filter

Filter by extension

Filter by extension

Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
Original file line number Diff line number Diff line change
@@ -0,0 +1,179 @@
{
"dataset_revision": "e8379541af4e31359cca9fbcf4b00f2671dba205",
"evaluation_time": 199.27316522598267,
"kg_co2_emissions": null,
"mteb_version": "1.15.3",
"scores": {
"test": [
{
"accuracy": 0.9453523238380809,
"ap": 0.6242264840836734,
"ap_weighted": 0.6242264840836734,
"f1": 0.8713103677336654,
"f1_weighted": 0.9485637995412656,
"hf_subset": "en-ext",
"languages": [
"eng-Latn"
],
"main_score": 0.9453523238380809,
"scores_per_experiment": [
{
"accuracy": 0.972263868065967,
"ap": 0.765857076600446,
"ap_weighted": 0.765857076600446,
"f1": 0.9259479304038374,
"f1_weighted": 0.9723077694286801
},
{
"accuracy": 0.9385307346326837,
"ap": 0.5929303585430976,
"ap_weighted": 0.5929303585430976,
"f1": 0.8589546542333717,
"f1_weighted": 0.9428192659116885
},
{
"accuracy": 0.9437781109445277,
"ap": 0.6172761239156941,
"ap_weighted": 0.6172761239156941,
"f1": 0.8693023055308442,
"f1_weighted": 0.9474020577819363
},
{
"accuracy": 0.9370314842578711,
"ap": 0.5767602868053913,
"ap_weighted": 0.5767602868053913,
"f1": 0.8532322539070483,
"f1_weighted": 0.9410219237983866
},
{
"accuracy": 0.9250374812593704,
"ap": 0.5405204784616745,
"ap_weighted": 0.5405204784616745,
"f1": 0.8347701149425288,
"f1_weighted": 0.9314459149735478
},
{
"accuracy": 0.9302848575712144,
"ap": 0.5615245170474433,
"ap_weighted": 0.5615245170474433,
"f1": 0.8444314868804664,
"f1_weighted": 0.9359160070110716
},
{
"accuracy": 0.9512743628185907,
"ap": 0.6404631060625905,
"ap_weighted": 0.6404631060625905,
"f1": 0.8803714541538878,
"f1_weighted": 0.9532764897429103
},
{
"accuracy": 0.9685157421289355,
"ap": 0.7364087610974286,
"ap_weighted": 0.7364087610974286,
"f1": 0.9156738207760151,
"f1_weighted": 0.9685157421289355
},
{
"accuracy": 0.9355322338830585,
"ap": 0.5785197882519725,
"ap_weighted": 0.5785197882519725,
"f1": 0.8528341774035477,
"f1_weighted": 0.940163325045911
},
{
"accuracy": 0.9512743628185907,
"ap": 0.632004344050996,
"ap_weighted": 0.632004344050996,
"f1": 0.8775854791051075,
"f1_weighted": 0.9527694995895889
}
]
},
{
"accuracy": 0.903134328358209,
"ap": 0.6342364739316405,
"ap_weighted": 0.6342364739316405,
"f1": 0.8554214552412622,
"f1_weighted": 0.9059539168268289,
"hf_subset": "en",
"languages": [
"eng-Latn"
],
"main_score": 0.903134328358209,
"scores_per_experiment": [
{
"accuracy": 0.9029850746268657,
"ap": 0.6202534220934586,
"ap_weighted": 0.6202534220934586,
"f1": 0.8496715556491393,
"f1_weighted": 0.9041876352053108
},
{
"accuracy": 0.908955223880597,
"ap": 0.6532598866715797,
"ap_weighted": 0.6532598866715797,
"f1": 0.865443686849565,
"f1_weighted": 0.9120387186308277
},
{
"accuracy": 0.9014925373134328,
"ap": 0.625611353500422,
"ap_weighted": 0.625611353500422,
"f1": 0.8525528169014085,
"f1_weighted": 0.9042818478032374
},
{
"accuracy": 0.9074626865671642,
"ap": 0.6477146043981483,
"ap_weighted": 0.6477146043981483,
"f1": 0.862893430502746,
"f1_weighted": 0.9104961961422293
},
{
"accuracy": 0.9149253731343283,
"ap": 0.6667297102274885,
"ap_weighted": 0.6667297102274885,
"f1": 0.8716643311523998,
"f1_weighted": 0.9170383803262476
},
{
"accuracy": 0.9,
"ap": 0.6121060646664538,
"ap_weighted": 0.6121060646664538,
"f1": 0.8458929386764438,
"f1_weighted": 0.9014991880971261
},
{
"accuracy": 0.9194029850746268,
"ap": 0.6748907174818423,
"ap_weighted": 0.6748907174818423,
"f1": 0.874766355140187,
"f1_weighted": 0.9202957176733156
},
{
"accuracy": 0.9283582089552239,
"ap": 0.7029731593764759,
"ap_weighted": 0.7029731593764759,
"f1": 0.8861334957300061,
"f1_weighted": 0.9283582089552239
},
{
"accuracy": 0.8820895522388059,
"ap": 0.5850775414601058,
"ap_weighted": 0.5850775414601058,
"f1": 0.8315463176018663,
"f1_weighted": 0.887736071639663
},
{
"accuracy": 0.8656716417910447,
"ap": 0.5537482794404294,
"ap_weighted": 0.5537482794404294,
"f1": 0.8136496242088608,
"f1_weighted": 0.8736072037951069
}
]
}
]
},
"task_name": "AmazonCounterfactualClassification"
}
Original file line number Diff line number Diff line change
@@ -0,0 +1,95 @@
{
"dataset_revision": "e2d317d38cd51312af73b3d32a06d1a08b442046",
"evaluation_time": 3150.8407397270203,
"kg_co2_emissions": null,
"mteb_version": "1.15.3",
"scores": {
"test": [
{
"accuracy": 0.9429605000000001,
"ap": 0.9130887530384255,
"ap_weighted": 0.9130887530384255,
"f1": 0.9429070662237378,
"f1_weighted": 0.9429070662237378,
"hf_subset": "default",
"languages": [
"eng-Latn"
],
"main_score": 0.9429605000000001,
"scores_per_experiment": [
{
"accuracy": 0.938855,
"ap": 0.9100520593271506,
"ap_weighted": 0.9100520593271506,
"f1": 0.9388533687775609,
"f1_weighted": 0.938853368777561
},
{
"accuracy": 0.9225475,
"ap": 0.8737119551432224,
"ap_weighted": 0.8737119551432224,
"f1": 0.9223566196264703,
"f1_weighted": 0.9223566196264703
},
{
"accuracy": 0.9533625,
"ap": 0.9361003901692877,
"ap_weighted": 0.9361003901692877,
"f1": 0.9533584941216818,
"f1_weighted": 0.9533584941216817
},
{
"accuracy": 0.947655,
"ap": 0.9347141077611154,
"ap_weighted": 0.9347141077611154,
"f1": 0.9476225906595737,
"f1_weighted": 0.9476225906595737
},
{
"accuracy": 0.94949,
"ap": 0.9254994242406176,
"ap_weighted": 0.9254994242406176,
"f1": 0.9494894811546877,
"f1_weighted": 0.9494894811546876
},
{
"accuracy": 0.943205,
"ap": 0.9064453723569432,
"ap_weighted": 0.9064453723569432,
"f1": 0.9431491435204298,
"f1_weighted": 0.9431491435204298
},
{
"accuracy": 0.93975,
"ap": 0.9036596651333859,
"ap_weighted": 0.9036596651333859,
"f1": 0.9397089133713968,
"f1_weighted": 0.9397089133713967
},
{
"accuracy": 0.9524525,
"ap": 0.9311988050383735,
"ap_weighted": 0.9311988050383735,
"f1": 0.9524524809783178,
"f1_weighted": 0.9524524809783178
},
{
"accuracy": 0.952835,
"ap": 0.9282735812161005,
"ap_weighted": 0.9282735812161005,
"f1": 0.9528320301053308,
"f1_weighted": 0.9528320301053308
},
{
"accuracy": 0.9294525,
"ap": 0.881232169998059,
"ap_weighted": 0.881232169998059,
"f1": 0.9292475399219274,
"f1_weighted": 0.9292475399219274
}
]
}
]
},
"task_name": "AmazonPolarityClassification"
}
Loading
Loading