Skip to content
Merged
Show file tree
Hide file tree
Changes from all commits
Commits
File filter

Filter by extension

Filter by extension

Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
Original file line number Diff line number Diff line change
@@ -0,0 +1,73 @@
{
"dataset_revision": "c0ba03d058e3e1b2f3fd20518875a4563dd12db4",
"task_name": "CEDRClassification",
"mteb_version": "1.38.0",
"scores": {
"test": [
{
"accuracy": 0.440489,
"f1": 0.331878,
"lrap": 0.665266,
"scores_per_experiment": [
{
"accuracy": 0.45271,
"f1": 0.289083,
"lrap": 0.649947
},
{
"accuracy": 0.504251,
"f1": 0.405432,
"lrap": 0.706323
},
{
"accuracy": 0.456429,
"f1": 0.300366,
"lrap": 0.644049
},
{
"accuracy": 0.478215,
"f1": 0.362387,
"lrap": 0.673964
},
{
"accuracy": 0.46068,
"f1": 0.354197,
"lrap": 0.681243
},
{
"accuracy": 0.391605,
"f1": 0.300982,
"lrap": 0.653507
},
{
"accuracy": 0.353348,
"f1": 0.3574,
"lrap": 0.669075
},
{
"accuracy": 0.489904,
"f1": 0.317934,
"lrap": 0.66525
},
{
"accuracy": 0.399044,
"f1": 0.268548,
"lrap": 0.649841
},
{
"accuracy": 0.418704,
"f1": 0.362453,
"lrap": 0.659458
}
],
"main_score": 0.440489,
"hf_subset": "default",
"languages": [
"rus-Cyrl"
]
}
]
},
"evaluation_time": 12.207267761230469,
"kg_co2_emissions": null
}
Original file line number Diff line number Diff line change
@@ -0,0 +1,73 @@
{
"dataset_revision": "3765c0d1de6b7d264bc459433c45e5a75513839c",
"task_name": "GeoreviewClassification",
"mteb_version": "1.38.0",
"scores": {
"test": [
{
"accuracy": 0.433936,
"f1": 0.431696,
"f1_weighted": 0.431706,
"scores_per_experiment": [
{
"accuracy": 0.444336,
"f1": 0.44221,
"f1_weighted": 0.442259
},
{
"accuracy": 0.444824,
"f1": 0.441732,
"f1_weighted": 0.441737
},
{
"accuracy": 0.444336,
"f1": 0.435421,
"f1_weighted": 0.43545
},
{
"accuracy": 0.414062,
"f1": 0.417061,
"f1_weighted": 0.417095
},
{
"accuracy": 0.421387,
"f1": 0.427308,
"f1_weighted": 0.427333
},
{
"accuracy": 0.420898,
"f1": 0.417962,
"f1_weighted": 0.417969
},
{
"accuracy": 0.467285,
"f1": 0.464286,
"f1_weighted": 0.464287
},
{
"accuracy": 0.472168,
"f1": 0.459022,
"f1_weighted": 0.458967
},
{
"accuracy": 0.374023,
"f1": 0.374138,
"f1_weighted": 0.374137
},
{
"accuracy": 0.436035,
"f1": 0.437821,
"f1_weighted": 0.437832
}
],
"main_score": 0.433936,
"hf_subset": "default",
"languages": [
"rus-Cyrl"
]
}
]
},
"evaluation_time": 9.894148588180542,
"kg_co2_emissions": null
}
Original file line number Diff line number Diff line change
@@ -0,0 +1,34 @@
{
"dataset_revision": "97a313c8fc85b47f13f33e7e9a95c1ad888c7fec",
"task_name": "GeoreviewClusteringP2P",
"mteb_version": "1.38.0",
"scores": {
"test": [
{
"v_measures": {
"Level 0": [
0.682167,
0.688579,
0.71259,
0.702199,
0.683211,
0.681277,
0.692437,
0.688077,
0.700128,
0.695916
]
},
"v_measure": 0.692658,
"v_measure_std": 0.009598,
"main_score": 0.692658,
"hf_subset": "default",
"languages": [
"rus-Cyrl"
]
}
]
},
"evaluation_time": 11.321293592453003,
"kg_co2_emissions": null
}
Original file line number Diff line number Diff line change
@@ -0,0 +1,73 @@
{
"dataset_revision": "2fe05ee6b5832cda29f2ef7aaad7b7fe6a3609eb",
"task_name": "HeadlineClassification",
"mteb_version": "1.38.0",
"scores": {
"test": [
{
"accuracy": 0.75874,
"f1": 0.758527,
"f1_weighted": 0.758529,
"scores_per_experiment": [
{
"accuracy": 0.704102,
"f1": 0.708731,
"f1_weighted": 0.708757
},
{
"accuracy": 0.769043,
"f1": 0.768265,
"f1_weighted": 0.768256
},
{
"accuracy": 0.765137,
"f1": 0.764255,
"f1_weighted": 0.764245
},
{
"accuracy": 0.755859,
"f1": 0.756639,
"f1_weighted": 0.756634
},
{
"accuracy": 0.782715,
"f1": 0.783646,
"f1_weighted": 0.783642
},
{
"accuracy": 0.780273,
"f1": 0.778675,
"f1_weighted": 0.778675
},
{
"accuracy": 0.745605,
"f1": 0.744286,
"f1_weighted": 0.744271
},
{
"accuracy": 0.753906,
"f1": 0.752349,
"f1_weighted": 0.752357
},
{
"accuracy": 0.73877,
"f1": 0.737767,
"f1_weighted": 0.7378
},
{
"accuracy": 0.791992,
"f1": 0.790661,
"f1_weighted": 0.790656
}
],
"main_score": 0.75874,
"hf_subset": "default",
"languages": [
"rus-Cyrl"
]
}
]
},
"evaluation_time": 5.842599630355835,
"kg_co2_emissions": null
}
Original file line number Diff line number Diff line change
@@ -0,0 +1,95 @@
{
"dataset_revision": "601651fdc45ef243751676e62dd7a19f491c0285",
"task_name": "InappropriatenessClassification",
"mteb_version": "1.38.0",
"scores": {
"test": [
{
"accuracy": 0.620605,
"f1": 0.615843,
"f1_weighted": 0.615843,
"ap": 0.575877,
"ap_weighted": 0.575877,
"scores_per_experiment": [
{
"accuracy": 0.647949,
"f1": 0.647808,
"f1_weighted": 0.647808,
"ap": 0.596777,
"ap_weighted": 0.596777
},
{
"accuracy": 0.640625,
"f1": 0.638473,
"f1_weighted": 0.638473,
"ap": 0.587444,
"ap_weighted": 0.587444
},
{
"accuracy": 0.620605,
"f1": 0.612567,
"f1_weighted": 0.612567,
"ap": 0.580735,
"ap_weighted": 0.580735
},
{
"accuracy": 0.603516,
"f1": 0.596073,
"f1_weighted": 0.596073,
"ap": 0.560185,
"ap_weighted": 0.560185
},
{
"accuracy": 0.615723,
"f1": 0.599433,
"f1_weighted": 0.599433,
"ap": 0.580305,
"ap_weighted": 0.580305
},
{
"accuracy": 0.556152,
"f1": 0.549493,
"f1_weighted": 0.549493,
"ap": 0.530613,
"ap_weighted": 0.530613
},
{
"accuracy": 0.633789,
"f1": 0.633257,
"f1_weighted": 0.633257,
"ap": 0.583527,
"ap_weighted": 0.583527
},
{
"accuracy": 0.647949,
"f1": 0.641862,
"f1_weighted": 0.641862,
"ap": 0.591337,
"ap_weighted": 0.591337
},
{
"accuracy": 0.607422,
"f1": 0.607272,
"f1_weighted": 0.607272,
"ap": 0.564817,
"ap_weighted": 0.564817
},
{
"accuracy": 0.632324,
"f1": 0.632191,
"f1_weighted": 0.632191,
"ap": 0.583029,
"ap_weighted": 0.583029
}
],
"main_score": 0.620605,
"hf_subset": "default",
"languages": [
"rus-Cyrl"
]
}
]
},
"evaluation_time": 5.455994129180908,
"kg_co2_emissions": null
}
Loading
Loading