Muennighoff's picture
Add MTEB evaluation
039f451
raw
history blame
366 Bytes
{
"test": {
"en": {
"accuracy": 0.703866845998655,
"accuracy_stderr": 0.00456616151893328,
"f1": 0.6896745198729209,
"f1_stderr": 0.006294250065644286,
"main_score": 0.703866845998655
},
"evaluation_time": 387.35
},
"dataset_version": null,
"mteb_version": "0.0.2"
}