{ "dataset_revision": "d80d48c1eb48d3562165c59d59d0034df9fff0bf", "evaluation_time": 3.4560580253601074, "kg_co2_emissions": null, "mteb_version": "1.14.5", "scores": { "test": [ { "accuracy": 0.988622891016872, "f1": 0.9876863201526358, "f1_weighted": 0.9886145312611918, "hf_subset": "en", "languages": [ "eng-Latn" ], "main_score": 0.988622891016872, "scores_per_experiment": [ { "accuracy": 0.9886000911992704, "f1": 0.9876686073808003, "f1_weighted": 0.9885948282716679 }, { "accuracy": 0.988828089375285, "f1": 0.9879181734502591, "f1_weighted": 0.9888214588322477 }, { "accuracy": 0.9883720930232558, "f1": 0.9874013413542513, "f1_weighted": 0.9883657451493001 }, { "accuracy": 0.9886000911992704, "f1": 0.9876712682458778, "f1_weighted": 0.9885902791398906 }, { "accuracy": 0.9886000911992704, "f1": 0.98760989347103, "f1_weighted": 0.9885942347349939 }, { "accuracy": 0.9886000911992704, "f1": 0.9877299821556481, "f1_weighted": 0.9885908726765649 }, { "accuracy": 0.9886000911992704, "f1": 0.9876681953349723, "f1_weighted": 0.9885866355762386 }, { "accuracy": 0.9886000911992704, "f1": 0.9877080293461734, "f1_weighted": 0.9885904354570986 }, { "accuracy": 0.9886000911992704, "f1": 0.9875128951653935, "f1_weighted": 0.9885936470949775 }, { "accuracy": 0.988828089375285, "f1": 0.9879748156219518, "f1_weighted": 0.988817175678938 } ] } ] }, "task_name": "MTOPDomainClassification" }