Pringled commited on
Commit
55b8b97
1 Parent(s): ed5b306

Upload README.md with huggingface_hub

Browse files
Files changed (1) hide show
  1. README.md +433 -434
README.md CHANGED
@@ -1,15 +1,21 @@
1
  ---
2
  library_name: model2vec
3
  license: mit
 
 
 
 
4
  model-index:
5
  - name: potion-base-4M
6
  results:
7
- - dataset:
8
- config: en-ext
 
9
  name: MTEB AmazonCounterfactualClassification (en-ext)
10
- revision: e8379541af4e31359cca9fbcf4b00f2671dba205
11
- split: test
12
  type: mteb/amazon_counterfactual
 
 
 
13
  metrics:
14
  - type: accuracy
15
  value: 67.5712143928036
@@ -23,14 +29,14 @@ model-index:
23
  value: 73.78539599566238
24
  - type: main_score
25
  value: 67.5712143928036
26
- task:
27
  type: Classification
28
- - dataset:
29
- config: en
30
  name: MTEB AmazonCounterfactualClassification (en)
31
- revision: e8379541af4e31359cca9fbcf4b00f2671dba205
32
- split: test
33
  type: mteb/amazon_counterfactual
 
 
 
34
  metrics:
35
  - type: accuracy
36
  value: 68.38805970149254
@@ -44,14 +50,14 @@ model-index:
44
  value: 71.3687043183968
45
  - type: main_score
46
  value: 68.38805970149254
47
- task:
48
  type: Classification
49
- - dataset:
50
- config: default
51
  name: MTEB AmazonPolarityClassification (default)
52
- revision: e2d317d38cd51312af73b3d32a06d1a08b442046
53
- split: test
54
  type: mteb/amazon_polarity
 
 
 
55
  metrics:
56
  - type: accuracy
57
  value: 71.69602499999999
@@ -65,14 +71,14 @@ model-index:
65
  value: 71.42572451950106
66
  - type: main_score
67
  value: 71.69602499999999
68
- task:
69
  type: Classification
70
- - dataset:
71
- config: en
72
  name: MTEB AmazonReviewsClassification (en)
73
- revision: 1399c76144fd37290681b995c656ef9b2e06e26d
74
- split: test
75
  type: mteb/amazon_reviews_multi
 
 
 
76
  metrics:
77
  - type: accuracy
78
  value: 34.321999999999996
@@ -82,14 +88,14 @@ model-index:
82
  value: 33.68591037434848
83
  - type: main_score
84
  value: 34.321999999999996
85
- task:
86
- type: Classification
87
- - dataset:
88
- config: default
89
  name: MTEB ArguAna (default)
90
- revision: c22ab2a51041ffd869aaddef7af8d8215647e41a
91
- split: test
92
  type: mteb/arguana
 
 
 
93
  metrics:
94
  - type: main_score
95
  value: 39.449
@@ -373,14 +379,14 @@ model-index:
373
  value: 39.26
374
  - type: recall_at_5
375
  value: 49.431000000000004
376
- task:
377
- type: Retrieval
378
- - dataset:
379
- config: default
380
  name: MTEB ArxivClusteringP2P (default)
381
- revision: a122ad7f3f0291bf49cc6f4d32aa80929df69d5d
382
- split: test
383
  type: mteb/arxiv-clustering-p2p
 
 
 
384
  metrics:
385
  - type: main_score
386
  value: 34.19332453253303
@@ -388,14 +394,14 @@ model-index:
388
  value: 34.19332453253303
389
  - type: v_measure_std
390
  value: 14.189152053888938
391
- task:
392
  type: Clustering
393
- - dataset:
394
- config: default
395
  name: MTEB ArxivClusteringS2S (default)
396
- revision: f910caf1a6075f7329cdf8c1a6135696f37dbd53
397
- split: test
398
  type: mteb/arxiv-clustering-s2s
 
 
 
399
  metrics:
400
  - type: main_score
401
  value: 24.104443846867934
@@ -403,14 +409,14 @@ model-index:
403
  value: 24.104443846867934
404
  - type: v_measure_std
405
  value: 15.164667270049211
406
- task:
407
- type: Clustering
408
- - dataset:
409
- config: default
410
  name: MTEB AskUbuntuDupQuestions (default)
411
- revision: 2000358ca161889fa9c082cb41daa8dcfb161a54
412
- split: test
413
  type: mteb/askubuntudupquestions-reranking
 
 
 
414
  metrics:
415
  - type: main_score
416
  value: 53.803562861722696
@@ -430,14 +436,14 @@ model-index:
430
  value: 21.221447117685774
431
  - type: nAUC_mrr_std
432
  value: 5.470983299654257
433
- task:
434
- type: Reranking
435
- - dataset:
436
- config: default
437
  name: MTEB BIOSSES (default)
438
- revision: d3fb88f8f02e40887cd149695127462bbcf29b4a
439
- split: test
440
  type: mteb/biosses-sts
 
 
 
441
  metrics:
442
  - type: cosine_pearson
443
  value: 75.18540989714172
@@ -457,14 +463,14 @@ model-index:
457
  value: 75.18540989714172
458
  - type: spearman
459
  value: 71.87392685280749
460
- task:
461
- type: STS
462
- - dataset:
463
- config: default
464
  name: MTEB Banking77Classification (default)
465
- revision: 0fd18e25b25c072e09e0d92ab615fda904d66300
466
- split: test
467
  type: mteb/banking77
 
 
 
468
  metrics:
469
  - type: accuracy
470
  value: 69.69480519480518
@@ -474,14 +480,14 @@ model-index:
474
  value: 68.4951487448326
475
  - type: main_score
476
  value: 69.69480519480518
477
- task:
478
- type: Classification
479
- - dataset:
480
- config: default
481
  name: MTEB BiorxivClusteringP2P (default)
482
- revision: 65b79d1d13f80053f67aca9498d9402c2d9f1f40
483
- split: test
484
  type: mteb/biorxiv-clustering-p2p
 
 
 
485
  metrics:
486
  - type: main_score
487
  value: 29.09840292638104
@@ -489,14 +495,14 @@ model-index:
489
  value: 29.09840292638104
490
  - type: v_measure_std
491
  value: 0.8130265845858159
492
- task:
493
  type: Clustering
494
- - dataset:
495
- config: default
496
  name: MTEB BiorxivClusteringS2S (default)
497
- revision: 258694dd0231531bc1fd9de6ceb52a0853c6d908
498
- split: test
499
  type: mteb/biorxiv-clustering-s2s
 
 
 
500
  metrics:
501
  - type: main_score
502
  value: 18.622216350398414
@@ -504,14 +510,14 @@ model-index:
504
  value: 18.622216350398414
505
  - type: v_measure_std
506
  value: 0.6293204014994399
507
- task:
508
- type: Clustering
509
- - dataset:
510
- config: default
511
  name: MTEB CQADupstackAndroidRetrieval (default)
512
- revision: f46a197baaae43b4f621051089b82a364682dfeb
513
- split: test
514
  type: mteb/cqadupstack-android
 
 
 
515
  metrics:
516
  - type: main_score
517
  value: 28.634999999999998
@@ -795,14 +801,14 @@ model-index:
795
  value: 26.143
796
  - type: recall_at_5
797
  value: 30.441000000000003
798
- task:
799
  type: Retrieval
800
- - dataset:
801
- config: default
802
  name: MTEB CQADupstackEnglishRetrieval (default)
803
- revision: ad9991cb51e31e31e430383c75ffb2885547b5f0
804
- split: test
805
  type: mteb/cqadupstack-english
 
 
 
806
  metrics:
807
  - type: main_score
808
  value: 23.933
@@ -1086,14 +1092,14 @@ model-index:
1086
  value: 22.034000000000002
1087
  - type: recall_at_5
1088
  value: 25.663000000000004
1089
- task:
1090
  type: Retrieval
1091
- - dataset:
1092
- config: default
1093
  name: MTEB CQADupstackGamingRetrieval (default)
1094
- revision: 4885aa143210c98657558c04aaf3dc47cfb54340
1095
- split: test
1096
  type: mteb/cqadupstack-gaming
 
 
 
1097
  metrics:
1098
  - type: main_score
1099
  value: 35.807
@@ -1377,14 +1383,14 @@ model-index:
1377
  value: 34.777
1378
  - type: recall_at_5
1379
  value: 40.527
1380
- task:
1381
  type: Retrieval
1382
- - dataset:
1383
- config: default
1384
  name: MTEB CQADupstackGisRetrieval (default)
1385
- revision: 5003b3064772da1887988e05400cf3806fe491f2
1386
- split: test
1387
  type: mteb/cqadupstack-gis
 
 
 
1388
  metrics:
1389
  - type: main_score
1390
  value: 17.629
@@ -1668,14 +1674,14 @@ model-index:
1668
  value: 16.329
1669
  - type: recall_at_5
1670
  value: 20.561
1671
- task:
1672
  type: Retrieval
1673
- - dataset:
1674
- config: default
1675
  name: MTEB CQADupstackMathematicaRetrieval (default)
1676
- revision: 90fceea13679c63fe563ded68f3b6f06e50061de
1677
- split: test
1678
  type: mteb/cqadupstack-mathematica
 
 
 
1679
  metrics:
1680
  - type: main_score
1681
  value: 11.477
@@ -1959,14 +1965,14 @@ model-index:
1959
  value: 10.395
1960
  - type: recall_at_5
1961
  value: 12.519
1962
- task:
1963
  type: Retrieval
1964
- - dataset:
1965
- config: default
1966
  name: MTEB CQADupstackPhysicsRetrieval (default)
1967
- revision: 79531abbd1fb92d06c6d6315a0cbbbf5bb247ea4
1968
- split: test
1969
  type: mteb/cqadupstack-physics
 
 
 
1970
  metrics:
1971
  - type: main_score
1972
  value: 24.339
@@ -2250,14 +2256,14 @@ model-index:
2250
  value: 22.708000000000002
2251
  - type: recall_at_5
2252
  value: 27.126
2253
- task:
2254
  type: Retrieval
2255
- - dataset:
2256
- config: default
2257
  name: MTEB CQADupstackProgrammersRetrieval (default)
2258
- revision: 6184bc1440d2dbc7612be22b50686b8826d22b32
2259
- split: test
2260
  type: mteb/cqadupstack-programmers
 
 
 
2261
  metrics:
2262
  - type: main_score
2263
  value: 18.572
@@ -2541,27 +2547,27 @@ model-index:
2541
  value: 17.283
2542
  - type: recall_at_5
2543
  value: 20.448
2544
- task:
2545
  type: Retrieval
2546
- - dataset:
2547
- config: default
2548
  name: MTEB CQADupstackRetrieval (default)
2549
- revision: CQADupstackRetrieval_is_a_combined_dataset
2550
- split: test
2551
  type: CQADupstackRetrieval_is_a_combined_dataset
 
 
 
2552
  metrics:
2553
  - type: main_score
2554
  value: 20.06633333333333
2555
  - type: ndcg_at_10
2556
  value: 20.06633333333333
2557
- task:
2558
  type: Retrieval
2559
- - dataset:
2560
- config: default
2561
  name: MTEB CQADupstackStatsRetrieval (default)
2562
- revision: 65ac3a16b8e91f9cee4c9828cc7c335575432a2a
2563
- split: test
2564
  type: mteb/cqadupstack-stats
 
 
 
2565
  metrics:
2566
  - type: main_score
2567
  value: 14.771999999999998
@@ -2845,14 +2851,14 @@ model-index:
2845
  value: 13.639000000000001
2846
  - type: recall_at_5
2847
  value: 17.197000000000003
2848
- task:
2849
  type: Retrieval
2850
- - dataset:
2851
- config: default
2852
  name: MTEB CQADupstackTexRetrieval (default)
2853
- revision: 46989137a86843e03a6195de44b09deda022eec7
2854
- split: test
2855
  type: mteb/cqadupstack-tex
 
 
 
2856
  metrics:
2857
  - type: main_score
2858
  value: 11.89
@@ -3136,14 +3142,14 @@ model-index:
3136
  value: 11.006
3137
  - type: recall_at_5
3138
  value: 13.242999999999999
3139
- task:
3140
  type: Retrieval
3141
- - dataset:
3142
- config: default
3143
  name: MTEB CQADupstackUnixRetrieval (default)
3144
- revision: 6c6430d3a6d36f8d2a829195bc5dc94d7e063e53
3145
- split: test
3146
  type: mteb/cqadupstack-unix
 
 
 
3147
  metrics:
3148
  - type: main_score
3149
  value: 17.681
@@ -3427,14 +3433,14 @@ model-index:
3427
  value: 16.892
3428
  - type: recall_at_5
3429
  value: 20.104
3430
- task:
3431
  type: Retrieval
3432
- - dataset:
3433
- config: default
3434
  name: MTEB CQADupstackWebmastersRetrieval (default)
3435
- revision: 160c094312a0e1facb97e55eeddb698c0abe3571
3436
- split: test
3437
  type: mteb/cqadupstack-webmasters
 
 
 
3438
  metrics:
3439
  - type: main_score
3440
  value: 21.083
@@ -3718,14 +3724,14 @@ model-index:
3718
  value: 18.917
3719
  - type: recall_at_5
3720
  value: 23.026
3721
- task:
3722
  type: Retrieval
3723
- - dataset:
3724
- config: default
3725
  name: MTEB CQADupstackWordpressRetrieval (default)
3726
- revision: 4ffe81d471b1924886b33c7567bfb200e9eec5c4
3727
- split: test
3728
  type: mteb/cqadupstack-wordpress
 
 
 
3729
  metrics:
3730
  - type: main_score
3731
  value: 14.978
@@ -4009,14 +4015,14 @@ model-index:
4009
  value: 13.869000000000002
4010
  - type: recall_at_5
4011
  value: 17.702
4012
- task:
4013
  type: Retrieval
4014
- - dataset:
4015
- config: default
4016
  name: MTEB ClimateFEVER (default)
4017
- revision: 47f2ac6acb640fc46020b02a5b59fdda04d39380
4018
- split: test
4019
  type: mteb/climate-fever
 
 
 
4020
  metrics:
4021
  - type: main_score
4022
  value: 14.771999999999998
@@ -4300,14 +4306,14 @@ model-index:
4300
  value: 10.41
4301
  - type: recall_at_5
4302
  value: 13.636999999999999
4303
- task:
4304
  type: Retrieval
4305
- - dataset:
4306
- config: default
4307
  name: MTEB DBPedia (default)
4308
- revision: c0f706b76e590d620bd6618b3ca8efdd34e2d659
4309
- split: test
4310
  type: mteb/dbpedia
 
 
 
4311
  metrics:
4312
  - type: main_score
4313
  value: 22.325999999999997
@@ -4591,14 +4597,14 @@ model-index:
4591
  value: 7.481999999999999
4592
  - type: recall_at_5
4593
  value: 9.869
4594
- task:
4595
- type: Retrieval
4596
- - dataset:
4597
- config: default
4598
  name: MTEB EmotionClassification (default)
4599
- revision: 4f58c6b202a23cf9a4da393831edf4f9183cad37
4600
- split: test
4601
  type: mteb/emotion
 
 
 
4602
  metrics:
4603
  - type: accuracy
4604
  value: 44.71
@@ -4608,14 +4614,14 @@ model-index:
4608
  value: 46.99336148050091
4609
  - type: main_score
4610
  value: 44.71
4611
- task:
4612
- type: Classification
4613
- - dataset:
4614
- config: default
4615
  name: MTEB FEVER (default)
4616
- revision: bea83ef9e8fb933d90a2f1d5515737465d613e12
4617
- split: test
4618
  type: mteb/fever
 
 
 
4619
  metrics:
4620
  - type: main_score
4621
  value: 26.815
@@ -4899,14 +4905,14 @@ model-index:
4899
  value: 25.728
4900
  - type: recall_at_5
4901
  value: 31.5
4902
- task:
4903
  type: Retrieval
4904
- - dataset:
4905
- config: default
4906
  name: MTEB FiQA2018 (default)
4907
- revision: 27a168819829fe9bcd655c2df245fb19452e8e06
4908
- split: test
4909
  type: mteb/fiqa
 
 
 
4910
  metrics:
4911
  - type: main_score
4912
  value: 14.184
@@ -5190,14 +5196,14 @@ model-index:
5190
  value: 10.904
5191
  - type: recall_at_5
5192
  value: 13.141
5193
- task:
5194
  type: Retrieval
5195
- - dataset:
5196
- config: default
5197
  name: MTEB HotpotQA (default)
5198
- revision: ab518f4d6fcca38d87c25209f94beba119d02014
5199
- split: test
5200
  type: mteb/hotpotqa
 
 
 
5201
  metrics:
5202
  - type: main_score
5203
  value: 31.903
@@ -5481,14 +5487,14 @@ model-index:
5481
  value: 26.516000000000002
5482
  - type: recall_at_5
5483
  value: 29.946
5484
- task:
5485
- type: Retrieval
5486
- - dataset:
5487
- config: default
5488
  name: MTEB ImdbClassification (default)
5489
- revision: 3d86128a09e091d6018b6d26cad27f2739fc2db7
5490
- split: test
5491
  type: mteb/imdb
 
 
 
5492
  metrics:
5493
  - type: accuracy
5494
  value: 70.78880000000001
@@ -5502,14 +5508,14 @@ model-index:
5502
  value: 70.58556203674722
5503
  - type: main_score
5504
  value: 70.78880000000001
5505
- task:
5506
- type: Classification
5507
- - dataset:
5508
- config: default
5509
  name: MTEB MSMARCO (default)
5510
- revision: c5a29a104738b98a9e76336939199e264163d4a0
5511
- split: test
5512
  type: mteb/msmarco
 
 
 
5513
  metrics:
5514
  - type: main_score
5515
  value: 32.823
@@ -5793,14 +5799,14 @@ model-index:
5793
  value: 2.625
5794
  - type: recall_at_5
5795
  value: 4.593
5796
- task:
5797
- type: Retrieval
5798
- - dataset:
5799
- config: en
5800
  name: MTEB MTOPDomainClassification (en)
5801
- revision: d80d48c1eb48d3562165c59d59d0034df9fff0bf
5802
- split: test
5803
  type: mteb/mtop_domain
 
 
 
5804
  metrics:
5805
  - type: accuracy
5806
  value: 85.6155950752394
@@ -5810,14 +5816,14 @@ model-index:
5810
  value: 85.62566329953071
5811
  - type: main_score
5812
  value: 85.6155950752394
5813
- task:
5814
  type: Classification
5815
- - dataset:
5816
- config: en
5817
  name: MTEB MTOPIntentClassification (en)
5818
- revision: ae001d0e6b1228650b7bd1c2c65fb50ad11a8aba
5819
- split: test
5820
  type: mteb/mtop_intent
 
 
 
5821
  metrics:
5822
  - type: accuracy
5823
  value: 51.507067943456455
@@ -5827,14 +5833,14 @@ model-index:
5827
  value: 54.70051724359958
5828
  - type: main_score
5829
  value: 51.507067943456455
5830
- task:
5831
  type: Classification
5832
- - dataset:
5833
- config: en
5834
  name: MTEB MassiveIntentClassification (en)
5835
- revision: 4672e20407010da34463acc759c162ca9734bca6
5836
- split: test
5837
  type: mteb/amazon_massive_intent
 
 
 
5838
  metrics:
5839
  - type: accuracy
5840
  value: 60.67249495628783
@@ -5844,14 +5850,14 @@ model-index:
5844
  value: 59.4792418446047
5845
  - type: main_score
5846
  value: 60.67249495628783
5847
- task:
5848
  type: Classification
5849
- - dataset:
5850
- config: en
5851
  name: MTEB MassiveScenarioClassification (en)
5852
- revision: fad2c6e8459f9e1c45d9315f4953d921437d70f8
5853
- split: test
5854
  type: mteb/amazon_massive_scenario
 
 
 
5855
  metrics:
5856
  - type: accuracy
5857
  value: 66.94687289845326
@@ -5861,14 +5867,14 @@ model-index:
5861
  value: 66.7594041012024
5862
  - type: main_score
5863
  value: 66.94687289845326
5864
- task:
5865
- type: Classification
5866
- - dataset:
5867
- config: default
5868
  name: MTEB MedrxivClusteringP2P (default)
5869
- revision: e7a26af6f3ae46b30dde8737f02c07b1505bcc73
5870
- split: test
5871
  type: mteb/medrxiv-clustering-p2p
 
 
 
5872
  metrics:
5873
  - type: main_score
5874
  value: 27.033544101894037
@@ -5876,14 +5882,14 @@ model-index:
5876
  value: 27.033544101894037
5877
  - type: v_measure_std
5878
  value: 1.3875075110570252
5879
- task:
5880
  type: Clustering
5881
- - dataset:
5882
- config: default
5883
  name: MTEB MedrxivClusteringS2S (default)
5884
- revision: 35191c8c0dca72d8ff3efcd72aa802307d469663
5885
- split: test
5886
  type: mteb/medrxiv-clustering-s2s
 
 
 
5887
  metrics:
5888
  - type: main_score
5889
  value: 23.235475122772748
@@ -5891,14 +5897,14 @@ model-index:
5891
  value: 23.235475122772748
5892
  - type: v_measure_std
5893
  value: 1.3669816155807724
5894
- task:
5895
- type: Clustering
5896
- - dataset:
5897
- config: default
5898
  name: MTEB MindSmallReranking (default)
5899
- revision: 59042f120c80e8afa9cdbb224f67076cec0fc9a7
5900
- split: test
5901
  type: mteb/mind_small
 
 
 
5902
  metrics:
5903
  - type: main_score
5904
  value: 29.025211119858856
@@ -5918,14 +5924,14 @@ model-index:
5918
  value: -18.006089017770165
5919
  - type: nAUC_mrr_std
5920
  value: -6.484973234862847
5921
- task:
5922
- type: Reranking
5923
- - dataset:
5924
- config: default
5925
  name: MTEB NFCorpus (default)
5926
- revision: ec0fa4fe99da2ff19ca1214b7966684033a58814
5927
- split: test
5928
  type: mteb/nfcorpus
 
 
 
5929
  metrics:
5930
  - type: main_score
5931
  value: 22.613
@@ -6209,14 +6215,14 @@ model-index:
6209
  value: 6.203
6210
  - type: recall_at_5
6211
  value: 7.688000000000001
6212
- task:
6213
  type: Retrieval
6214
- - dataset:
6215
- config: default
6216
  name: MTEB NQ (default)
6217
- revision: b774495ed302d8c44a3a7ea25c90dbce03968f31
6218
- split: test
6219
  type: mteb/nq
 
 
 
6220
  metrics:
6221
  - type: main_score
6222
  value: 17.163999999999998
@@ -6500,14 +6506,14 @@ model-index:
6500
  value: 15.486
6501
  - type: recall_at_5
6502
  value: 20.471
6503
- task:
6504
  type: Retrieval
6505
- - dataset:
6506
- config: default
6507
  name: MTEB QuoraRetrieval (default)
6508
- revision: e4e08e0b7dbe3c8700f0daef558ff32256715259
6509
- split: test
6510
  type: mteb/quora
 
 
 
6511
  metrics:
6512
  - type: main_score
6513
  value: 78.944
@@ -6791,14 +6797,14 @@ model-index:
6791
  value: 77.594
6792
  - type: recall_at_5
6793
  value: 82.431
6794
- task:
6795
- type: Retrieval
6796
- - dataset:
6797
- config: default
6798
  name: MTEB RedditClustering (default)
6799
- revision: 24640382cdbf8abc73003fb0fa6d111a705499eb
6800
- split: test
6801
  type: mteb/reddit-clustering
 
 
 
6802
  metrics:
6803
  - type: main_score
6804
  value: 37.24085606808143
@@ -6806,14 +6812,14 @@ model-index:
6806
  value: 37.24085606808143
6807
  - type: v_measure_std
6808
  value: 4.75871654527573
6809
- task:
6810
  type: Clustering
6811
- - dataset:
6812
- config: default
6813
  name: MTEB RedditClusteringP2P (default)
6814
- revision: 385e3cb46b4cfa89021f56c4380204149d0efe33
6815
- split: test
6816
  type: mteb/reddit-clustering-p2p
 
 
 
6817
  metrics:
6818
  - type: main_score
6819
  value: 44.53294634700315
@@ -6821,14 +6827,14 @@ model-index:
6821
  value: 44.53294634700315
6822
  - type: v_measure_std
6823
  value: 11.281270481309583
6824
- task:
6825
- type: Clustering
6826
- - dataset:
6827
- config: default
6828
  name: MTEB SCIDOCS (default)
6829
- revision: f8c2fcf00f625baaa80f62ec5bd9e1fff3b8ae88
6830
- split: test
6831
  type: mteb/scidocs
 
 
 
6832
  metrics:
6833
  - type: main_score
6834
  value: 10.975999999999999
@@ -7112,14 +7118,14 @@ model-index:
7112
  value: 6.008
7113
  - type: recall_at_5
7114
  value: 8.187999999999999
7115
- task:
7116
- type: Retrieval
7117
- - dataset:
7118
- config: default
7119
  name: MTEB SICK-R (default)
7120
- revision: 20a6d6f312dd54037fe07a32d58e5e168867909d
7121
- split: test
7122
  type: mteb/sickr-sts
 
 
 
7123
  metrics:
7124
  - type: cosine_pearson
7125
  value: 75.3765633045698
@@ -7139,14 +7145,14 @@ model-index:
7139
  value: 75.3765633045698
7140
  - type: spearman
7141
  value: 63.98613448451591
7142
- task:
7143
  type: STS
7144
- - dataset:
7145
- config: default
7146
  name: MTEB STS12 (default)
7147
- revision: a0d554a64d88156834ff5ae9920b964011b16384
7148
- split: test
7149
  type: mteb/sts12-sts
 
 
 
7150
  metrics:
7151
  - type: cosine_pearson
7152
  value: 72.15434642732643
@@ -7166,14 +7172,14 @@ model-index:
7166
  value: 72.15434642732643
7167
  - type: spearman
7168
  value: 62.37621006422606
7169
- task:
7170
  type: STS
7171
- - dataset:
7172
- config: default
7173
  name: MTEB STS13 (default)
7174
- revision: 7e90230a92c190f1bf69ae9002b8cea547a64cca
7175
- split: test
7176
  type: mteb/sts13-sts
 
 
 
7177
  metrics:
7178
  - type: cosine_pearson
7179
  value: 75.8217088748812
@@ -7193,14 +7199,14 @@ model-index:
7193
  value: 75.8217088748812
7194
  - type: spearman
7195
  value: 76.81750749576634
7196
- task:
7197
  type: STS
7198
- - dataset:
7199
- config: default
7200
  name: MTEB STS14 (default)
7201
- revision: 6031580fec1f6af667f0bd2da0a551cf4f0b2375
7202
- split: test
7203
  type: mteb/sts14-sts
 
 
 
7204
  metrics:
7205
  - type: cosine_pearson
7206
  value: 75.44045207503216
@@ -7220,14 +7226,14 @@ model-index:
7220
  value: 75.44045207503216
7221
  - type: spearman
7222
  value: 70.98200491385815
7223
- task:
7224
  type: STS
7225
- - dataset:
7226
- config: default
7227
  name: MTEB STS15 (default)
7228
- revision: ae752c7c21bf194d8b67fd573edf7ae58183cbe3
7229
- split: test
7230
  type: mteb/sts15-sts
 
 
 
7231
  metrics:
7232
  - type: cosine_pearson
7233
  value: 77.94210244961548
@@ -7247,14 +7253,14 @@ model-index:
7247
  value: 77.94210244961548
7248
  - type: spearman
7249
  value: 78.66432028882325
7250
- task:
7251
  type: STS
7252
- - dataset:
7253
- config: default
7254
  name: MTEB STS16 (default)
7255
- revision: 4d8694f8f0e0100860b497b999b3dbed754a0513
7256
- split: test
7257
  type: mteb/sts16-sts
 
 
 
7258
  metrics:
7259
  - type: cosine_pearson
7260
  value: 74.09988267613177
@@ -7274,14 +7280,14 @@ model-index:
7274
  value: 74.09988267613177
7275
  - type: spearman
7276
  value: 75.09842774068727
7277
- task:
7278
  type: STS
7279
- - dataset:
7280
- config: en-tr
7281
  name: MTEB STS17 (en-tr)
7282
- revision: faeb762787bd10488a50c8b5be4a3b82e411949c
7283
- split: test
7284
  type: mteb/sts17-crosslingual-sts
 
 
 
7285
  metrics:
7286
  - type: cosine_pearson
7287
  value: 14.399972034167586
@@ -7301,14 +7307,14 @@ model-index:
7301
  value: 14.399972034167586
7302
  - type: spearman
7303
  value: 11.974036233167949
7304
- task:
7305
  type: STS
7306
- - dataset:
7307
- config: en-ar
7308
  name: MTEB STS17 (en-ar)
7309
- revision: faeb762787bd10488a50c8b5be4a3b82e411949c
7310
- split: test
7311
  type: mteb/sts17-crosslingual-sts
 
 
 
7312
  metrics:
7313
  - type: cosine_pearson
7314
  value: 0.48735904843064204
@@ -7328,14 +7334,14 @@ model-index:
7328
  value: 0.48735904843064204
7329
  - type: spearman
7330
  value: -0.5586853869425744
7331
- task:
7332
  type: STS
7333
- - dataset:
7334
- config: en-en
7335
  name: MTEB STS17 (en-en)
7336
- revision: faeb762787bd10488a50c8b5be4a3b82e411949c
7337
- split: test
7338
  type: mteb/sts17-crosslingual-sts
 
 
 
7339
  metrics:
7340
  - type: cosine_pearson
7341
  value: 83.08779566633636
@@ -7355,14 +7361,14 @@ model-index:
7355
  value: 83.08779566633636
7356
  - type: spearman
7357
  value: 84.49159199501017
7358
- task:
7359
  type: STS
7360
- - dataset:
7361
- config: nl-en
7362
  name: MTEB STS17 (nl-en)
7363
- revision: faeb762787bd10488a50c8b5be4a3b82e411949c
7364
- split: test
7365
  type: mteb/sts17-crosslingual-sts
 
 
 
7366
  metrics:
7367
  - type: cosine_pearson
7368
  value: 28.253544786263518
@@ -7382,14 +7388,14 @@ model-index:
7382
  value: 28.253544786263518
7383
  - type: spearman
7384
  value: 24.619138389390237
7385
- task:
7386
  type: STS
7387
- - dataset:
7388
- config: fr-en
7389
  name: MTEB STS17 (fr-en)
7390
- revision: faeb762787bd10488a50c8b5be4a3b82e411949c
7391
- split: test
7392
  type: mteb/sts17-crosslingual-sts
 
 
 
7393
  metrics:
7394
  - type: cosine_pearson
7395
  value: 32.599578555181466
@@ -7409,14 +7415,14 @@ model-index:
7409
  value: 32.599578555181466
7410
  - type: spearman
7411
  value: 31.33358514527386
7412
- task:
7413
  type: STS
7414
- - dataset:
7415
- config: it-en
7416
  name: MTEB STS17 (it-en)
7417
- revision: faeb762787bd10488a50c8b5be4a3b82e411949c
7418
- split: test
7419
  type: mteb/sts17-crosslingual-sts
 
 
 
7420
  metrics:
7421
  - type: cosine_pearson
7422
  value: 23.75577872956453
@@ -7436,14 +7442,14 @@ model-index:
7436
  value: 23.75577872956453
7437
  - type: spearman
7438
  value: 21.321554136347544
7439
- task:
7440
  type: STS
7441
- - dataset:
7442
- config: en-de
7443
  name: MTEB STS17 (en-de)
7444
- revision: faeb762787bd10488a50c8b5be4a3b82e411949c
7445
- split: test
7446
  type: mteb/sts17-crosslingual-sts
 
 
 
7447
  metrics:
7448
  - type: cosine_pearson
7449
  value: 27.69563703760355
@@ -7463,14 +7469,14 @@ model-index:
7463
  value: 27.69563703760355
7464
  - type: spearman
7465
  value: 24.76336397535317
7466
- task:
7467
  type: STS
7468
- - dataset:
7469
- config: es-en
7470
  name: MTEB STS17 (es-en)
7471
- revision: faeb762787bd10488a50c8b5be4a3b82e411949c
7472
- split: test
7473
  type: mteb/sts17-crosslingual-sts
 
 
 
7474
  metrics:
7475
  - type: cosine_pearson
7476
  value: 16.346177540651276
@@ -7490,14 +7496,14 @@ model-index:
7490
  value: 16.346177540651276
7491
  - type: spearman
7492
  value: 13.601344161361006
7493
- task:
7494
  type: STS
7495
- - dataset:
7496
- config: pl-en
7497
  name: MTEB STS22 (pl-en)
7498
- revision: de9d86b3b84231dc21f76c7b7af1f28e2f57f6e3
7499
- split: test
7500
  type: mteb/sts22-crosslingual-sts
 
 
 
7501
  metrics:
7502
  - type: cosine_pearson
7503
  value: 8.721258908976726
@@ -7517,14 +7523,14 @@ model-index:
7517
  value: 8.721258908976726
7518
  - type: spearman
7519
  value: 18.35245351814682
7520
- task:
7521
  type: STS
7522
- - dataset:
7523
- config: zh-en
7524
  name: MTEB STS22 (zh-en)
7525
- revision: de9d86b3b84231dc21f76c7b7af1f28e2f57f6e3
7526
- split: test
7527
  type: mteb/sts22-crosslingual-sts
 
 
 
7528
  metrics:
7529
  - type: cosine_pearson
7530
  value: 11.49126181142697
@@ -7544,14 +7550,14 @@ model-index:
7544
  value: 11.49126181142697
7545
  - type: spearman
7546
  value: 16.21762340764813
7547
- task:
7548
  type: STS
7549
- - dataset:
7550
- config: de-en
7551
  name: MTEB STS22 (de-en)
7552
- revision: de9d86b3b84231dc21f76c7b7af1f28e2f57f6e3
7553
- split: test
7554
  type: mteb/sts22-crosslingual-sts
 
 
 
7555
  metrics:
7556
  - type: cosine_pearson
7557
  value: 26.168037910219216
@@ -7571,14 +7577,14 @@ model-index:
7571
  value: 26.168037910219216
7572
  - type: spearman
7573
  value: 31.68384203851073
7574
- task:
7575
  type: STS
7576
- - dataset:
7577
- config: es-en
7578
  name: MTEB STS22 (es-en)
7579
- revision: de9d86b3b84231dc21f76c7b7af1f28e2f57f6e3
7580
- split: test
7581
  type: mteb/sts22-crosslingual-sts
 
 
 
7582
  metrics:
7583
  - type: cosine_pearson
7584
  value: 9.071484790364082
@@ -7598,14 +7604,14 @@ model-index:
7598
  value: 9.071484790364082
7599
  - type: spearman
7600
  value: 9.726488833098573
7601
- task:
7602
  type: STS
7603
- - dataset:
7604
- config: en
7605
  name: MTEB STS22 (en)
7606
- revision: de9d86b3b84231dc21f76c7b7af1f28e2f57f6e3
7607
- split: test
7608
  type: mteb/sts22-crosslingual-sts
 
 
 
7609
  metrics:
7610
  - type: cosine_pearson
7611
  value: 58.835618895152855
@@ -7625,14 +7631,14 @@ model-index:
7625
  value: 58.835618895152855
7626
  - type: spearman
7627
  value: 63.35773898104199
7628
- task:
7629
  type: STS
7630
- - dataset:
7631
- config: default
7632
  name: MTEB STSBenchmark (default)
7633
- revision: b0fddb56ed78048fa8b90373c8a3cfc37b684831
7634
- split: test
7635
  type: mteb/stsbenchmark-sts
 
 
 
7636
  metrics:
7637
  - type: cosine_pearson
7638
  value: 75.85762627227177
@@ -7652,14 +7658,14 @@ model-index:
7652
  value: 75.85762627227177
7653
  - type: spearman
7654
  value: 74.28444387078392
7655
- task:
7656
- type: STS
7657
- - dataset:
7658
- config: default
7659
  name: MTEB SciDocsRR (default)
7660
- revision: d3c5e1fc0b855ab6097bf1cda04dd73947d7caab
7661
- split: test
7662
  type: mteb/scidocs-reranking
 
 
 
7663
  metrics:
7664
  - type: main_score
7665
  value: 70.28519419296406
@@ -7679,14 +7685,14 @@ model-index:
7679
  value: 74.99462642243834
7680
  - type: nAUC_mrr_std
7681
  value: 71.03708831031823
7682
- task:
7683
- type: Reranking
7684
- - dataset:
7685
- config: default
7686
  name: MTEB SciFact (default)
7687
- revision: 0228b52cf27578f30900b9e5271d331663a030d7
7688
- split: test
7689
  type: mteb/scifact
 
 
 
7690
  metrics:
7691
  - type: main_score
7692
  value: 46.039
@@ -7970,14 +7976,14 @@ model-index:
7970
  value: 44.806000000000004
7971
  - type: recall_at_5
7972
  value: 52.111
7973
- task:
7974
- type: Retrieval
7975
- - dataset:
7976
- config: default
7977
  name: MTEB SprintDuplicateQuestions (default)
7978
- revision: d66bd1f72af766a5cc4b0ca5e00c162f89e8cc46
7979
- split: test
7980
  type: mteb/sprintduplicatequestions-pairclassification
 
 
 
7981
  metrics:
7982
  - type: cosine_accuracy
7983
  value: 99.71188118811881
@@ -8061,14 +8067,14 @@ model-index:
8061
  value: 86.43371017471738
8062
  - type: similarity_recall
8063
  value: 84.1
8064
- task:
8065
- type: PairClassification
8066
- - dataset:
8067
- config: default
8068
  name: MTEB StackExchangeClustering (default)
8069
- revision: 6cbc1f7b2bc0622f2e39d2c77fa502909748c259
8070
- split: test
8071
  type: mteb/stackexchange-clustering
 
 
 
8072
  metrics:
8073
  - type: main_score
8074
  value: 45.28623588344632
@@ -8076,14 +8082,14 @@ model-index:
8076
  value: 45.28623588344632
8077
  - type: v_measure_std
8078
  value: 3.699756512710844
8079
- task:
8080
  type: Clustering
8081
- - dataset:
8082
- config: default
8083
  name: MTEB StackExchangeClusteringP2P (default)
8084
- revision: 815ca46b2622cec33ccafc3735d572c266efdb44
8085
- split: test
8086
  type: mteb/stackexchange-clustering-p2p
 
 
 
8087
  metrics:
8088
  - type: main_score
8089
  value: 31.36188637863386
@@ -8091,14 +8097,14 @@ model-index:
8091
  value: 31.36188637863386
8092
  - type: v_measure_std
8093
  value: 1.479842367330655
8094
- task:
8095
- type: Clustering
8096
- - dataset:
8097
- config: default
8098
  name: MTEB StackOverflowDupQuestions (default)
8099
- revision: e185fbe320c72810689fc5848eb6114e1ef5ec69
8100
- split: test
8101
  type: mteb/stackoverflowdupquestions-reranking
 
 
 
8102
  metrics:
8103
  - type: main_score
8104
  value: 41.90071199374392
@@ -8118,14 +8124,14 @@ model-index:
8118
  value: 14.254825331290972
8119
  - type: nAUC_mrr_std
8120
  value: 4.529615444769097
8121
- task:
8122
- type: Reranking
8123
- - dataset:
8124
- config: default
8125
  name: MTEB SummEval (default)
8126
- revision: cda12ad7615edc362dbf25a00fdd61d3b1eaf93c
8127
- split: test
8128
  type: mteb/summeval
 
 
 
8129
  metrics:
8130
  - type: cosine_pearson
8131
  value: 30.390849690222034
@@ -8141,14 +8147,14 @@ model-index:
8141
  value: 30.390849690222034
8142
  - type: spearman
8143
  value: 28.891662569282467
8144
- task:
8145
- type: Summarization
8146
- - dataset:
8147
- config: default
8148
  name: MTEB TRECCOVID (default)
8149
- revision: bb9466bac8153a0349341eb1b22e06409e78ef4e
8150
- split: test
8151
  type: mteb/trec-covid
 
 
 
8152
  metrics:
8153
  - type: main_score
8154
  value: 43.236000000000004
@@ -8432,14 +8438,14 @@ model-index:
8432
  value: 0.38
8433
  - type: recall_at_5
8434
  value: 0.598
8435
- task:
8436
  type: Retrieval
8437
- - dataset:
8438
- config: default
8439
  name: MTEB Touche2020 (default)
8440
- revision: a34f9a33db75fa0cbb21bb5cfc3dae8dc8bec93f
8441
- split: test
8442
  type: mteb/touche2020
 
 
 
8443
  metrics:
8444
  - type: main_score
8445
  value: 15.369
@@ -8723,14 +8729,14 @@ model-index:
8723
  value: 3.7379999999999995
8724
  - type: recall_at_5
8725
  value: 6.2170000000000005
8726
- task:
8727
- type: Retrieval
8728
- - dataset:
8729
- config: default
8730
  name: MTEB ToxicConversationsClassification (default)
8731
- revision: edfaf9da55d3dd50d43143d90c1ac476895ae6de
8732
- split: test
8733
  type: mteb/toxic_conversations_50k
 
 
 
8734
  metrics:
8735
  - type: accuracy
8736
  value: 67.724609375
@@ -8744,14 +8750,14 @@ model-index:
8744
  value: 74.90031680337547
8745
  - type: main_score
8746
  value: 67.724609375
8747
- task:
8748
  type: Classification
8749
- - dataset:
8750
- config: default
8751
  name: MTEB TweetSentimentExtractionClassification (default)
8752
- revision: d604517c81ca91fe16a244d1248fc021f9ecee7a
8753
- split: test
8754
  type: mteb/tweet_sentiment_extraction
 
 
 
8755
  metrics:
8756
  - type: accuracy
8757
  value: 54.20486700622524
@@ -8761,14 +8767,14 @@ model-index:
8761
  value: 53.770651858268025
8762
  - type: main_score
8763
  value: 54.20486700622524
8764
- task:
8765
- type: Classification
8766
- - dataset:
8767
- config: default
8768
  name: MTEB TwentyNewsgroupsClustering (default)
8769
- revision: 6125ec4e24fa026cec8a478383ee943acfbd5449
8770
- split: test
8771
  type: mteb/twentynewsgroups-clustering
 
 
 
8772
  metrics:
8773
  - type: main_score
8774
  value: 31.416458059499856
@@ -8776,14 +8782,14 @@ model-index:
8776
  value: 31.416458059499856
8777
  - type: v_measure_std
8778
  value: 1.7923550883286548
8779
- task:
8780
- type: Clustering
8781
- - dataset:
8782
- config: default
8783
  name: MTEB TwitterSemEval2015 (default)
8784
- revision: 70970daeab8776df92f5ea462b6173c0b46fd2d1
8785
- split: test
8786
  type: mteb/twittersemeval2015-pairclassification
 
 
 
8787
  metrics:
8788
  - type: cosine_accuracy
8789
  value: 81.09912380044108
@@ -8867,14 +8873,14 @@ model-index:
8867
  value: 47.37899176541474
8868
  - type: similarity_recall
8869
  value: 62.24274406332454
8870
- task:
8871
  type: PairClassification
8872
- - dataset:
8873
- config: default
8874
  name: MTEB TwitterURLCorpus (default)
8875
- revision: 8b6510b0b1fa4e4c4f879467980e9be563ec1cdf
8876
- split: test
8877
  type: mteb/twitterurlcorpus-pairclassification
 
 
 
8878
  metrics:
8879
  - type: cosine_accuracy
8880
  value: 86.66123336049986
@@ -8958,13 +8964,6 @@ model-index:
8958
  value: 69.54226574980899
8959
  - type: similarity_recall
8960
  value: 77.08654142285187
8961
- task:
8962
- type: PairClassification
8963
- model_name: potion-base-4M
8964
- tags:
8965
- - embeddings
8966
- - static-embeddings
8967
- - mteb
8968
  ---
8969
 
8970
  # potion-base-4M Model Card
 
1
  ---
2
  library_name: model2vec
3
  license: mit
4
+ tags:
5
+ - embeddings
6
+ - static-embeddings
7
+ - mteb
8
  model-index:
9
  - name: potion-base-4M
10
  results:
11
+ - task:
12
+ type: Classification
13
+ dataset:
14
  name: MTEB AmazonCounterfactualClassification (en-ext)
 
 
15
  type: mteb/amazon_counterfactual
16
+ config: en-ext
17
+ split: test
18
+ revision: e8379541af4e31359cca9fbcf4b00f2671dba205
19
  metrics:
20
  - type: accuracy
21
  value: 67.5712143928036
 
29
  value: 73.78539599566238
30
  - type: main_score
31
  value: 67.5712143928036
32
+ - task:
33
  type: Classification
34
+ dataset:
 
35
  name: MTEB AmazonCounterfactualClassification (en)
 
 
36
  type: mteb/amazon_counterfactual
37
+ config: en
38
+ split: test
39
+ revision: e8379541af4e31359cca9fbcf4b00f2671dba205
40
  metrics:
41
  - type: accuracy
42
  value: 68.38805970149254
 
50
  value: 71.3687043183968
51
  - type: main_score
52
  value: 68.38805970149254
53
+ - task:
54
  type: Classification
55
+ dataset:
 
56
  name: MTEB AmazonPolarityClassification (default)
 
 
57
  type: mteb/amazon_polarity
58
+ config: default
59
+ split: test
60
+ revision: e2d317d38cd51312af73b3d32a06d1a08b442046
61
  metrics:
62
  - type: accuracy
63
  value: 71.69602499999999
 
71
  value: 71.42572451950106
72
  - type: main_score
73
  value: 71.69602499999999
74
+ - task:
75
  type: Classification
76
+ dataset:
 
77
  name: MTEB AmazonReviewsClassification (en)
 
 
78
  type: mteb/amazon_reviews_multi
79
+ config: en
80
+ split: test
81
+ revision: 1399c76144fd37290681b995c656ef9b2e06e26d
82
  metrics:
83
  - type: accuracy
84
  value: 34.321999999999996
 
88
  value: 33.68591037434848
89
  - type: main_score
90
  value: 34.321999999999996
91
+ - task:
92
+ type: Retrieval
93
+ dataset:
 
94
  name: MTEB ArguAna (default)
 
 
95
  type: mteb/arguana
96
+ config: default
97
+ split: test
98
+ revision: c22ab2a51041ffd869aaddef7af8d8215647e41a
99
  metrics:
100
  - type: main_score
101
  value: 39.449
 
379
  value: 39.26
380
  - type: recall_at_5
381
  value: 49.431000000000004
382
+ - task:
383
+ type: Clustering
384
+ dataset:
 
385
  name: MTEB ArxivClusteringP2P (default)
 
 
386
  type: mteb/arxiv-clustering-p2p
387
+ config: default
388
+ split: test
389
+ revision: a122ad7f3f0291bf49cc6f4d32aa80929df69d5d
390
  metrics:
391
  - type: main_score
392
  value: 34.19332453253303
 
394
  value: 34.19332453253303
395
  - type: v_measure_std
396
  value: 14.189152053888938
397
+ - task:
398
  type: Clustering
399
+ dataset:
 
400
  name: MTEB ArxivClusteringS2S (default)
 
 
401
  type: mteb/arxiv-clustering-s2s
402
+ config: default
403
+ split: test
404
+ revision: f910caf1a6075f7329cdf8c1a6135696f37dbd53
405
  metrics:
406
  - type: main_score
407
  value: 24.104443846867934
 
409
  value: 24.104443846867934
410
  - type: v_measure_std
411
  value: 15.164667270049211
412
+ - task:
413
+ type: Reranking
414
+ dataset:
 
415
  name: MTEB AskUbuntuDupQuestions (default)
 
 
416
  type: mteb/askubuntudupquestions-reranking
417
+ config: default
418
+ split: test
419
+ revision: 2000358ca161889fa9c082cb41daa8dcfb161a54
420
  metrics:
421
  - type: main_score
422
  value: 53.803562861722696
 
436
  value: 21.221447117685774
437
  - type: nAUC_mrr_std
438
  value: 5.470983299654257
439
+ - task:
440
+ type: STS
441
+ dataset:
 
442
  name: MTEB BIOSSES (default)
 
 
443
  type: mteb/biosses-sts
444
+ config: default
445
+ split: test
446
+ revision: d3fb88f8f02e40887cd149695127462bbcf29b4a
447
  metrics:
448
  - type: cosine_pearson
449
  value: 75.18540989714172
 
463
  value: 75.18540989714172
464
  - type: spearman
465
  value: 71.87392685280749
466
+ - task:
467
+ type: Classification
468
+ dataset:
 
469
  name: MTEB Banking77Classification (default)
 
 
470
  type: mteb/banking77
471
+ config: default
472
+ split: test
473
+ revision: 0fd18e25b25c072e09e0d92ab615fda904d66300
474
  metrics:
475
  - type: accuracy
476
  value: 69.69480519480518
 
480
  value: 68.4951487448326
481
  - type: main_score
482
  value: 69.69480519480518
483
+ - task:
484
+ type: Clustering
485
+ dataset:
 
486
  name: MTEB BiorxivClusteringP2P (default)
 
 
487
  type: mteb/biorxiv-clustering-p2p
488
+ config: default
489
+ split: test
490
+ revision: 65b79d1d13f80053f67aca9498d9402c2d9f1f40
491
  metrics:
492
  - type: main_score
493
  value: 29.09840292638104
 
495
  value: 29.09840292638104
496
  - type: v_measure_std
497
  value: 0.8130265845858159
498
+ - task:
499
  type: Clustering
500
+ dataset:
 
501
  name: MTEB BiorxivClusteringS2S (default)
 
 
502
  type: mteb/biorxiv-clustering-s2s
503
+ config: default
504
+ split: test
505
+ revision: 258694dd0231531bc1fd9de6ceb52a0853c6d908
506
  metrics:
507
  - type: main_score
508
  value: 18.622216350398414
 
510
  value: 18.622216350398414
511
  - type: v_measure_std
512
  value: 0.6293204014994399
513
+ - task:
514
+ type: Retrieval
515
+ dataset:
 
516
  name: MTEB CQADupstackAndroidRetrieval (default)
 
 
517
  type: mteb/cqadupstack-android
518
+ config: default
519
+ split: test
520
+ revision: f46a197baaae43b4f621051089b82a364682dfeb
521
  metrics:
522
  - type: main_score
523
  value: 28.634999999999998
 
801
  value: 26.143
802
  - type: recall_at_5
803
  value: 30.441000000000003
804
+ - task:
805
  type: Retrieval
806
+ dataset:
 
807
  name: MTEB CQADupstackEnglishRetrieval (default)
 
 
808
  type: mteb/cqadupstack-english
809
+ config: default
810
+ split: test
811
+ revision: ad9991cb51e31e31e430383c75ffb2885547b5f0
812
  metrics:
813
  - type: main_score
814
  value: 23.933
 
1092
  value: 22.034000000000002
1093
  - type: recall_at_5
1094
  value: 25.663000000000004
1095
+ - task:
1096
  type: Retrieval
1097
+ dataset:
 
1098
  name: MTEB CQADupstackGamingRetrieval (default)
 
 
1099
  type: mteb/cqadupstack-gaming
1100
+ config: default
1101
+ split: test
1102
+ revision: 4885aa143210c98657558c04aaf3dc47cfb54340
1103
  metrics:
1104
  - type: main_score
1105
  value: 35.807
 
1383
  value: 34.777
1384
  - type: recall_at_5
1385
  value: 40.527
1386
+ - task:
1387
  type: Retrieval
1388
+ dataset:
 
1389
  name: MTEB CQADupstackGisRetrieval (default)
 
 
1390
  type: mteb/cqadupstack-gis
1391
+ config: default
1392
+ split: test
1393
+ revision: 5003b3064772da1887988e05400cf3806fe491f2
1394
  metrics:
1395
  - type: main_score
1396
  value: 17.629
 
1674
  value: 16.329
1675
  - type: recall_at_5
1676
  value: 20.561
1677
+ - task:
1678
  type: Retrieval
1679
+ dataset:
 
1680
  name: MTEB CQADupstackMathematicaRetrieval (default)
 
 
1681
  type: mteb/cqadupstack-mathematica
1682
+ config: default
1683
+ split: test
1684
+ revision: 90fceea13679c63fe563ded68f3b6f06e50061de
1685
  metrics:
1686
  - type: main_score
1687
  value: 11.477
 
1965
  value: 10.395
1966
  - type: recall_at_5
1967
  value: 12.519
1968
+ - task:
1969
  type: Retrieval
1970
+ dataset:
 
1971
  name: MTEB CQADupstackPhysicsRetrieval (default)
 
 
1972
  type: mteb/cqadupstack-physics
1973
+ config: default
1974
+ split: test
1975
+ revision: 79531abbd1fb92d06c6d6315a0cbbbf5bb247ea4
1976
  metrics:
1977
  - type: main_score
1978
  value: 24.339
 
2256
  value: 22.708000000000002
2257
  - type: recall_at_5
2258
  value: 27.126
2259
+ - task:
2260
  type: Retrieval
2261
+ dataset:
 
2262
  name: MTEB CQADupstackProgrammersRetrieval (default)
 
 
2263
  type: mteb/cqadupstack-programmers
2264
+ config: default
2265
+ split: test
2266
+ revision: 6184bc1440d2dbc7612be22b50686b8826d22b32
2267
  metrics:
2268
  - type: main_score
2269
  value: 18.572
 
2547
  value: 17.283
2548
  - type: recall_at_5
2549
  value: 20.448
2550
+ - task:
2551
  type: Retrieval
2552
+ dataset:
 
2553
  name: MTEB CQADupstackRetrieval (default)
 
 
2554
  type: CQADupstackRetrieval_is_a_combined_dataset
2555
+ config: default
2556
+ split: test
2557
+ revision: CQADupstackRetrieval_is_a_combined_dataset
2558
  metrics:
2559
  - type: main_score
2560
  value: 20.06633333333333
2561
  - type: ndcg_at_10
2562
  value: 20.06633333333333
2563
+ - task:
2564
  type: Retrieval
2565
+ dataset:
 
2566
  name: MTEB CQADupstackStatsRetrieval (default)
 
 
2567
  type: mteb/cqadupstack-stats
2568
+ config: default
2569
+ split: test
2570
+ revision: 65ac3a16b8e91f9cee4c9828cc7c335575432a2a
2571
  metrics:
2572
  - type: main_score
2573
  value: 14.771999999999998
 
2851
  value: 13.639000000000001
2852
  - type: recall_at_5
2853
  value: 17.197000000000003
2854
+ - task:
2855
  type: Retrieval
2856
+ dataset:
 
2857
  name: MTEB CQADupstackTexRetrieval (default)
 
 
2858
  type: mteb/cqadupstack-tex
2859
+ config: default
2860
+ split: test
2861
+ revision: 46989137a86843e03a6195de44b09deda022eec7
2862
  metrics:
2863
  - type: main_score
2864
  value: 11.89
 
3142
  value: 11.006
3143
  - type: recall_at_5
3144
  value: 13.242999999999999
3145
+ - task:
3146
  type: Retrieval
3147
+ dataset:
 
3148
  name: MTEB CQADupstackUnixRetrieval (default)
 
 
3149
  type: mteb/cqadupstack-unix
3150
+ config: default
3151
+ split: test
3152
+ revision: 6c6430d3a6d36f8d2a829195bc5dc94d7e063e53
3153
  metrics:
3154
  - type: main_score
3155
  value: 17.681
 
3433
  value: 16.892
3434
  - type: recall_at_5
3435
  value: 20.104
3436
+ - task:
3437
  type: Retrieval
3438
+ dataset:
 
3439
  name: MTEB CQADupstackWebmastersRetrieval (default)
 
 
3440
  type: mteb/cqadupstack-webmasters
3441
+ config: default
3442
+ split: test
3443
+ revision: 160c094312a0e1facb97e55eeddb698c0abe3571
3444
  metrics:
3445
  - type: main_score
3446
  value: 21.083
 
3724
  value: 18.917
3725
  - type: recall_at_5
3726
  value: 23.026
3727
+ - task:
3728
  type: Retrieval
3729
+ dataset:
 
3730
  name: MTEB CQADupstackWordpressRetrieval (default)
 
 
3731
  type: mteb/cqadupstack-wordpress
3732
+ config: default
3733
+ split: test
3734
+ revision: 4ffe81d471b1924886b33c7567bfb200e9eec5c4
3735
  metrics:
3736
  - type: main_score
3737
  value: 14.978
 
4015
  value: 13.869000000000002
4016
  - type: recall_at_5
4017
  value: 17.702
4018
+ - task:
4019
  type: Retrieval
4020
+ dataset:
 
4021
  name: MTEB ClimateFEVER (default)
 
 
4022
  type: mteb/climate-fever
4023
+ config: default
4024
+ split: test
4025
+ revision: 47f2ac6acb640fc46020b02a5b59fdda04d39380
4026
  metrics:
4027
  - type: main_score
4028
  value: 14.771999999999998
 
4306
  value: 10.41
4307
  - type: recall_at_5
4308
  value: 13.636999999999999
4309
+ - task:
4310
  type: Retrieval
4311
+ dataset:
 
4312
  name: MTEB DBPedia (default)
 
 
4313
  type: mteb/dbpedia
4314
+ config: default
4315
+ split: test
4316
+ revision: c0f706b76e590d620bd6618b3ca8efdd34e2d659
4317
  metrics:
4318
  - type: main_score
4319
  value: 22.325999999999997
 
4597
  value: 7.481999999999999
4598
  - type: recall_at_5
4599
  value: 9.869
4600
+ - task:
4601
+ type: Classification
4602
+ dataset:
 
4603
  name: MTEB EmotionClassification (default)
 
 
4604
  type: mteb/emotion
4605
+ config: default
4606
+ split: test
4607
+ revision: 4f58c6b202a23cf9a4da393831edf4f9183cad37
4608
  metrics:
4609
  - type: accuracy
4610
  value: 44.71
 
4614
  value: 46.99336148050091
4615
  - type: main_score
4616
  value: 44.71
4617
+ - task:
4618
+ type: Retrieval
4619
+ dataset:
 
4620
  name: MTEB FEVER (default)
 
 
4621
  type: mteb/fever
4622
+ config: default
4623
+ split: test
4624
+ revision: bea83ef9e8fb933d90a2f1d5515737465d613e12
4625
  metrics:
4626
  - type: main_score
4627
  value: 26.815
 
4905
  value: 25.728
4906
  - type: recall_at_5
4907
  value: 31.5
4908
+ - task:
4909
  type: Retrieval
4910
+ dataset:
 
4911
  name: MTEB FiQA2018 (default)
 
 
4912
  type: mteb/fiqa
4913
+ config: default
4914
+ split: test
4915
+ revision: 27a168819829fe9bcd655c2df245fb19452e8e06
4916
  metrics:
4917
  - type: main_score
4918
  value: 14.184
 
5196
  value: 10.904
5197
  - type: recall_at_5
5198
  value: 13.141
5199
+ - task:
5200
  type: Retrieval
5201
+ dataset:
 
5202
  name: MTEB HotpotQA (default)
 
 
5203
  type: mteb/hotpotqa
5204
+ config: default
5205
+ split: test
5206
+ revision: ab518f4d6fcca38d87c25209f94beba119d02014
5207
  metrics:
5208
  - type: main_score
5209
  value: 31.903
 
5487
  value: 26.516000000000002
5488
  - type: recall_at_5
5489
  value: 29.946
5490
+ - task:
5491
+ type: Classification
5492
+ dataset:
 
5493
  name: MTEB ImdbClassification (default)
 
 
5494
  type: mteb/imdb
5495
+ config: default
5496
+ split: test
5497
+ revision: 3d86128a09e091d6018b6d26cad27f2739fc2db7
5498
  metrics:
5499
  - type: accuracy
5500
  value: 70.78880000000001
 
5508
  value: 70.58556203674722
5509
  - type: main_score
5510
  value: 70.78880000000001
5511
+ - task:
5512
+ type: Retrieval
5513
+ dataset:
 
5514
  name: MTEB MSMARCO (default)
 
 
5515
  type: mteb/msmarco
5516
+ config: default
5517
+ split: test
5518
+ revision: c5a29a104738b98a9e76336939199e264163d4a0
5519
  metrics:
5520
  - type: main_score
5521
  value: 32.823
 
5799
  value: 2.625
5800
  - type: recall_at_5
5801
  value: 4.593
5802
+ - task:
5803
+ type: Classification
5804
+ dataset:
 
5805
  name: MTEB MTOPDomainClassification (en)
 
 
5806
  type: mteb/mtop_domain
5807
+ config: en
5808
+ split: test
5809
+ revision: d80d48c1eb48d3562165c59d59d0034df9fff0bf
5810
  metrics:
5811
  - type: accuracy
5812
  value: 85.6155950752394
 
5816
  value: 85.62566329953071
5817
  - type: main_score
5818
  value: 85.6155950752394
5819
+ - task:
5820
  type: Classification
5821
+ dataset:
 
5822
  name: MTEB MTOPIntentClassification (en)
 
 
5823
  type: mteb/mtop_intent
5824
+ config: en
5825
+ split: test
5826
+ revision: ae001d0e6b1228650b7bd1c2c65fb50ad11a8aba
5827
  metrics:
5828
  - type: accuracy
5829
  value: 51.507067943456455
 
5833
  value: 54.70051724359958
5834
  - type: main_score
5835
  value: 51.507067943456455
5836
+ - task:
5837
  type: Classification
5838
+ dataset:
 
5839
  name: MTEB MassiveIntentClassification (en)
 
 
5840
  type: mteb/amazon_massive_intent
5841
+ config: en
5842
+ split: test
5843
+ revision: 4672e20407010da34463acc759c162ca9734bca6
5844
  metrics:
5845
  - type: accuracy
5846
  value: 60.67249495628783
 
5850
  value: 59.4792418446047
5851
  - type: main_score
5852
  value: 60.67249495628783
5853
+ - task:
5854
  type: Classification
5855
+ dataset:
 
5856
  name: MTEB MassiveScenarioClassification (en)
 
 
5857
  type: mteb/amazon_massive_scenario
5858
+ config: en
5859
+ split: test
5860
+ revision: fad2c6e8459f9e1c45d9315f4953d921437d70f8
5861
  metrics:
5862
  - type: accuracy
5863
  value: 66.94687289845326
 
5867
  value: 66.7594041012024
5868
  - type: main_score
5869
  value: 66.94687289845326
5870
+ - task:
5871
+ type: Clustering
5872
+ dataset:
 
5873
  name: MTEB MedrxivClusteringP2P (default)
 
 
5874
  type: mteb/medrxiv-clustering-p2p
5875
+ config: default
5876
+ split: test
5877
+ revision: e7a26af6f3ae46b30dde8737f02c07b1505bcc73
5878
  metrics:
5879
  - type: main_score
5880
  value: 27.033544101894037
 
5882
  value: 27.033544101894037
5883
  - type: v_measure_std
5884
  value: 1.3875075110570252
5885
+ - task:
5886
  type: Clustering
5887
+ dataset:
 
5888
  name: MTEB MedrxivClusteringS2S (default)
 
 
5889
  type: mteb/medrxiv-clustering-s2s
5890
+ config: default
5891
+ split: test
5892
+ revision: 35191c8c0dca72d8ff3efcd72aa802307d469663
5893
  metrics:
5894
  - type: main_score
5895
  value: 23.235475122772748
 
5897
  value: 23.235475122772748
5898
  - type: v_measure_std
5899
  value: 1.3669816155807724
5900
+ - task:
5901
+ type: Reranking
5902
+ dataset:
 
5903
  name: MTEB MindSmallReranking (default)
 
 
5904
  type: mteb/mind_small
5905
+ config: default
5906
+ split: test
5907
+ revision: 59042f120c80e8afa9cdbb224f67076cec0fc9a7
5908
  metrics:
5909
  - type: main_score
5910
  value: 29.025211119858856
 
5924
  value: -18.006089017770165
5925
  - type: nAUC_mrr_std
5926
  value: -6.484973234862847
5927
+ - task:
5928
+ type: Retrieval
5929
+ dataset:
 
5930
  name: MTEB NFCorpus (default)
 
 
5931
  type: mteb/nfcorpus
5932
+ config: default
5933
+ split: test
5934
+ revision: ec0fa4fe99da2ff19ca1214b7966684033a58814
5935
  metrics:
5936
  - type: main_score
5937
  value: 22.613
 
6215
  value: 6.203
6216
  - type: recall_at_5
6217
  value: 7.688000000000001
6218
+ - task:
6219
  type: Retrieval
6220
+ dataset:
 
6221
  name: MTEB NQ (default)
 
 
6222
  type: mteb/nq
6223
+ config: default
6224
+ split: test
6225
+ revision: b774495ed302d8c44a3a7ea25c90dbce03968f31
6226
  metrics:
6227
  - type: main_score
6228
  value: 17.163999999999998
 
6506
  value: 15.486
6507
  - type: recall_at_5
6508
  value: 20.471
6509
+ - task:
6510
  type: Retrieval
6511
+ dataset:
 
6512
  name: MTEB QuoraRetrieval (default)
 
 
6513
  type: mteb/quora
6514
+ config: default
6515
+ split: test
6516
+ revision: e4e08e0b7dbe3c8700f0daef558ff32256715259
6517
  metrics:
6518
  - type: main_score
6519
  value: 78.944
 
6797
  value: 77.594
6798
  - type: recall_at_5
6799
  value: 82.431
6800
+ - task:
6801
+ type: Clustering
6802
+ dataset:
 
6803
  name: MTEB RedditClustering (default)
 
 
6804
  type: mteb/reddit-clustering
6805
+ config: default
6806
+ split: test
6807
+ revision: 24640382cdbf8abc73003fb0fa6d111a705499eb
6808
  metrics:
6809
  - type: main_score
6810
  value: 37.24085606808143
 
6812
  value: 37.24085606808143
6813
  - type: v_measure_std
6814
  value: 4.75871654527573
6815
+ - task:
6816
  type: Clustering
6817
+ dataset:
 
6818
  name: MTEB RedditClusteringP2P (default)
 
 
6819
  type: mteb/reddit-clustering-p2p
6820
+ config: default
6821
+ split: test
6822
+ revision: 385e3cb46b4cfa89021f56c4380204149d0efe33
6823
  metrics:
6824
  - type: main_score
6825
  value: 44.53294634700315
 
6827
  value: 44.53294634700315
6828
  - type: v_measure_std
6829
  value: 11.281270481309583
6830
+ - task:
6831
+ type: Retrieval
6832
+ dataset:
 
6833
  name: MTEB SCIDOCS (default)
 
 
6834
  type: mteb/scidocs
6835
+ config: default
6836
+ split: test
6837
+ revision: f8c2fcf00f625baaa80f62ec5bd9e1fff3b8ae88
6838
  metrics:
6839
  - type: main_score
6840
  value: 10.975999999999999
 
7118
  value: 6.008
7119
  - type: recall_at_5
7120
  value: 8.187999999999999
7121
+ - task:
7122
+ type: STS
7123
+ dataset:
 
7124
  name: MTEB SICK-R (default)
 
 
7125
  type: mteb/sickr-sts
7126
+ config: default
7127
+ split: test
7128
+ revision: 20a6d6f312dd54037fe07a32d58e5e168867909d
7129
  metrics:
7130
  - type: cosine_pearson
7131
  value: 75.3765633045698
 
7145
  value: 75.3765633045698
7146
  - type: spearman
7147
  value: 63.98613448451591
7148
+ - task:
7149
  type: STS
7150
+ dataset:
 
7151
  name: MTEB STS12 (default)
 
 
7152
  type: mteb/sts12-sts
7153
+ config: default
7154
+ split: test
7155
+ revision: a0d554a64d88156834ff5ae9920b964011b16384
7156
  metrics:
7157
  - type: cosine_pearson
7158
  value: 72.15434642732643
 
7172
  value: 72.15434642732643
7173
  - type: spearman
7174
  value: 62.37621006422606
7175
+ - task:
7176
  type: STS
7177
+ dataset:
 
7178
  name: MTEB STS13 (default)
 
 
7179
  type: mteb/sts13-sts
7180
+ config: default
7181
+ split: test
7182
+ revision: 7e90230a92c190f1bf69ae9002b8cea547a64cca
7183
  metrics:
7184
  - type: cosine_pearson
7185
  value: 75.8217088748812
 
7199
  value: 75.8217088748812
7200
  - type: spearman
7201
  value: 76.81750749576634
7202
+ - task:
7203
  type: STS
7204
+ dataset:
 
7205
  name: MTEB STS14 (default)
 
 
7206
  type: mteb/sts14-sts
7207
+ config: default
7208
+ split: test
7209
+ revision: 6031580fec1f6af667f0bd2da0a551cf4f0b2375
7210
  metrics:
7211
  - type: cosine_pearson
7212
  value: 75.44045207503216
 
7226
  value: 75.44045207503216
7227
  - type: spearman
7228
  value: 70.98200491385815
7229
+ - task:
7230
  type: STS
7231
+ dataset:
 
7232
  name: MTEB STS15 (default)
 
 
7233
  type: mteb/sts15-sts
7234
+ config: default
7235
+ split: test
7236
+ revision: ae752c7c21bf194d8b67fd573edf7ae58183cbe3
7237
  metrics:
7238
  - type: cosine_pearson
7239
  value: 77.94210244961548
 
7253
  value: 77.94210244961548
7254
  - type: spearman
7255
  value: 78.66432028882325
7256
+ - task:
7257
  type: STS
7258
+ dataset:
 
7259
  name: MTEB STS16 (default)
 
 
7260
  type: mteb/sts16-sts
7261
+ config: default
7262
+ split: test
7263
+ revision: 4d8694f8f0e0100860b497b999b3dbed754a0513
7264
  metrics:
7265
  - type: cosine_pearson
7266
  value: 74.09988267613177
 
7280
  value: 74.09988267613177
7281
  - type: spearman
7282
  value: 75.09842774068727
7283
+ - task:
7284
  type: STS
7285
+ dataset:
 
7286
  name: MTEB STS17 (en-tr)
 
 
7287
  type: mteb/sts17-crosslingual-sts
7288
+ config: en-tr
7289
+ split: test
7290
+ revision: faeb762787bd10488a50c8b5be4a3b82e411949c
7291
  metrics:
7292
  - type: cosine_pearson
7293
  value: 14.399972034167586
 
7307
  value: 14.399972034167586
7308
  - type: spearman
7309
  value: 11.974036233167949
7310
+ - task:
7311
  type: STS
7312
+ dataset:
 
7313
  name: MTEB STS17 (en-ar)
 
 
7314
  type: mteb/sts17-crosslingual-sts
7315
+ config: en-ar
7316
+ split: test
7317
+ revision: faeb762787bd10488a50c8b5be4a3b82e411949c
7318
  metrics:
7319
  - type: cosine_pearson
7320
  value: 0.48735904843064204
 
7334
  value: 0.48735904843064204
7335
  - type: spearman
7336
  value: -0.5586853869425744
7337
+ - task:
7338
  type: STS
7339
+ dataset:
 
7340
  name: MTEB STS17 (en-en)
 
 
7341
  type: mteb/sts17-crosslingual-sts
7342
+ config: en-en
7343
+ split: test
7344
+ revision: faeb762787bd10488a50c8b5be4a3b82e411949c
7345
  metrics:
7346
  - type: cosine_pearson
7347
  value: 83.08779566633636
 
7361
  value: 83.08779566633636
7362
  - type: spearman
7363
  value: 84.49159199501017
7364
+ - task:
7365
  type: STS
7366
+ dataset:
 
7367
  name: MTEB STS17 (nl-en)
 
 
7368
  type: mteb/sts17-crosslingual-sts
7369
+ config: nl-en
7370
+ split: test
7371
+ revision: faeb762787bd10488a50c8b5be4a3b82e411949c
7372
  metrics:
7373
  - type: cosine_pearson
7374
  value: 28.253544786263518
 
7388
  value: 28.253544786263518
7389
  - type: spearman
7390
  value: 24.619138389390237
7391
+ - task:
7392
  type: STS
7393
+ dataset:
 
7394
  name: MTEB STS17 (fr-en)
 
 
7395
  type: mteb/sts17-crosslingual-sts
7396
+ config: fr-en
7397
+ split: test
7398
+ revision: faeb762787bd10488a50c8b5be4a3b82e411949c
7399
  metrics:
7400
  - type: cosine_pearson
7401
  value: 32.599578555181466
 
7415
  value: 32.599578555181466
7416
  - type: spearman
7417
  value: 31.33358514527386
7418
+ - task:
7419
  type: STS
7420
+ dataset:
 
7421
  name: MTEB STS17 (it-en)
 
 
7422
  type: mteb/sts17-crosslingual-sts
7423
+ config: it-en
7424
+ split: test
7425
+ revision: faeb762787bd10488a50c8b5be4a3b82e411949c
7426
  metrics:
7427
  - type: cosine_pearson
7428
  value: 23.75577872956453
 
7442
  value: 23.75577872956453
7443
  - type: spearman
7444
  value: 21.321554136347544
7445
+ - task:
7446
  type: STS
7447
+ dataset:
 
7448
  name: MTEB STS17 (en-de)
 
 
7449
  type: mteb/sts17-crosslingual-sts
7450
+ config: en-de
7451
+ split: test
7452
+ revision: faeb762787bd10488a50c8b5be4a3b82e411949c
7453
  metrics:
7454
  - type: cosine_pearson
7455
  value: 27.69563703760355
 
7469
  value: 27.69563703760355
7470
  - type: spearman
7471
  value: 24.76336397535317
7472
+ - task:
7473
  type: STS
7474
+ dataset:
 
7475
  name: MTEB STS17 (es-en)
 
 
7476
  type: mteb/sts17-crosslingual-sts
7477
+ config: es-en
7478
+ split: test
7479
+ revision: faeb762787bd10488a50c8b5be4a3b82e411949c
7480
  metrics:
7481
  - type: cosine_pearson
7482
  value: 16.346177540651276
 
7496
  value: 16.346177540651276
7497
  - type: spearman
7498
  value: 13.601344161361006
7499
+ - task:
7500
  type: STS
7501
+ dataset:
 
7502
  name: MTEB STS22 (pl-en)
 
 
7503
  type: mteb/sts22-crosslingual-sts
7504
+ config: pl-en
7505
+ split: test
7506
+ revision: de9d86b3b84231dc21f76c7b7af1f28e2f57f6e3
7507
  metrics:
7508
  - type: cosine_pearson
7509
  value: 8.721258908976726
 
7523
  value: 8.721258908976726
7524
  - type: spearman
7525
  value: 18.35245351814682
7526
+ - task:
7527
  type: STS
7528
+ dataset:
 
7529
  name: MTEB STS22 (zh-en)
 
 
7530
  type: mteb/sts22-crosslingual-sts
7531
+ config: zh-en
7532
+ split: test
7533
+ revision: de9d86b3b84231dc21f76c7b7af1f28e2f57f6e3
7534
  metrics:
7535
  - type: cosine_pearson
7536
  value: 11.49126181142697
 
7550
  value: 11.49126181142697
7551
  - type: spearman
7552
  value: 16.21762340764813
7553
+ - task:
7554
  type: STS
7555
+ dataset:
 
7556
  name: MTEB STS22 (de-en)
 
 
7557
  type: mteb/sts22-crosslingual-sts
7558
+ config: de-en
7559
+ split: test
7560
+ revision: de9d86b3b84231dc21f76c7b7af1f28e2f57f6e3
7561
  metrics:
7562
  - type: cosine_pearson
7563
  value: 26.168037910219216
 
7577
  value: 26.168037910219216
7578
  - type: spearman
7579
  value: 31.68384203851073
7580
+ - task:
7581
  type: STS
7582
+ dataset:
 
7583
  name: MTEB STS22 (es-en)
 
 
7584
  type: mteb/sts22-crosslingual-sts
7585
+ config: es-en
7586
+ split: test
7587
+ revision: de9d86b3b84231dc21f76c7b7af1f28e2f57f6e3
7588
  metrics:
7589
  - type: cosine_pearson
7590
  value: 9.071484790364082
 
7604
  value: 9.071484790364082
7605
  - type: spearman
7606
  value: 9.726488833098573
7607
+ - task:
7608
  type: STS
7609
+ dataset:
 
7610
  name: MTEB STS22 (en)
 
 
7611
  type: mteb/sts22-crosslingual-sts
7612
+ config: en
7613
+ split: test
7614
+ revision: de9d86b3b84231dc21f76c7b7af1f28e2f57f6e3
7615
  metrics:
7616
  - type: cosine_pearson
7617
  value: 58.835618895152855
 
7631
  value: 58.835618895152855
7632
  - type: spearman
7633
  value: 63.35773898104199
7634
+ - task:
7635
  type: STS
7636
+ dataset:
 
7637
  name: MTEB STSBenchmark (default)
 
 
7638
  type: mteb/stsbenchmark-sts
7639
+ config: default
7640
+ split: test
7641
+ revision: b0fddb56ed78048fa8b90373c8a3cfc37b684831
7642
  metrics:
7643
  - type: cosine_pearson
7644
  value: 75.85762627227177
 
7658
  value: 75.85762627227177
7659
  - type: spearman
7660
  value: 74.28444387078392
7661
+ - task:
7662
+ type: Reranking
7663
+ dataset:
 
7664
  name: MTEB SciDocsRR (default)
 
 
7665
  type: mteb/scidocs-reranking
7666
+ config: default
7667
+ split: test
7668
+ revision: d3c5e1fc0b855ab6097bf1cda04dd73947d7caab
7669
  metrics:
7670
  - type: main_score
7671
  value: 70.28519419296406
 
7685
  value: 74.99462642243834
7686
  - type: nAUC_mrr_std
7687
  value: 71.03708831031823
7688
+ - task:
7689
+ type: Retrieval
7690
+ dataset:
 
7691
  name: MTEB SciFact (default)
 
 
7692
  type: mteb/scifact
7693
+ config: default
7694
+ split: test
7695
+ revision: 0228b52cf27578f30900b9e5271d331663a030d7
7696
  metrics:
7697
  - type: main_score
7698
  value: 46.039
 
7976
  value: 44.806000000000004
7977
  - type: recall_at_5
7978
  value: 52.111
7979
+ - task:
7980
+ type: PairClassification
7981
+ dataset:
 
7982
  name: MTEB SprintDuplicateQuestions (default)
 
 
7983
  type: mteb/sprintduplicatequestions-pairclassification
7984
+ config: default
7985
+ split: test
7986
+ revision: d66bd1f72af766a5cc4b0ca5e00c162f89e8cc46
7987
  metrics:
7988
  - type: cosine_accuracy
7989
  value: 99.71188118811881
 
8067
  value: 86.43371017471738
8068
  - type: similarity_recall
8069
  value: 84.1
8070
+ - task:
8071
+ type: Clustering
8072
+ dataset:
 
8073
  name: MTEB StackExchangeClustering (default)
 
 
8074
  type: mteb/stackexchange-clustering
8075
+ config: default
8076
+ split: test
8077
+ revision: 6cbc1f7b2bc0622f2e39d2c77fa502909748c259
8078
  metrics:
8079
  - type: main_score
8080
  value: 45.28623588344632
 
8082
  value: 45.28623588344632
8083
  - type: v_measure_std
8084
  value: 3.699756512710844
8085
+ - task:
8086
  type: Clustering
8087
+ dataset:
 
8088
  name: MTEB StackExchangeClusteringP2P (default)
 
 
8089
  type: mteb/stackexchange-clustering-p2p
8090
+ config: default
8091
+ split: test
8092
+ revision: 815ca46b2622cec33ccafc3735d572c266efdb44
8093
  metrics:
8094
  - type: main_score
8095
  value: 31.36188637863386
 
8097
  value: 31.36188637863386
8098
  - type: v_measure_std
8099
  value: 1.479842367330655
8100
+ - task:
8101
+ type: Reranking
8102
+ dataset:
 
8103
  name: MTEB StackOverflowDupQuestions (default)
 
 
8104
  type: mteb/stackoverflowdupquestions-reranking
8105
+ config: default
8106
+ split: test
8107
+ revision: e185fbe320c72810689fc5848eb6114e1ef5ec69
8108
  metrics:
8109
  - type: main_score
8110
  value: 41.90071199374392
 
8124
  value: 14.254825331290972
8125
  - type: nAUC_mrr_std
8126
  value: 4.529615444769097
8127
+ - task:
8128
+ type: Summarization
8129
+ dataset:
 
8130
  name: MTEB SummEval (default)
 
 
8131
  type: mteb/summeval
8132
+ config: default
8133
+ split: test
8134
+ revision: cda12ad7615edc362dbf25a00fdd61d3b1eaf93c
8135
  metrics:
8136
  - type: cosine_pearson
8137
  value: 30.390849690222034
 
8147
  value: 30.390849690222034
8148
  - type: spearman
8149
  value: 28.891662569282467
8150
+ - task:
8151
+ type: Retrieval
8152
+ dataset:
 
8153
  name: MTEB TRECCOVID (default)
 
 
8154
  type: mteb/trec-covid
8155
+ config: default
8156
+ split: test
8157
+ revision: bb9466bac8153a0349341eb1b22e06409e78ef4e
8158
  metrics:
8159
  - type: main_score
8160
  value: 43.236000000000004
 
8438
  value: 0.38
8439
  - type: recall_at_5
8440
  value: 0.598
8441
+ - task:
8442
  type: Retrieval
8443
+ dataset:
 
8444
  name: MTEB Touche2020 (default)
 
 
8445
  type: mteb/touche2020
8446
+ config: default
8447
+ split: test
8448
+ revision: a34f9a33db75fa0cbb21bb5cfc3dae8dc8bec93f
8449
  metrics:
8450
  - type: main_score
8451
  value: 15.369
 
8729
  value: 3.7379999999999995
8730
  - type: recall_at_5
8731
  value: 6.2170000000000005
8732
+ - task:
8733
+ type: Classification
8734
+ dataset:
 
8735
  name: MTEB ToxicConversationsClassification (default)
 
 
8736
  type: mteb/toxic_conversations_50k
8737
+ config: default
8738
+ split: test
8739
+ revision: edfaf9da55d3dd50d43143d90c1ac476895ae6de
8740
  metrics:
8741
  - type: accuracy
8742
  value: 67.724609375
 
8750
  value: 74.90031680337547
8751
  - type: main_score
8752
  value: 67.724609375
8753
+ - task:
8754
  type: Classification
8755
+ dataset:
 
8756
  name: MTEB TweetSentimentExtractionClassification (default)
 
 
8757
  type: mteb/tweet_sentiment_extraction
8758
+ config: default
8759
+ split: test
8760
+ revision: d604517c81ca91fe16a244d1248fc021f9ecee7a
8761
  metrics:
8762
  - type: accuracy
8763
  value: 54.20486700622524
 
8767
  value: 53.770651858268025
8768
  - type: main_score
8769
  value: 54.20486700622524
8770
+ - task:
8771
+ type: Clustering
8772
+ dataset:
 
8773
  name: MTEB TwentyNewsgroupsClustering (default)
 
 
8774
  type: mteb/twentynewsgroups-clustering
8775
+ config: default
8776
+ split: test
8777
+ revision: 6125ec4e24fa026cec8a478383ee943acfbd5449
8778
  metrics:
8779
  - type: main_score
8780
  value: 31.416458059499856
 
8782
  value: 31.416458059499856
8783
  - type: v_measure_std
8784
  value: 1.7923550883286548
8785
+ - task:
8786
+ type: PairClassification
8787
+ dataset:
 
8788
  name: MTEB TwitterSemEval2015 (default)
 
 
8789
  type: mteb/twittersemeval2015-pairclassification
8790
+ config: default
8791
+ split: test
8792
+ revision: 70970daeab8776df92f5ea462b6173c0b46fd2d1
8793
  metrics:
8794
  - type: cosine_accuracy
8795
  value: 81.09912380044108
 
8873
  value: 47.37899176541474
8874
  - type: similarity_recall
8875
  value: 62.24274406332454
8876
+ - task:
8877
  type: PairClassification
8878
+ dataset:
 
8879
  name: MTEB TwitterURLCorpus (default)
 
 
8880
  type: mteb/twitterurlcorpus-pairclassification
8881
+ config: default
8882
+ split: test
8883
+ revision: 8b6510b0b1fa4e4c4f879467980e9be563ec1cdf
8884
  metrics:
8885
  - type: cosine_accuracy
8886
  value: 86.66123336049986
 
8964
  value: 69.54226574980899
8965
  - type: similarity_recall
8966
  value: 77.08654142285187
 
 
 
 
 
 
 
8967
  ---
8968
 
8969
  # potion-base-4M Model Card