Pringled commited on
Commit
b0a2f11
·
verified ·
1 Parent(s): c33f20d

Upload folder using huggingface_hub

Browse files
Files changed (1) hide show
  1. README.md +439 -433
README.md CHANGED
@@ -1,21 +1,15 @@
1
  ---
2
  library_name: model2vec
3
  license: mit
4
- tags:
5
- - embeddings
6
- - static-embeddings
7
- - mteb
8
  model-index:
9
  - name: potion-base-8M
10
  results:
11
- - task:
12
- type: Classification
13
- dataset:
14
- name: MTEB AmazonCounterfactualClassification (en-ext)
15
- type: mteb/amazon_counterfactual
16
  config: en-ext
17
- split: test
18
  revision: e8379541af4e31359cca9fbcf4b00f2671dba205
 
 
19
  metrics:
20
  - type: accuracy
21
  value: 72.15142428785607
@@ -29,14 +23,14 @@ model-index:
29
  value: 77.34906471545477
30
  - type: main_score
31
  value: 72.15142428785607
32
- - task:
33
  type: Classification
34
- dataset:
35
- name: MTEB AmazonCounterfactualClassification (en)
36
- type: mteb/amazon_counterfactual
37
  config: en
38
- split: test
39
  revision: e8379541af4e31359cca9fbcf4b00f2671dba205
 
 
40
  metrics:
41
  - type: accuracy
42
  value: 71.7910447761194
@@ -50,14 +44,14 @@ model-index:
50
  value: 74.32788084269461
51
  - type: main_score
52
  value: 71.7910447761194
53
- - task:
54
  type: Classification
55
- dataset:
56
- name: MTEB AmazonPolarityClassification (default)
57
- type: mteb/amazon_polarity
58
  config: default
59
- split: test
60
  revision: e2d317d38cd51312af73b3d32a06d1a08b442046
 
 
61
  metrics:
62
  - type: accuracy
63
  value: 72.47644999999999
@@ -71,14 +65,14 @@ model-index:
71
  value: 72.2600863044581
72
  - type: main_score
73
  value: 72.47644999999999
74
- - task:
75
  type: Classification
76
- dataset:
77
- name: MTEB AmazonReviewsClassification (en)
78
- type: mteb/amazon_reviews_multi
79
  config: en
80
- split: test
81
  revision: 1399c76144fd37290681b995c656ef9b2e06e26d
 
 
82
  metrics:
83
  - type: accuracy
84
  value: 36.012
@@ -88,14 +82,14 @@ model-index:
88
  value: 35.38209336470206
89
  - type: main_score
90
  value: 36.012
91
- - task:
92
- type: Retrieval
93
- dataset:
94
- name: MTEB ArguAna (default)
95
- type: mteb/arguana
96
  config: default
97
- split: test
98
  revision: c22ab2a51041ffd869aaddef7af8d8215647e41a
 
 
99
  metrics:
100
  - type: main_score
101
  value: 41.966
@@ -379,14 +373,14 @@ model-index:
379
  value: 42.888
380
  - type: recall_at_5
381
  value: 52.063
382
- - task:
383
- type: Clustering
384
- dataset:
385
- name: MTEB ArxivClusteringP2P (default)
386
- type: mteb/arxiv-clustering-p2p
387
  config: default
388
- split: test
389
  revision: a122ad7f3f0291bf49cc6f4d32aa80929df69d5d
 
 
390
  metrics:
391
  - type: main_score
392
  value: 35.387660145946825
@@ -394,14 +388,14 @@ model-index:
394
  value: 35.387660145946825
395
  - type: v_measure_std
396
  value: 14.022525689022785
397
- - task:
398
  type: Clustering
399
- dataset:
400
- name: MTEB ArxivClusteringS2S (default)
401
- type: mteb/arxiv-clustering-s2s
402
  config: default
403
- split: test
404
  revision: f910caf1a6075f7329cdf8c1a6135696f37dbd53
 
 
405
  metrics:
406
  - type: main_score
407
  value: 25.26058942964131
@@ -409,14 +403,14 @@ model-index:
409
  value: 25.26058942964131
410
  - type: v_measure_std
411
  value: 14.850432186356857
412
- - task:
413
- type: Reranking
414
- dataset:
415
- name: MTEB AskUbuntuDupQuestions (default)
416
- type: mteb/askubuntudupquestions-reranking
417
  config: default
418
- split: test
419
  revision: 2000358ca161889fa9c082cb41daa8dcfb161a54
 
 
420
  metrics:
421
  - type: main_score
422
  value: 54.13950871400633
@@ -436,14 +430,14 @@ model-index:
436
  value: 21.861692980537956
437
  - type: nAUC_mrr_std
438
  value: 8.256966784037171
439
- - task:
440
- type: STS
441
- dataset:
442
- name: MTEB BIOSSES (default)
443
- type: mteb/biosses-sts
444
  config: default
445
- split: test
446
  revision: d3fb88f8f02e40887cd149695127462bbcf29b4a
 
 
447
  metrics:
448
  - type: cosine_pearson
449
  value: 79.11612010879227
@@ -463,14 +457,14 @@ model-index:
463
  value: 79.11612010879227
464
  - type: spearman
465
  value: 75.85775256673794
466
- - task:
467
- type: Classification
468
- dataset:
469
- name: MTEB Banking77Classification (default)
470
- type: mteb/banking77
471
  config: default
472
- split: test
473
  revision: 0fd18e25b25c072e09e0d92ab615fda904d66300
 
 
474
  metrics:
475
  - type: accuracy
476
  value: 72.63636363636363
@@ -480,14 +474,14 @@ model-index:
480
  value: 71.69751597573539
481
  - type: main_score
482
  value: 72.63636363636363
483
- - task:
484
- type: Clustering
485
- dataset:
486
- name: MTEB BiorxivClusteringP2P (default)
487
- type: mteb/biorxiv-clustering-p2p
488
  config: default
489
- split: test
490
  revision: 65b79d1d13f80053f67aca9498d9402c2d9f1f40
 
 
491
  metrics:
492
  - type: main_score
493
  value: 30.861840536151014
@@ -495,14 +489,14 @@ model-index:
495
  value: 30.861840536151014
496
  - type: v_measure_std
497
  value: 0.8096483751274005
498
- - task:
499
  type: Clustering
500
- dataset:
501
- name: MTEB BiorxivClusteringS2S (default)
502
- type: mteb/biorxiv-clustering-s2s
503
  config: default
504
- split: test
505
  revision: 258694dd0231531bc1fd9de6ceb52a0853c6d908
 
 
506
  metrics:
507
  - type: main_score
508
  value: 20.219544420664455
@@ -510,14 +504,14 @@ model-index:
510
  value: 20.219544420664455
511
  - type: v_measure_std
512
  value: 0.7431903039116942
513
- - task:
514
- type: Retrieval
515
- dataset:
516
- name: MTEB CQADupstackAndroidRetrieval (default)
517
- type: mteb/cqadupstack-android
518
  config: default
519
- split: test
520
  revision: f46a197baaae43b4f621051089b82a364682dfeb
 
 
521
  metrics:
522
  - type: main_score
523
  value: 31.835
@@ -801,14 +795,14 @@ model-index:
801
  value: 29.301
802
  - type: recall_at_5
803
  value: 33.918
804
- - task:
805
  type: Retrieval
806
- dataset:
807
- name: MTEB CQADupstackEnglishRetrieval (default)
808
- type: mteb/cqadupstack-english
809
  config: default
810
- split: test
811
  revision: ad9991cb51e31e31e430383c75ffb2885547b5f0
 
 
812
  metrics:
813
  - type: main_score
814
  value: 25.734
@@ -1092,14 +1086,14 @@ model-index:
1092
  value: 23.54
1093
  - type: recall_at_5
1094
  value: 28.054000000000002
1095
- - task:
1096
  type: Retrieval
1097
- dataset:
1098
- name: MTEB CQADupstackGamingRetrieval (default)
1099
- type: mteb/cqadupstack-gaming
1100
  config: default
1101
- split: test
1102
  revision: 4885aa143210c98657558c04aaf3dc47cfb54340
 
 
1103
  metrics:
1104
  - type: main_score
1105
  value: 39.235
@@ -1383,14 +1377,14 @@ model-index:
1383
  value: 37.830999999999996
1384
  - type: recall_at_5
1385
  value: 43.811
1386
- - task:
1387
  type: Retrieval
1388
- dataset:
1389
- name: MTEB CQADupstackGisRetrieval (default)
1390
- type: mteb/cqadupstack-gis
1391
  config: default
1392
- split: test
1393
  revision: 5003b3064772da1887988e05400cf3806fe491f2
 
 
1394
  metrics:
1395
  - type: main_score
1396
  value: 19.197
@@ -1674,14 +1668,14 @@ model-index:
1674
  value: 19.417
1675
  - type: recall_at_5
1676
  value: 23.577
1677
- - task:
1678
  type: Retrieval
1679
- dataset:
1680
- name: MTEB CQADupstackMathematicaRetrieval (default)
1681
- type: mteb/cqadupstack-mathematica
1682
  config: default
1683
- split: test
1684
  revision: 90fceea13679c63fe563ded68f3b6f06e50061de
 
 
1685
  metrics:
1686
  - type: main_score
1687
  value: 12.145
@@ -1965,14 +1959,14 @@ model-index:
1965
  value: 11.178
1966
  - type: recall_at_5
1967
  value: 13.877
1968
- - task:
1969
  type: Retrieval
1970
- dataset:
1971
- name: MTEB CQADupstackPhysicsRetrieval (default)
1972
- type: mteb/cqadupstack-physics
1973
  config: default
1974
- split: test
1975
  revision: 79531abbd1fb92d06c6d6315a0cbbbf5bb247ea4
 
 
1976
  metrics:
1977
  - type: main_score
1978
  value: 26.667999999999996
@@ -2256,14 +2250,14 @@ model-index:
2256
  value: 24.524
2257
  - type: recall_at_5
2258
  value: 28.787000000000003
2259
- - task:
2260
  type: Retrieval
2261
- dataset:
2262
- name: MTEB CQADupstackProgrammersRetrieval (default)
2263
- type: mteb/cqadupstack-programmers
2264
  config: default
2265
- split: test
2266
  revision: 6184bc1440d2dbc7612be22b50686b8826d22b32
 
 
2267
  metrics:
2268
  - type: main_score
2269
  value: 21.66
@@ -2547,27 +2541,27 @@ model-index:
2547
  value: 19.79
2548
  - type: recall_at_5
2549
  value: 23.376
2550
- - task:
2551
  type: Retrieval
2552
- dataset:
2553
- name: MTEB CQADupstackRetrieval (default)
2554
- type: CQADupstackRetrieval_is_a_combined_dataset
2555
  config: default
2556
- split: test
2557
  revision: CQADupstackRetrieval_is_a_combined_dataset
 
 
2558
  metrics:
2559
  - type: main_score
2560
  value: 22.302333333333333
2561
  - type: ndcg_at_10
2562
  value: 22.302333333333333
2563
- - task:
2564
  type: Retrieval
2565
- dataset:
2566
- name: MTEB CQADupstackStatsRetrieval (default)
2567
- type: mteb/cqadupstack-stats
2568
  config: default
2569
- split: test
2570
  revision: 65ac3a16b8e91f9cee4c9828cc7c335575432a2a
 
 
2571
  metrics:
2572
  - type: main_score
2573
  value: 17.253
@@ -2851,14 +2845,14 @@ model-index:
2851
  value: 17.134
2852
  - type: recall_at_5
2853
  value: 19.72
2854
- - task:
2855
  type: Retrieval
2856
- dataset:
2857
- name: MTEB CQADupstackTexRetrieval (default)
2858
- type: mteb/cqadupstack-tex
2859
  config: default
2860
- split: test
2861
  revision: 46989137a86843e03a6195de44b09deda022eec7
 
 
2862
  metrics:
2863
  - type: main_score
2864
  value: 13.308
@@ -3142,14 +3136,14 @@ model-index:
3142
  value: 12.199
3143
  - type: recall_at_5
3144
  value: 14.87
3145
- - task:
3146
  type: Retrieval
3147
- dataset:
3148
- name: MTEB CQADupstackUnixRetrieval (default)
3149
- type: mteb/cqadupstack-unix
3150
  config: default
3151
- split: test
3152
  revision: 6c6430d3a6d36f8d2a829195bc5dc94d7e063e53
 
 
3153
  metrics:
3154
  - type: main_score
3155
  value: 20.061999999999998
@@ -3433,14 +3427,14 @@ model-index:
3433
  value: 18.859
3434
  - type: recall_at_5
3435
  value: 22.435
3436
- - task:
3437
  type: Retrieval
3438
- dataset:
3439
- name: MTEB CQADupstackWebmastersRetrieval (default)
3440
- type: mteb/cqadupstack-webmasters
3441
  config: default
3442
- split: test
3443
  revision: 160c094312a0e1facb97e55eeddb698c0abe3571
 
 
3444
  metrics:
3445
  - type: main_score
3446
  value: 23.735999999999997
@@ -3724,14 +3718,14 @@ model-index:
3724
  value: 21.397
3725
  - type: recall_at_5
3726
  value: 24.863
3727
- - task:
3728
  type: Retrieval
3729
- dataset:
3730
- name: MTEB CQADupstackWordpressRetrieval (default)
3731
- type: mteb/cqadupstack-wordpress
3732
  config: default
3733
- split: test
3734
  revision: 4ffe81d471b1924886b33c7567bfb200e9eec5c4
 
 
3735
  metrics:
3736
  - type: main_score
3737
  value: 16.794999999999998
@@ -4015,14 +4009,14 @@ model-index:
4015
  value: 15.440999999999999
4016
  - type: recall_at_5
4017
  value: 18.829
4018
- - task:
4019
  type: Retrieval
4020
- dataset:
4021
- name: MTEB ClimateFEVER (default)
4022
- type: mteb/climate-fever
4023
  config: default
4024
- split: test
4025
  revision: 47f2ac6acb640fc46020b02a5b59fdda04d39380
 
 
4026
  metrics:
4027
  - type: main_score
4028
  value: 17.288999999999998
@@ -4306,14 +4300,14 @@ model-index:
4306
  value: 12.232
4307
  - type: recall_at_5
4308
  value: 15.540999999999999
4309
- - task:
4310
  type: Retrieval
4311
- dataset:
4312
- name: MTEB DBPedia (default)
4313
- type: mteb/dbpedia
4314
  config: default
4315
- split: test
4316
  revision: c0f706b76e590d620bd6618b3ca8efdd34e2d659
 
 
4317
  metrics:
4318
  - type: main_score
4319
  value: 24.235
@@ -4597,14 +4591,14 @@ model-index:
4597
  value: 8.698
4598
  - type: recall_at_5
4599
  value: 11.052
4600
- - task:
4601
- type: Classification
4602
- dataset:
4603
- name: MTEB EmotionClassification (default)
4604
- type: mteb/emotion
4605
  config: default
4606
- split: test
4607
  revision: 4f58c6b202a23cf9a4da393831edf4f9183cad37
 
 
4608
  metrics:
4609
  - type: accuracy
4610
  value: 45.09
@@ -4614,14 +4608,14 @@ model-index:
4614
  value: 47.129694558751545
4615
  - type: main_score
4616
  value: 45.09
4617
- - task:
4618
- type: Retrieval
4619
- dataset:
4620
- name: MTEB FEVER (default)
4621
- type: mteb/fever
4622
  config: default
4623
- split: test
4624
  revision: bea83ef9e8fb933d90a2f1d5515737465d613e12
 
 
4625
  metrics:
4626
  - type: main_score
4627
  value: 30.267
@@ -4905,14 +4899,14 @@ model-index:
4905
  value: 29.42
4906
  - type: recall_at_5
4907
  value: 36.112
4908
- - task:
4909
  type: Retrieval
4910
- dataset:
4911
- name: MTEB FiQA2018 (default)
4912
- type: mteb/fiqa
4913
  config: default
4914
- split: test
4915
  revision: 27a168819829fe9bcd655c2df245fb19452e8e06
 
 
4916
  metrics:
4917
  - type: main_score
4918
  value: 16.619
@@ -5196,14 +5190,14 @@ model-index:
5196
  value: 12.485
5197
  - type: recall_at_5
5198
  value: 15.862000000000002
5199
- - task:
5200
  type: Retrieval
5201
- dataset:
5202
- name: MTEB HotpotQA (default)
5203
- type: mteb/hotpotqa
5204
  config: default
5205
- split: test
5206
  revision: ab518f4d6fcca38d87c25209f94beba119d02014
 
 
5207
  metrics:
5208
  - type: main_score
5209
  value: 36.217
@@ -5487,14 +5481,14 @@ model-index:
5487
  value: 30.311
5488
  - type: recall_at_5
5489
  value: 34.092
5490
- - task:
5491
- type: Classification
5492
- dataset:
5493
- name: MTEB ImdbClassification (default)
5494
- type: mteb/imdb
5495
  config: default
5496
- split: test
5497
  revision: 3d86128a09e091d6018b6d26cad27f2739fc2db7
 
 
5498
  metrics:
5499
  - type: accuracy
5500
  value: 70.78
@@ -5508,14 +5502,14 @@ model-index:
5508
  value: 70.56498155979408
5509
  - type: main_score
5510
  value: 70.78
5511
- - task:
5512
- type: Retrieval
5513
- dataset:
5514
- name: MTEB MSMARCO (default)
5515
- type: mteb/msmarco
5516
  config: default
5517
- split: test
5518
  revision: c5a29a104738b98a9e76336939199e264163d4a0
 
 
5519
  metrics:
5520
  - type: main_score
5521
  value: 34.981
@@ -5799,14 +5793,14 @@ model-index:
5799
  value: 2.8240000000000003
5800
  - type: recall_at_5
5801
  value: 4.476999999999999
5802
- - task:
5803
- type: Classification
5804
- dataset:
5805
- name: MTEB MTOPDomainClassification (en)
5806
- type: mteb/mtop_domain
5807
  config: en
5808
- split: test
5809
  revision: d80d48c1eb48d3562165c59d59d0034df9fff0bf
 
 
5810
  metrics:
5811
  - type: accuracy
5812
  value: 88.95576835385319
@@ -5816,14 +5810,14 @@ model-index:
5816
  value: 89.00721562093213
5817
  - type: main_score
5818
  value: 88.95576835385319
5819
- - task:
5820
  type: Classification
5821
- dataset:
5822
- name: MTEB MTOPIntentClassification (en)
5823
- type: mteb/mtop_intent
5824
  config: en
5825
- split: test
5826
  revision: ae001d0e6b1228650b7bd1c2c65fb50ad11a8aba
 
 
5827
  metrics:
5828
  - type: accuracy
5829
  value: 56.99726402188783
@@ -5833,14 +5827,14 @@ model-index:
5833
  value: 59.96788951671549
5834
  - type: main_score
5835
  value: 56.99726402188783
5836
- - task:
5837
  type: Classification
5838
- dataset:
5839
- name: MTEB MassiveIntentClassification (en)
5840
- type: mteb/amazon_massive_intent
5841
  config: en
5842
- split: test
5843
  revision: 4672e20407010da34463acc759c162ca9734bca6
 
 
5844
  metrics:
5845
  - type: accuracy
5846
  value: 63.79287155346336
@@ -5850,14 +5844,14 @@ model-index:
5850
  value: 62.567311481126055
5851
  - type: main_score
5852
  value: 63.79287155346336
5853
- - task:
5854
  type: Classification
5855
- dataset:
5856
- name: MTEB MassiveScenarioClassification (en)
5857
- type: mteb/amazon_massive_scenario
5858
  config: en
5859
- split: test
5860
  revision: fad2c6e8459f9e1c45d9315f4953d921437d70f8
 
 
5861
  metrics:
5862
  - type: accuracy
5863
  value: 70.30934767989241
@@ -5867,14 +5861,14 @@ model-index:
5867
  value: 70.1128179307388
5868
  - type: main_score
5869
  value: 70.30934767989241
5870
- - task:
5871
- type: Clustering
5872
- dataset:
 
5873
  name: MTEB MedrxivClusteringP2P (default)
 
 
5874
  type: mteb/medrxiv-clustering-p2p
5875
- config: default
5876
- split: test
5877
- revision: e7a26af6f3ae46b30dde8737f02c07b1505bcc73
5878
  metrics:
5879
  - type: main_score
5880
  value: 27.61734940907637
@@ -5882,14 +5876,14 @@ model-index:
5882
  value: 27.61734940907637
5883
  - type: v_measure_std
5884
  value: 1.2248100208316097
5885
- - task:
5886
  type: Clustering
5887
- dataset:
5888
- name: MTEB MedrxivClusteringS2S (default)
5889
- type: mteb/medrxiv-clustering-s2s
5890
  config: default
5891
- split: test
5892
  revision: 35191c8c0dca72d8ff3efcd72aa802307d469663
 
 
5893
  metrics:
5894
  - type: main_score
5895
  value: 23.802943866708308
@@ -5897,14 +5891,14 @@ model-index:
5897
  value: 23.802943866708308
5898
  - type: v_measure_std
5899
  value: 1.4975518910969763
5900
- - task:
5901
- type: Reranking
5902
- dataset:
5903
- name: MTEB MindSmallReranking (default)
5904
- type: mteb/mind_small
5905
  config: default
5906
- split: test
5907
  revision: 59042f120c80e8afa9cdbb224f67076cec0fc9a7
 
 
5908
  metrics:
5909
  - type: main_score
5910
  value: 29.431722284942175
@@ -5924,14 +5918,14 @@ model-index:
5924
  value: -17.42779158552557
5925
  - type: nAUC_mrr_std
5926
  value: -5.997215692334967
5927
- - task:
5928
- type: Retrieval
5929
- dataset:
5930
- name: MTEB NFCorpus (default)
5931
- type: mteb/nfcorpus
5932
  config: default
5933
- split: test
5934
  revision: ec0fa4fe99da2ff19ca1214b7966684033a58814
 
 
5935
  metrics:
5936
  - type: main_score
5937
  value: 24.267
@@ -6215,14 +6209,14 @@ model-index:
6215
  value: 7.236
6216
  - type: recall_at_5
6217
  value: 9.003
6218
- - task:
6219
  type: Retrieval
6220
- dataset:
6221
- name: MTEB NQ (default)
6222
- type: mteb/nq
6223
  config: default
6224
- split: test
6225
  revision: b774495ed302d8c44a3a7ea25c90dbce03968f31
 
 
6226
  metrics:
6227
  - type: main_score
6228
  value: 19.543
@@ -6506,14 +6500,14 @@ model-index:
6506
  value: 17.669
6507
  - type: recall_at_5
6508
  value: 23.488999999999997
6509
- - task:
6510
  type: Retrieval
6511
- dataset:
6512
- name: MTEB QuoraRetrieval (default)
6513
- type: mteb/quora
6514
  config: default
6515
- split: test
6516
  revision: e4e08e0b7dbe3c8700f0daef558ff32256715259
 
 
6517
  metrics:
6518
  - type: main_score
6519
  value: 80.598
@@ -6797,14 +6791,14 @@ model-index:
6797
  value: 79.074
6798
  - type: recall_at_5
6799
  value: 84.143
6800
- - task:
6801
- type: Clustering
6802
- dataset:
6803
- name: MTEB RedditClustering (default)
6804
- type: mteb/reddit-clustering
6805
  config: default
6806
- split: test
6807
  revision: 24640382cdbf8abc73003fb0fa6d111a705499eb
 
 
6808
  metrics:
6809
  - type: main_score
6810
  value: 39.5465127563479
@@ -6812,14 +6806,14 @@ model-index:
6812
  value: 39.5465127563479
6813
  - type: v_measure_std
6814
  value: 5.038703300031419
6815
- - task:
6816
  type: Clustering
6817
- dataset:
6818
- name: MTEB RedditClusteringP2P (default)
6819
- type: mteb/reddit-clustering-p2p
6820
  config: default
6821
- split: test
6822
  revision: 385e3cb46b4cfa89021f56c4380204149d0efe33
 
 
6823
  metrics:
6824
  - type: main_score
6825
  value: 47.07911795189491
@@ -6827,14 +6821,14 @@ model-index:
6827
  value: 47.07911795189491
6828
  - type: v_measure_std
6829
  value: 11.546436135362846
6830
- - task:
6831
- type: Retrieval
6832
- dataset:
6833
- name: MTEB SCIDOCS (default)
6834
- type: mteb/scidocs
6835
  config: default
6836
- split: test
6837
  revision: f8c2fcf00f625baaa80f62ec5bd9e1fff3b8ae88
 
 
6838
  metrics:
6839
  - type: main_score
6840
  value: 12.386999999999999
@@ -7118,14 +7112,14 @@ model-index:
7118
  value: 6.553000000000001
7119
  - type: recall_at_5
7120
  value: 9.013
7121
- - task:
7122
- type: STS
7123
- dataset:
7124
- name: MTEB SICK-R (default)
7125
- type: mteb/sickr-sts
7126
  config: default
7127
- split: test
7128
  revision: 20a6d6f312dd54037fe07a32d58e5e168867909d
 
 
7129
  metrics:
7130
  - type: cosine_pearson
7131
  value: 75.67336823619708
@@ -7145,14 +7139,14 @@ model-index:
7145
  value: 75.67336823619708
7146
  - type: spearman
7147
  value: 64.6753400763881
7148
- - task:
7149
  type: STS
7150
- dataset:
7151
- name: MTEB STS12 (default)
7152
- type: mteb/sts12-sts
7153
  config: default
7154
- split: test
7155
  revision: a0d554a64d88156834ff5ae9920b964011b16384
 
 
7156
  metrics:
7157
  - type: cosine_pearson
7158
  value: 72.06681927996405
@@ -7172,14 +7166,14 @@ model-index:
7172
  value: 72.06681927996405
7173
  - type: spearman
7174
  value: 62.248985055530525
7175
- - task:
7176
  type: STS
7177
- dataset:
7178
- name: MTEB STS13 (default)
7179
- type: mteb/sts13-sts
7180
  config: default
7181
- split: test
7182
  revision: 7e90230a92c190f1bf69ae9002b8cea547a64cca
 
 
7183
  metrics:
7184
  - type: cosine_pearson
7185
  value: 76.53983680018591
@@ -7199,14 +7193,14 @@ model-index:
7199
  value: 76.53983680018591
7200
  - type: spearman
7201
  value: 77.27600787572996
7202
- - task:
7203
  type: STS
7204
- dataset:
7205
- name: MTEB STS14 (default)
7206
- type: mteb/sts14-sts
7207
  config: default
7208
- split: test
7209
  revision: 6031580fec1f6af667f0bd2da0a551cf4f0b2375
 
 
7210
  metrics:
7211
  - type: cosine_pearson
7212
  value: 76.20854411766629
@@ -7226,14 +7220,14 @@ model-index:
7226
  value: 76.20854411766629
7227
  - type: spearman
7228
  value: 71.914099628002
7229
- - task:
7230
  type: STS
7231
- dataset:
7232
- name: MTEB STS15 (default)
7233
- type: mteb/sts15-sts
7234
  config: default
7235
- split: test
7236
  revision: ae752c7c21bf194d8b67fd573edf7ae58183cbe3
 
 
7237
  metrics:
7238
  - type: cosine_pearson
7239
  value: 79.24273419832653
@@ -7253,14 +7247,14 @@ model-index:
7253
  value: 79.24273419832653
7254
  - type: spearman
7255
  value: 79.75345871163103
7256
- - task:
7257
  type: STS
7258
- dataset:
7259
- name: MTEB STS16 (default)
7260
- type: mteb/sts16-sts
7261
  config: default
7262
- split: test
7263
  revision: 4d8694f8f0e0100860b497b999b3dbed754a0513
 
 
7264
  metrics:
7265
  - type: cosine_pearson
7266
  value: 75.64452330127995
@@ -7280,14 +7274,14 @@ model-index:
7280
  value: 75.64452330127995
7281
  - type: spearman
7282
  value: 76.26343823222666
7283
- - task:
7284
  type: STS
7285
- dataset:
7286
- name: MTEB STS17 (es-en)
7287
- type: mteb/sts17-crosslingual-sts
7288
  config: es-en
7289
- split: test
7290
  revision: faeb762787bd10488a50c8b5be4a3b82e411949c
 
 
7291
  metrics:
7292
  - type: cosine_pearson
7293
  value: 17.52217310066287
@@ -7307,14 +7301,14 @@ model-index:
7307
  value: 17.52217310066287
7308
  - type: spearman
7309
  value: 14.729958484232528
7310
- - task:
7311
  type: STS
7312
- dataset:
7313
- name: MTEB STS17 (en-en)
7314
- type: mteb/sts17-crosslingual-sts
7315
  config: en-en
7316
- split: test
7317
  revision: faeb762787bd10488a50c8b5be4a3b82e411949c
 
 
7318
  metrics:
7319
  - type: cosine_pearson
7320
  value: 84.67406984717113
@@ -7334,14 +7328,14 @@ model-index:
7334
  value: 84.67406984717113
7335
  - type: spearman
7336
  value: 85.96709815630739
7337
- - task:
7338
  type: STS
7339
- dataset:
7340
- name: MTEB STS17 (fr-en)
7341
- type: mteb/sts17-crosslingual-sts
7342
  config: fr-en
7343
- split: test
7344
  revision: faeb762787bd10488a50c8b5be4a3b82e411949c
 
 
7345
  metrics:
7346
  - type: cosine_pearson
7347
  value: 36.02331692863771
@@ -7361,14 +7355,14 @@ model-index:
7361
  value: 36.02331692863771
7362
  - type: spearman
7363
  value: 34.28540470062557
7364
- - task:
7365
  type: STS
7366
- dataset:
7367
- name: MTEB STS17 (en-tr)
7368
- type: mteb/sts17-crosslingual-sts
7369
  config: en-tr
7370
- split: test
7371
  revision: faeb762787bd10488a50c8b5be4a3b82e411949c
 
 
7372
  metrics:
7373
  - type: cosine_pearson
7374
  value: 13.925983981770388
@@ -7388,14 +7382,14 @@ model-index:
7388
  value: 13.925983981770388
7389
  - type: spearman
7390
  value: 11.193291331109325
7391
- - task:
7392
  type: STS
7393
- dataset:
7394
- name: MTEB STS17 (en-de)
7395
- type: mteb/sts17-crosslingual-sts
7396
  config: en-de
7397
- split: test
7398
  revision: faeb762787bd10488a50c8b5be4a3b82e411949c
 
 
7399
  metrics:
7400
  - type: cosine_pearson
7401
  value: 26.77839285232968
@@ -7415,14 +7409,14 @@ model-index:
7415
  value: 26.77839285232968
7416
  - type: spearman
7417
  value: 23.010015986939717
7418
- - task:
7419
  type: STS
7420
- dataset:
7421
- name: MTEB STS17 (it-en)
7422
- type: mteb/sts17-crosslingual-sts
7423
  config: it-en
7424
- split: test
7425
  revision: faeb762787bd10488a50c8b5be4a3b82e411949c
 
 
7426
  metrics:
7427
  - type: cosine_pearson
7428
  value: 25.330935194314364
@@ -7442,14 +7436,14 @@ model-index:
7442
  value: 25.330935194314364
7443
  - type: spearman
7444
  value: 23.143555348782797
7445
- - task:
7446
  type: STS
7447
- dataset:
7448
- name: MTEB STS17 (en-ar)
7449
- type: mteb/sts17-crosslingual-sts
7450
  config: en-ar
7451
- split: test
7452
  revision: faeb762787bd10488a50c8b5be4a3b82e411949c
 
 
7453
  metrics:
7454
  - type: cosine_pearson
7455
  value: 6.61651078645899
@@ -7469,14 +7463,14 @@ model-index:
7469
  value: 6.61651078645899
7470
  - type: spearman
7471
  value: 5.415104433010482
7472
- - task:
7473
  type: STS
7474
- dataset:
7475
- name: MTEB STS17 (nl-en)
7476
- type: mteb/sts17-crosslingual-sts
7477
  config: nl-en
7478
- split: test
7479
  revision: faeb762787bd10488a50c8b5be4a3b82e411949c
 
 
7480
  metrics:
7481
  - type: cosine_pearson
7482
  value: 32.718045784523184
@@ -7496,14 +7490,14 @@ model-index:
7496
  value: 32.718045784523184
7497
  - type: spearman
7498
  value: 27.52844368619317
7499
- - task:
7500
  type: STS
7501
- dataset:
7502
- name: MTEB STS22 (es-en)
7503
- type: mteb/sts22-crosslingual-sts
7504
  config: es-en
7505
- split: test
7506
  revision: de9d86b3b84231dc21f76c7b7af1f28e2f57f6e3
 
 
7507
  metrics:
7508
  - type: cosine_pearson
7509
  value: 9.98410299881163
@@ -7523,14 +7517,14 @@ model-index:
7523
  value: 9.98410299881163
7524
  - type: spearman
7525
  value: 10.98684405086525
7526
- - task:
7527
  type: STS
7528
- dataset:
7529
- name: MTEB STS22 (en)
7530
- type: mteb/sts22-crosslingual-sts
7531
  config: en
7532
- split: test
7533
  revision: de9d86b3b84231dc21f76c7b7af1f28e2f57f6e3
 
 
7534
  metrics:
7535
  - type: cosine_pearson
7536
  value: 59.3180680265132
@@ -7550,14 +7544,14 @@ model-index:
7550
  value: 59.3180680265132
7551
  - type: spearman
7552
  value: 63.07956002739231
7553
- - task:
7554
  type: STS
7555
- dataset:
7556
- name: MTEB STS22 (de-en)
7557
- type: mteb/sts22-crosslingual-sts
7558
  config: de-en
7559
- split: test
7560
  revision: de9d86b3b84231dc21f76c7b7af1f28e2f57f6e3
 
 
7561
  metrics:
7562
  - type: cosine_pearson
7563
  value: 29.061215770374826
@@ -7577,14 +7571,14 @@ model-index:
7577
  value: 29.061215770374826
7578
  - type: spearman
7579
  value: 36.21441725938738
7580
- - task:
7581
  type: STS
7582
- dataset:
7583
- name: MTEB STS22 (zh-en)
7584
- type: mteb/sts22-crosslingual-sts
7585
  config: zh-en
7586
- split: test
7587
  revision: de9d86b3b84231dc21f76c7b7af1f28e2f57f6e3
 
 
7588
  metrics:
7589
  - type: cosine_pearson
7590
  value: 11.266385865086239
@@ -7604,14 +7598,14 @@ model-index:
7604
  value: 11.266385865086239
7605
  - type: spearman
7606
  value: 17.291293843893733
7607
- - task:
7608
  type: STS
7609
- dataset:
7610
- name: MTEB STS22 (pl-en)
7611
- type: mteb/sts22-crosslingual-sts
7612
  config: pl-en
7613
- split: test
7614
  revision: de9d86b3b84231dc21f76c7b7af1f28e2f57f6e3
 
 
7615
  metrics:
7616
  - type: cosine_pearson
7617
  value: 9.647587208410648
@@ -7631,14 +7625,14 @@ model-index:
7631
  value: 9.647587208410648
7632
  - type: spearman
7633
  value: 21.33739699413266
7634
- - task:
7635
  type: STS
7636
- dataset:
7637
- name: MTEB STSBenchmark (default)
7638
- type: mteb/stsbenchmark-sts
7639
  config: default
7640
- split: test
7641
  revision: b0fddb56ed78048fa8b90373c8a3cfc37b684831
 
 
7642
  metrics:
7643
  - type: cosine_pearson
7644
  value: 77.2598255013409
@@ -7658,14 +7652,14 @@ model-index:
7658
  value: 77.2598255013409
7659
  - type: spearman
7660
  value: 75.40519061413276
7661
- - task:
7662
- type: Reranking
7663
- dataset:
7664
- name: MTEB SciDocsRR (default)
7665
- type: mteb/scidocs-reranking
7666
  config: default
7667
- split: test
7668
  revision: d3c5e1fc0b855ab6097bf1cda04dd73947d7caab
 
 
7669
  metrics:
7670
  - type: main_score
7671
  value: 72.10127087089839
@@ -7685,14 +7679,14 @@ model-index:
7685
  value: 74.90946175462605
7686
  - type: nAUC_mrr_std
7687
  value: 71.81332269641806
7688
- - task:
7689
- type: Retrieval
7690
- dataset:
7691
- name: MTEB SciFact (default)
7692
- type: mteb/scifact
7693
  config: default
7694
- split: test
7695
  revision: 0228b52cf27578f30900b9e5271d331663a030d7
 
 
7696
  metrics:
7697
  - type: main_score
7698
  value: 50.63999999999999
@@ -7976,14 +7970,14 @@ model-index:
7976
  value: 50.306
7977
  - type: recall_at_5
7978
  value: 57.443999999999996
7979
- - task:
7980
- type: PairClassification
7981
- dataset:
7982
- name: MTEB SprintDuplicateQuestions (default)
7983
- type: mteb/sprintduplicatequestions-pairclassification
7984
  config: default
7985
- split: test
7986
  revision: d66bd1f72af766a5cc4b0ca5e00c162f89e8cc46
 
 
7987
  metrics:
7988
  - type: cosine_accuracy
7989
  value: 99.71386138613862
@@ -8067,14 +8061,14 @@ model-index:
8067
  value: 85.7727737973388
8068
  - type: similarity_recall
8069
  value: 83.8
8070
- - task:
8071
- type: Clustering
8072
- dataset:
8073
- name: MTEB StackExchangeClustering (default)
8074
- type: mteb/stackexchange-clustering
8075
  config: default
8076
- split: test
8077
  revision: 6cbc1f7b2bc0622f2e39d2c77fa502909748c259
 
 
8078
  metrics:
8079
  - type: main_score
8080
  value: 48.18939518021159
@@ -8082,14 +8076,14 @@ model-index:
8082
  value: 48.18939518021159
8083
  - type: v_measure_std
8084
  value: 4.6189444340187995
8085
- - task:
8086
  type: Clustering
8087
- dataset:
8088
- name: MTEB StackExchangeClusteringP2P (default)
8089
- type: mteb/stackexchange-clustering-p2p
8090
  config: default
8091
- split: test
8092
  revision: 815ca46b2622cec33ccafc3735d572c266efdb44
 
 
8093
  metrics:
8094
  - type: main_score
8095
  value: 30.743938802421265
@@ -8097,14 +8091,14 @@ model-index:
8097
  value: 30.743938802421265
8098
  - type: v_measure_std
8099
  value: 1.4645401677053824
8100
- - task:
8101
- type: Reranking
8102
- dataset:
8103
- name: MTEB StackOverflowDupQuestions (default)
8104
- type: mteb/stackoverflowdupquestions-reranking
8105
  config: default
8106
- split: test
8107
  revision: e185fbe320c72810689fc5848eb6114e1ef5ec69
 
 
8108
  metrics:
8109
  - type: main_score
8110
  value: 43.254152892780986
@@ -8124,14 +8118,14 @@ model-index:
8124
  value: 14.093758435205075
8125
  - type: nAUC_mrr_std
8126
  value: 4.198791420159734
8127
- - task:
8128
- type: Summarization
8129
- dataset:
8130
- name: MTEB SummEval (default)
8131
- type: mteb/summeval
8132
  config: default
8133
- split: test
8134
  revision: cda12ad7615edc362dbf25a00fdd61d3b1eaf93c
 
 
8135
  metrics:
8136
  - type: cosine_pearson
8137
  value: 29.88360050203766
@@ -8147,14 +8141,14 @@ model-index:
8147
  value: 29.88360050203766
8148
  - type: spearman
8149
  value: 29.275185932109494
8150
- - task:
8151
- type: Retrieval
8152
- dataset:
8153
- name: MTEB TRECCOVID (default)
8154
- type: mteb/trec-covid
8155
  config: default
8156
- split: test
8157
  revision: bb9466bac8153a0349341eb1b22e06409e78ef4e
 
 
8158
  metrics:
8159
  - type: main_score
8160
  value: 45.747
@@ -8438,14 +8432,14 @@ model-index:
8438
  value: 0.395
8439
  - type: recall_at_5
8440
  value: 0.628
8441
- - task:
8442
  type: Retrieval
8443
- dataset:
8444
- name: MTEB Touche2020 (default)
8445
- type: mteb/touche2020
8446
  config: default
8447
- split: test
8448
  revision: a34f9a33db75fa0cbb21bb5cfc3dae8dc8bec93f
 
 
8449
  metrics:
8450
  - type: main_score
8451
  value: 18.618000000000002
@@ -8729,14 +8723,14 @@ model-index:
8729
  value: 4.2540000000000004
8730
  - type: recall_at_5
8731
  value: 7.142999999999999
8732
- - task:
8733
- type: Classification
8734
- dataset:
8735
- name: MTEB ToxicConversationsClassification (default)
8736
- type: mteb/toxic_conversations_50k
8737
  config: default
8738
- split: test
8739
  revision: edfaf9da55d3dd50d43143d90c1ac476895ae6de
 
 
8740
  metrics:
8741
  - type: accuracy
8742
  value: 69.0625
@@ -8750,14 +8744,14 @@ model-index:
8750
  value: 75.91632061778698
8751
  - type: main_score
8752
  value: 69.0625
8753
- - task:
8754
  type: Classification
8755
- dataset:
8756
- name: MTEB TweetSentimentExtractionClassification (default)
8757
- type: mteb/tweet_sentiment_extraction
8758
  config: default
8759
- split: test
8760
  revision: d604517c81ca91fe16a244d1248fc021f9ecee7a
 
 
8761
  metrics:
8762
  - type: accuracy
8763
  value: 55.387662705149964
@@ -8767,14 +8761,14 @@ model-index:
8767
  value: 55.01561915660653
8768
  - type: main_score
8769
  value: 55.387662705149964
8770
- - task:
8771
- type: Clustering
8772
- dataset:
8773
- name: MTEB TwentyNewsgroupsClustering (default)
8774
- type: mteb/twentynewsgroups-clustering
8775
  config: default
8776
- split: test
8777
  revision: 6125ec4e24fa026cec8a478383ee943acfbd5449
 
 
8778
  metrics:
8779
  - type: main_score
8780
  value: 33.535908963951435
@@ -8782,14 +8776,14 @@ model-index:
8782
  value: 33.535908963951435
8783
  - type: v_measure_std
8784
  value: 1.8862804680454297
8785
- - task:
8786
- type: PairClassification
8787
- dataset:
8788
- name: MTEB TwitterSemEval2015 (default)
8789
- type: mteb/twittersemeval2015-pairclassification
8790
  config: default
8791
- split: test
8792
  revision: 70970daeab8776df92f5ea462b6173c0b46fd2d1
 
 
8793
  metrics:
8794
  - type: cosine_accuracy
8795
  value: 81.57000655659535
@@ -8873,14 +8867,14 @@ model-index:
8873
  value: 47.93632075471698
8874
  - type: similarity_recall
8875
  value: 64.35356200527704
8876
- - task:
8877
  type: PairClassification
8878
- dataset:
8879
- name: MTEB TwitterURLCorpus (default)
8880
- type: mteb/twitterurlcorpus-pairclassification
8881
  config: default
8882
- split: test
8883
  revision: 8b6510b0b1fa4e4c4f879467980e9be563ec1cdf
 
 
8884
  metrics:
8885
  - type: cosine_accuracy
8886
  value: 87.09977878682035
@@ -8964,10 +8958,22 @@ model-index:
8964
  value: 70.7599466704091
8965
  - type: similarity_recall
8966
  value: 77.64089929165382
 
 
 
 
 
 
 
8967
  ---
8968
 
8969
  # potion-base-8M Model Card
8970
 
 
 
 
 
 
8971
  This [Model2Vec](https://github.com/MinishLab/model2vec) model is pre-trained using [Tokenlearn](https://github.com/MinishLab/tokenlearn). It is a distilled version of the [baai/bge-base-en-v1.5](https://huggingface.co/baai/bge-base-en-v1.5) Sentence Transformer. It uses static embeddings, allowing text embeddings to be computed orders of magnitude faster on both GPU and CPU. It is designed for applications where computational resources are limited or where real-time performance is critical.
8972
 
8973
 
 
1
  ---
2
  library_name: model2vec
3
  license: mit
 
 
 
 
4
  model-index:
5
  - name: potion-base-8M
6
  results:
7
+ - dataset:
 
 
 
 
8
  config: en-ext
9
+ name: MTEB AmazonCounterfactualClassification (en-ext)
10
  revision: e8379541af4e31359cca9fbcf4b00f2671dba205
11
+ split: test
12
+ type: mteb/amazon_counterfactual
13
  metrics:
14
  - type: accuracy
15
  value: 72.15142428785607
 
23
  value: 77.34906471545477
24
  - type: main_score
25
  value: 72.15142428785607
26
+ task:
27
  type: Classification
28
+ - dataset:
 
 
29
  config: en
30
+ name: MTEB AmazonCounterfactualClassification (en)
31
  revision: e8379541af4e31359cca9fbcf4b00f2671dba205
32
+ split: test
33
+ type: mteb/amazon_counterfactual
34
  metrics:
35
  - type: accuracy
36
  value: 71.7910447761194
 
44
  value: 74.32788084269461
45
  - type: main_score
46
  value: 71.7910447761194
47
+ task:
48
  type: Classification
49
+ - dataset:
 
 
50
  config: default
51
+ name: MTEB AmazonPolarityClassification (default)
52
  revision: e2d317d38cd51312af73b3d32a06d1a08b442046
53
+ split: test
54
+ type: mteb/amazon_polarity
55
  metrics:
56
  - type: accuracy
57
  value: 72.47644999999999
 
65
  value: 72.2600863044581
66
  - type: main_score
67
  value: 72.47644999999999
68
+ task:
69
  type: Classification
70
+ - dataset:
 
 
71
  config: en
72
+ name: MTEB AmazonReviewsClassification (en)
73
  revision: 1399c76144fd37290681b995c656ef9b2e06e26d
74
+ split: test
75
+ type: mteb/amazon_reviews_multi
76
  metrics:
77
  - type: accuracy
78
  value: 36.012
 
82
  value: 35.38209336470206
83
  - type: main_score
84
  value: 36.012
85
+ task:
86
+ type: Classification
87
+ - dataset:
 
 
88
  config: default
89
+ name: MTEB ArguAna (default)
90
  revision: c22ab2a51041ffd869aaddef7af8d8215647e41a
91
+ split: test
92
+ type: mteb/arguana
93
  metrics:
94
  - type: main_score
95
  value: 41.966
 
373
  value: 42.888
374
  - type: recall_at_5
375
  value: 52.063
376
+ task:
377
+ type: Retrieval
378
+ - dataset:
 
 
379
  config: default
380
+ name: MTEB ArxivClusteringP2P (default)
381
  revision: a122ad7f3f0291bf49cc6f4d32aa80929df69d5d
382
+ split: test
383
+ type: mteb/arxiv-clustering-p2p
384
  metrics:
385
  - type: main_score
386
  value: 35.387660145946825
 
388
  value: 35.387660145946825
389
  - type: v_measure_std
390
  value: 14.022525689022785
391
+ task:
392
  type: Clustering
393
+ - dataset:
 
 
394
  config: default
395
+ name: MTEB ArxivClusteringS2S (default)
396
  revision: f910caf1a6075f7329cdf8c1a6135696f37dbd53
397
+ split: test
398
+ type: mteb/arxiv-clustering-s2s
399
  metrics:
400
  - type: main_score
401
  value: 25.26058942964131
 
403
  value: 25.26058942964131
404
  - type: v_measure_std
405
  value: 14.850432186356857
406
+ task:
407
+ type: Clustering
408
+ - dataset:
 
 
409
  config: default
410
+ name: MTEB AskUbuntuDupQuestions (default)
411
  revision: 2000358ca161889fa9c082cb41daa8dcfb161a54
412
+ split: test
413
+ type: mteb/askubuntudupquestions-reranking
414
  metrics:
415
  - type: main_score
416
  value: 54.13950871400633
 
430
  value: 21.861692980537956
431
  - type: nAUC_mrr_std
432
  value: 8.256966784037171
433
+ task:
434
+ type: Reranking
435
+ - dataset:
 
 
436
  config: default
437
+ name: MTEB BIOSSES (default)
438
  revision: d3fb88f8f02e40887cd149695127462bbcf29b4a
439
+ split: test
440
+ type: mteb/biosses-sts
441
  metrics:
442
  - type: cosine_pearson
443
  value: 79.11612010879227
 
457
  value: 79.11612010879227
458
  - type: spearman
459
  value: 75.85775256673794
460
+ task:
461
+ type: STS
462
+ - dataset:
 
 
463
  config: default
464
+ name: MTEB Banking77Classification (default)
465
  revision: 0fd18e25b25c072e09e0d92ab615fda904d66300
466
+ split: test
467
+ type: mteb/banking77
468
  metrics:
469
  - type: accuracy
470
  value: 72.63636363636363
 
474
  value: 71.69751597573539
475
  - type: main_score
476
  value: 72.63636363636363
477
+ task:
478
+ type: Classification
479
+ - dataset:
 
 
480
  config: default
481
+ name: MTEB BiorxivClusteringP2P (default)
482
  revision: 65b79d1d13f80053f67aca9498d9402c2d9f1f40
483
+ split: test
484
+ type: mteb/biorxiv-clustering-p2p
485
  metrics:
486
  - type: main_score
487
  value: 30.861840536151014
 
489
  value: 30.861840536151014
490
  - type: v_measure_std
491
  value: 0.8096483751274005
492
+ task:
493
  type: Clustering
494
+ - dataset:
 
 
495
  config: default
496
+ name: MTEB BiorxivClusteringS2S (default)
497
  revision: 258694dd0231531bc1fd9de6ceb52a0853c6d908
498
+ split: test
499
+ type: mteb/biorxiv-clustering-s2s
500
  metrics:
501
  - type: main_score
502
  value: 20.219544420664455
 
504
  value: 20.219544420664455
505
  - type: v_measure_std
506
  value: 0.7431903039116942
507
+ task:
508
+ type: Clustering
509
+ - dataset:
 
 
510
  config: default
511
+ name: MTEB CQADupstackAndroidRetrieval (default)
512
  revision: f46a197baaae43b4f621051089b82a364682dfeb
513
+ split: test
514
+ type: mteb/cqadupstack-android
515
  metrics:
516
  - type: main_score
517
  value: 31.835
 
795
  value: 29.301
796
  - type: recall_at_5
797
  value: 33.918
798
+ task:
799
  type: Retrieval
800
+ - dataset:
 
 
801
  config: default
802
+ name: MTEB CQADupstackEnglishRetrieval (default)
803
  revision: ad9991cb51e31e31e430383c75ffb2885547b5f0
804
+ split: test
805
+ type: mteb/cqadupstack-english
806
  metrics:
807
  - type: main_score
808
  value: 25.734
 
1086
  value: 23.54
1087
  - type: recall_at_5
1088
  value: 28.054000000000002
1089
+ task:
1090
  type: Retrieval
1091
+ - dataset:
 
 
1092
  config: default
1093
+ name: MTEB CQADupstackGamingRetrieval (default)
1094
  revision: 4885aa143210c98657558c04aaf3dc47cfb54340
1095
+ split: test
1096
+ type: mteb/cqadupstack-gaming
1097
  metrics:
1098
  - type: main_score
1099
  value: 39.235
 
1377
  value: 37.830999999999996
1378
  - type: recall_at_5
1379
  value: 43.811
1380
+ task:
1381
  type: Retrieval
1382
+ - dataset:
 
 
1383
  config: default
1384
+ name: MTEB CQADupstackGisRetrieval (default)
1385
  revision: 5003b3064772da1887988e05400cf3806fe491f2
1386
+ split: test
1387
+ type: mteb/cqadupstack-gis
1388
  metrics:
1389
  - type: main_score
1390
  value: 19.197
 
1668
  value: 19.417
1669
  - type: recall_at_5
1670
  value: 23.577
1671
+ task:
1672
  type: Retrieval
1673
+ - dataset:
 
 
1674
  config: default
1675
+ name: MTEB CQADupstackMathematicaRetrieval (default)
1676
  revision: 90fceea13679c63fe563ded68f3b6f06e50061de
1677
+ split: test
1678
+ type: mteb/cqadupstack-mathematica
1679
  metrics:
1680
  - type: main_score
1681
  value: 12.145
 
1959
  value: 11.178
1960
  - type: recall_at_5
1961
  value: 13.877
1962
+ task:
1963
  type: Retrieval
1964
+ - dataset:
 
 
1965
  config: default
1966
+ name: MTEB CQADupstackPhysicsRetrieval (default)
1967
  revision: 79531abbd1fb92d06c6d6315a0cbbbf5bb247ea4
1968
+ split: test
1969
+ type: mteb/cqadupstack-physics
1970
  metrics:
1971
  - type: main_score
1972
  value: 26.667999999999996
 
2250
  value: 24.524
2251
  - type: recall_at_5
2252
  value: 28.787000000000003
2253
+ task:
2254
  type: Retrieval
2255
+ - dataset:
 
 
2256
  config: default
2257
+ name: MTEB CQADupstackProgrammersRetrieval (default)
2258
  revision: 6184bc1440d2dbc7612be22b50686b8826d22b32
2259
+ split: test
2260
+ type: mteb/cqadupstack-programmers
2261
  metrics:
2262
  - type: main_score
2263
  value: 21.66
 
2541
  value: 19.79
2542
  - type: recall_at_5
2543
  value: 23.376
2544
+ task:
2545
  type: Retrieval
2546
+ - dataset:
 
 
2547
  config: default
2548
+ name: MTEB CQADupstackRetrieval (default)
2549
  revision: CQADupstackRetrieval_is_a_combined_dataset
2550
+ split: test
2551
+ type: CQADupstackRetrieval_is_a_combined_dataset
2552
  metrics:
2553
  - type: main_score
2554
  value: 22.302333333333333
2555
  - type: ndcg_at_10
2556
  value: 22.302333333333333
2557
+ task:
2558
  type: Retrieval
2559
+ - dataset:
 
 
2560
  config: default
2561
+ name: MTEB CQADupstackStatsRetrieval (default)
2562
  revision: 65ac3a16b8e91f9cee4c9828cc7c335575432a2a
2563
+ split: test
2564
+ type: mteb/cqadupstack-stats
2565
  metrics:
2566
  - type: main_score
2567
  value: 17.253
 
2845
  value: 17.134
2846
  - type: recall_at_5
2847
  value: 19.72
2848
+ task:
2849
  type: Retrieval
2850
+ - dataset:
 
 
2851
  config: default
2852
+ name: MTEB CQADupstackTexRetrieval (default)
2853
  revision: 46989137a86843e03a6195de44b09deda022eec7
2854
+ split: test
2855
+ type: mteb/cqadupstack-tex
2856
  metrics:
2857
  - type: main_score
2858
  value: 13.308
 
3136
  value: 12.199
3137
  - type: recall_at_5
3138
  value: 14.87
3139
+ task:
3140
  type: Retrieval
3141
+ - dataset:
 
 
3142
  config: default
3143
+ name: MTEB CQADupstackUnixRetrieval (default)
3144
  revision: 6c6430d3a6d36f8d2a829195bc5dc94d7e063e53
3145
+ split: test
3146
+ type: mteb/cqadupstack-unix
3147
  metrics:
3148
  - type: main_score
3149
  value: 20.061999999999998
 
3427
  value: 18.859
3428
  - type: recall_at_5
3429
  value: 22.435
3430
+ task:
3431
  type: Retrieval
3432
+ - dataset:
 
 
3433
  config: default
3434
+ name: MTEB CQADupstackWebmastersRetrieval (default)
3435
  revision: 160c094312a0e1facb97e55eeddb698c0abe3571
3436
+ split: test
3437
+ type: mteb/cqadupstack-webmasters
3438
  metrics:
3439
  - type: main_score
3440
  value: 23.735999999999997
 
3718
  value: 21.397
3719
  - type: recall_at_5
3720
  value: 24.863
3721
+ task:
3722
  type: Retrieval
3723
+ - dataset:
 
 
3724
  config: default
3725
+ name: MTEB CQADupstackWordpressRetrieval (default)
3726
  revision: 4ffe81d471b1924886b33c7567bfb200e9eec5c4
3727
+ split: test
3728
+ type: mteb/cqadupstack-wordpress
3729
  metrics:
3730
  - type: main_score
3731
  value: 16.794999999999998
 
4009
  value: 15.440999999999999
4010
  - type: recall_at_5
4011
  value: 18.829
4012
+ task:
4013
  type: Retrieval
4014
+ - dataset:
 
 
4015
  config: default
4016
+ name: MTEB ClimateFEVER (default)
4017
  revision: 47f2ac6acb640fc46020b02a5b59fdda04d39380
4018
+ split: test
4019
+ type: mteb/climate-fever
4020
  metrics:
4021
  - type: main_score
4022
  value: 17.288999999999998
 
4300
  value: 12.232
4301
  - type: recall_at_5
4302
  value: 15.540999999999999
4303
+ task:
4304
  type: Retrieval
4305
+ - dataset:
 
 
4306
  config: default
4307
+ name: MTEB DBPedia (default)
4308
  revision: c0f706b76e590d620bd6618b3ca8efdd34e2d659
4309
+ split: test
4310
+ type: mteb/dbpedia
4311
  metrics:
4312
  - type: main_score
4313
  value: 24.235
 
4591
  value: 8.698
4592
  - type: recall_at_5
4593
  value: 11.052
4594
+ task:
4595
+ type: Retrieval
4596
+ - dataset:
 
 
4597
  config: default
4598
+ name: MTEB EmotionClassification (default)
4599
  revision: 4f58c6b202a23cf9a4da393831edf4f9183cad37
4600
+ split: test
4601
+ type: mteb/emotion
4602
  metrics:
4603
  - type: accuracy
4604
  value: 45.09
 
4608
  value: 47.129694558751545
4609
  - type: main_score
4610
  value: 45.09
4611
+ task:
4612
+ type: Classification
4613
+ - dataset:
 
 
4614
  config: default
4615
+ name: MTEB FEVER (default)
4616
  revision: bea83ef9e8fb933d90a2f1d5515737465d613e12
4617
+ split: test
4618
+ type: mteb/fever
4619
  metrics:
4620
  - type: main_score
4621
  value: 30.267
 
4899
  value: 29.42
4900
  - type: recall_at_5
4901
  value: 36.112
4902
+ task:
4903
  type: Retrieval
4904
+ - dataset:
 
 
4905
  config: default
4906
+ name: MTEB FiQA2018 (default)
4907
  revision: 27a168819829fe9bcd655c2df245fb19452e8e06
4908
+ split: test
4909
+ type: mteb/fiqa
4910
  metrics:
4911
  - type: main_score
4912
  value: 16.619
 
5190
  value: 12.485
5191
  - type: recall_at_5
5192
  value: 15.862000000000002
5193
+ task:
5194
  type: Retrieval
5195
+ - dataset:
 
 
5196
  config: default
5197
+ name: MTEB HotpotQA (default)
5198
  revision: ab518f4d6fcca38d87c25209f94beba119d02014
5199
+ split: test
5200
+ type: mteb/hotpotqa
5201
  metrics:
5202
  - type: main_score
5203
  value: 36.217
 
5481
  value: 30.311
5482
  - type: recall_at_5
5483
  value: 34.092
5484
+ task:
5485
+ type: Retrieval
5486
+ - dataset:
 
 
5487
  config: default
5488
+ name: MTEB ImdbClassification (default)
5489
  revision: 3d86128a09e091d6018b6d26cad27f2739fc2db7
5490
+ split: test
5491
+ type: mteb/imdb
5492
  metrics:
5493
  - type: accuracy
5494
  value: 70.78
 
5502
  value: 70.56498155979408
5503
  - type: main_score
5504
  value: 70.78
5505
+ task:
5506
+ type: Classification
5507
+ - dataset:
 
 
5508
  config: default
5509
+ name: MTEB MSMARCO (default)
5510
  revision: c5a29a104738b98a9e76336939199e264163d4a0
5511
+ split: test
5512
+ type: mteb/msmarco
5513
  metrics:
5514
  - type: main_score
5515
  value: 34.981
 
5793
  value: 2.8240000000000003
5794
  - type: recall_at_5
5795
  value: 4.476999999999999
5796
+ task:
5797
+ type: Retrieval
5798
+ - dataset:
 
 
5799
  config: en
5800
+ name: MTEB MTOPDomainClassification (en)
5801
  revision: d80d48c1eb48d3562165c59d59d0034df9fff0bf
5802
+ split: test
5803
+ type: mteb/mtop_domain
5804
  metrics:
5805
  - type: accuracy
5806
  value: 88.95576835385319
 
5810
  value: 89.00721562093213
5811
  - type: main_score
5812
  value: 88.95576835385319
5813
+ task:
5814
  type: Classification
5815
+ - dataset:
 
 
5816
  config: en
5817
+ name: MTEB MTOPIntentClassification (en)
5818
  revision: ae001d0e6b1228650b7bd1c2c65fb50ad11a8aba
5819
+ split: test
5820
+ type: mteb/mtop_intent
5821
  metrics:
5822
  - type: accuracy
5823
  value: 56.99726402188783
 
5827
  value: 59.96788951671549
5828
  - type: main_score
5829
  value: 56.99726402188783
5830
+ task:
5831
  type: Classification
5832
+ - dataset:
 
 
5833
  config: en
5834
+ name: MTEB MassiveIntentClassification (en)
5835
  revision: 4672e20407010da34463acc759c162ca9734bca6
5836
+ split: test
5837
+ type: mteb/amazon_massive_intent
5838
  metrics:
5839
  - type: accuracy
5840
  value: 63.79287155346336
 
5844
  value: 62.567311481126055
5845
  - type: main_score
5846
  value: 63.79287155346336
5847
+ task:
5848
  type: Classification
5849
+ - dataset:
 
 
5850
  config: en
5851
+ name: MTEB MassiveScenarioClassification (en)
5852
  revision: fad2c6e8459f9e1c45d9315f4953d921437d70f8
5853
+ split: test
5854
+ type: mteb/amazon_massive_scenario
5855
  metrics:
5856
  - type: accuracy
5857
  value: 70.30934767989241
 
5861
  value: 70.1128179307388
5862
  - type: main_score
5863
  value: 70.30934767989241
5864
+ task:
5865
+ type: Classification
5866
+ - dataset:
5867
+ config: default
5868
  name: MTEB MedrxivClusteringP2P (default)
5869
+ revision: e7a26af6f3ae46b30dde8737f02c07b1505bcc73
5870
+ split: test
5871
  type: mteb/medrxiv-clustering-p2p
 
 
 
5872
  metrics:
5873
  - type: main_score
5874
  value: 27.61734940907637
 
5876
  value: 27.61734940907637
5877
  - type: v_measure_std
5878
  value: 1.2248100208316097
5879
+ task:
5880
  type: Clustering
5881
+ - dataset:
 
 
5882
  config: default
5883
+ name: MTEB MedrxivClusteringS2S (default)
5884
  revision: 35191c8c0dca72d8ff3efcd72aa802307d469663
5885
+ split: test
5886
+ type: mteb/medrxiv-clustering-s2s
5887
  metrics:
5888
  - type: main_score
5889
  value: 23.802943866708308
 
5891
  value: 23.802943866708308
5892
  - type: v_measure_std
5893
  value: 1.4975518910969763
5894
+ task:
5895
+ type: Clustering
5896
+ - dataset:
 
 
5897
  config: default
5898
+ name: MTEB MindSmallReranking (default)
5899
  revision: 59042f120c80e8afa9cdbb224f67076cec0fc9a7
5900
+ split: test
5901
+ type: mteb/mind_small
5902
  metrics:
5903
  - type: main_score
5904
  value: 29.431722284942175
 
5918
  value: -17.42779158552557
5919
  - type: nAUC_mrr_std
5920
  value: -5.997215692334967
5921
+ task:
5922
+ type: Reranking
5923
+ - dataset:
 
 
5924
  config: default
5925
+ name: MTEB NFCorpus (default)
5926
  revision: ec0fa4fe99da2ff19ca1214b7966684033a58814
5927
+ split: test
5928
+ type: mteb/nfcorpus
5929
  metrics:
5930
  - type: main_score
5931
  value: 24.267
 
6209
  value: 7.236
6210
  - type: recall_at_5
6211
  value: 9.003
6212
+ task:
6213
  type: Retrieval
6214
+ - dataset:
 
 
6215
  config: default
6216
+ name: MTEB NQ (default)
6217
  revision: b774495ed302d8c44a3a7ea25c90dbce03968f31
6218
+ split: test
6219
+ type: mteb/nq
6220
  metrics:
6221
  - type: main_score
6222
  value: 19.543
 
6500
  value: 17.669
6501
  - type: recall_at_5
6502
  value: 23.488999999999997
6503
+ task:
6504
  type: Retrieval
6505
+ - dataset:
 
 
6506
  config: default
6507
+ name: MTEB QuoraRetrieval (default)
6508
  revision: e4e08e0b7dbe3c8700f0daef558ff32256715259
6509
+ split: test
6510
+ type: mteb/quora
6511
  metrics:
6512
  - type: main_score
6513
  value: 80.598
 
6791
  value: 79.074
6792
  - type: recall_at_5
6793
  value: 84.143
6794
+ task:
6795
+ type: Retrieval
6796
+ - dataset:
 
 
6797
  config: default
6798
+ name: MTEB RedditClustering (default)
6799
  revision: 24640382cdbf8abc73003fb0fa6d111a705499eb
6800
+ split: test
6801
+ type: mteb/reddit-clustering
6802
  metrics:
6803
  - type: main_score
6804
  value: 39.5465127563479
 
6806
  value: 39.5465127563479
6807
  - type: v_measure_std
6808
  value: 5.038703300031419
6809
+ task:
6810
  type: Clustering
6811
+ - dataset:
 
 
6812
  config: default
6813
+ name: MTEB RedditClusteringP2P (default)
6814
  revision: 385e3cb46b4cfa89021f56c4380204149d0efe33
6815
+ split: test
6816
+ type: mteb/reddit-clustering-p2p
6817
  metrics:
6818
  - type: main_score
6819
  value: 47.07911795189491
 
6821
  value: 47.07911795189491
6822
  - type: v_measure_std
6823
  value: 11.546436135362846
6824
+ task:
6825
+ type: Clustering
6826
+ - dataset:
 
 
6827
  config: default
6828
+ name: MTEB SCIDOCS (default)
6829
  revision: f8c2fcf00f625baaa80f62ec5bd9e1fff3b8ae88
6830
+ split: test
6831
+ type: mteb/scidocs
6832
  metrics:
6833
  - type: main_score
6834
  value: 12.386999999999999
 
7112
  value: 6.553000000000001
7113
  - type: recall_at_5
7114
  value: 9.013
7115
+ task:
7116
+ type: Retrieval
7117
+ - dataset:
 
 
7118
  config: default
7119
+ name: MTEB SICK-R (default)
7120
  revision: 20a6d6f312dd54037fe07a32d58e5e168867909d
7121
+ split: test
7122
+ type: mteb/sickr-sts
7123
  metrics:
7124
  - type: cosine_pearson
7125
  value: 75.67336823619708
 
7139
  value: 75.67336823619708
7140
  - type: spearman
7141
  value: 64.6753400763881
7142
+ task:
7143
  type: STS
7144
+ - dataset:
 
 
7145
  config: default
7146
+ name: MTEB STS12 (default)
7147
  revision: a0d554a64d88156834ff5ae9920b964011b16384
7148
+ split: test
7149
+ type: mteb/sts12-sts
7150
  metrics:
7151
  - type: cosine_pearson
7152
  value: 72.06681927996405
 
7166
  value: 72.06681927996405
7167
  - type: spearman
7168
  value: 62.248985055530525
7169
+ task:
7170
  type: STS
7171
+ - dataset:
 
 
7172
  config: default
7173
+ name: MTEB STS13 (default)
7174
  revision: 7e90230a92c190f1bf69ae9002b8cea547a64cca
7175
+ split: test
7176
+ type: mteb/sts13-sts
7177
  metrics:
7178
  - type: cosine_pearson
7179
  value: 76.53983680018591
 
7193
  value: 76.53983680018591
7194
  - type: spearman
7195
  value: 77.27600787572996
7196
+ task:
7197
  type: STS
7198
+ - dataset:
 
 
7199
  config: default
7200
+ name: MTEB STS14 (default)
7201
  revision: 6031580fec1f6af667f0bd2da0a551cf4f0b2375
7202
+ split: test
7203
+ type: mteb/sts14-sts
7204
  metrics:
7205
  - type: cosine_pearson
7206
  value: 76.20854411766629
 
7220
  value: 76.20854411766629
7221
  - type: spearman
7222
  value: 71.914099628002
7223
+ task:
7224
  type: STS
7225
+ - dataset:
 
 
7226
  config: default
7227
+ name: MTEB STS15 (default)
7228
  revision: ae752c7c21bf194d8b67fd573edf7ae58183cbe3
7229
+ split: test
7230
+ type: mteb/sts15-sts
7231
  metrics:
7232
  - type: cosine_pearson
7233
  value: 79.24273419832653
 
7247
  value: 79.24273419832653
7248
  - type: spearman
7249
  value: 79.75345871163103
7250
+ task:
7251
  type: STS
7252
+ - dataset:
 
 
7253
  config: default
7254
+ name: MTEB STS16 (default)
7255
  revision: 4d8694f8f0e0100860b497b999b3dbed754a0513
7256
+ split: test
7257
+ type: mteb/sts16-sts
7258
  metrics:
7259
  - type: cosine_pearson
7260
  value: 75.64452330127995
 
7274
  value: 75.64452330127995
7275
  - type: spearman
7276
  value: 76.26343823222666
7277
+ task:
7278
  type: STS
7279
+ - dataset:
 
 
7280
  config: es-en
7281
+ name: MTEB STS17 (es-en)
7282
  revision: faeb762787bd10488a50c8b5be4a3b82e411949c
7283
+ split: test
7284
+ type: mteb/sts17-crosslingual-sts
7285
  metrics:
7286
  - type: cosine_pearson
7287
  value: 17.52217310066287
 
7301
  value: 17.52217310066287
7302
  - type: spearman
7303
  value: 14.729958484232528
7304
+ task:
7305
  type: STS
7306
+ - dataset:
 
 
7307
  config: en-en
7308
+ name: MTEB STS17 (en-en)
7309
  revision: faeb762787bd10488a50c8b5be4a3b82e411949c
7310
+ split: test
7311
+ type: mteb/sts17-crosslingual-sts
7312
  metrics:
7313
  - type: cosine_pearson
7314
  value: 84.67406984717113
 
7328
  value: 84.67406984717113
7329
  - type: spearman
7330
  value: 85.96709815630739
7331
+ task:
7332
  type: STS
7333
+ - dataset:
 
 
7334
  config: fr-en
7335
+ name: MTEB STS17 (fr-en)
7336
  revision: faeb762787bd10488a50c8b5be4a3b82e411949c
7337
+ split: test
7338
+ type: mteb/sts17-crosslingual-sts
7339
  metrics:
7340
  - type: cosine_pearson
7341
  value: 36.02331692863771
 
7355
  value: 36.02331692863771
7356
  - type: spearman
7357
  value: 34.28540470062557
7358
+ task:
7359
  type: STS
7360
+ - dataset:
 
 
7361
  config: en-tr
7362
+ name: MTEB STS17 (en-tr)
7363
  revision: faeb762787bd10488a50c8b5be4a3b82e411949c
7364
+ split: test
7365
+ type: mteb/sts17-crosslingual-sts
7366
  metrics:
7367
  - type: cosine_pearson
7368
  value: 13.925983981770388
 
7382
  value: 13.925983981770388
7383
  - type: spearman
7384
  value: 11.193291331109325
7385
+ task:
7386
  type: STS
7387
+ - dataset:
 
 
7388
  config: en-de
7389
+ name: MTEB STS17 (en-de)
7390
  revision: faeb762787bd10488a50c8b5be4a3b82e411949c
7391
+ split: test
7392
+ type: mteb/sts17-crosslingual-sts
7393
  metrics:
7394
  - type: cosine_pearson
7395
  value: 26.77839285232968
 
7409
  value: 26.77839285232968
7410
  - type: spearman
7411
  value: 23.010015986939717
7412
+ task:
7413
  type: STS
7414
+ - dataset:
 
 
7415
  config: it-en
7416
+ name: MTEB STS17 (it-en)
7417
  revision: faeb762787bd10488a50c8b5be4a3b82e411949c
7418
+ split: test
7419
+ type: mteb/sts17-crosslingual-sts
7420
  metrics:
7421
  - type: cosine_pearson
7422
  value: 25.330935194314364
 
7436
  value: 25.330935194314364
7437
  - type: spearman
7438
  value: 23.143555348782797
7439
+ task:
7440
  type: STS
7441
+ - dataset:
 
 
7442
  config: en-ar
7443
+ name: MTEB STS17 (en-ar)
7444
  revision: faeb762787bd10488a50c8b5be4a3b82e411949c
7445
+ split: test
7446
+ type: mteb/sts17-crosslingual-sts
7447
  metrics:
7448
  - type: cosine_pearson
7449
  value: 6.61651078645899
 
7463
  value: 6.61651078645899
7464
  - type: spearman
7465
  value: 5.415104433010482
7466
+ task:
7467
  type: STS
7468
+ - dataset:
 
 
7469
  config: nl-en
7470
+ name: MTEB STS17 (nl-en)
7471
  revision: faeb762787bd10488a50c8b5be4a3b82e411949c
7472
+ split: test
7473
+ type: mteb/sts17-crosslingual-sts
7474
  metrics:
7475
  - type: cosine_pearson
7476
  value: 32.718045784523184
 
7490
  value: 32.718045784523184
7491
  - type: spearman
7492
  value: 27.52844368619317
7493
+ task:
7494
  type: STS
7495
+ - dataset:
 
 
7496
  config: es-en
7497
+ name: MTEB STS22 (es-en)
7498
  revision: de9d86b3b84231dc21f76c7b7af1f28e2f57f6e3
7499
+ split: test
7500
+ type: mteb/sts22-crosslingual-sts
7501
  metrics:
7502
  - type: cosine_pearson
7503
  value: 9.98410299881163
 
7517
  value: 9.98410299881163
7518
  - type: spearman
7519
  value: 10.98684405086525
7520
+ task:
7521
  type: STS
7522
+ - dataset:
 
 
7523
  config: en
7524
+ name: MTEB STS22 (en)
7525
  revision: de9d86b3b84231dc21f76c7b7af1f28e2f57f6e3
7526
+ split: test
7527
+ type: mteb/sts22-crosslingual-sts
7528
  metrics:
7529
  - type: cosine_pearson
7530
  value: 59.3180680265132
 
7544
  value: 59.3180680265132
7545
  - type: spearman
7546
  value: 63.07956002739231
7547
+ task:
7548
  type: STS
7549
+ - dataset:
 
 
7550
  config: de-en
7551
+ name: MTEB STS22 (de-en)
7552
  revision: de9d86b3b84231dc21f76c7b7af1f28e2f57f6e3
7553
+ split: test
7554
+ type: mteb/sts22-crosslingual-sts
7555
  metrics:
7556
  - type: cosine_pearson
7557
  value: 29.061215770374826
 
7571
  value: 29.061215770374826
7572
  - type: spearman
7573
  value: 36.21441725938738
7574
+ task:
7575
  type: STS
7576
+ - dataset:
 
 
7577
  config: zh-en
7578
+ name: MTEB STS22 (zh-en)
7579
  revision: de9d86b3b84231dc21f76c7b7af1f28e2f57f6e3
7580
+ split: test
7581
+ type: mteb/sts22-crosslingual-sts
7582
  metrics:
7583
  - type: cosine_pearson
7584
  value: 11.266385865086239
 
7598
  value: 11.266385865086239
7599
  - type: spearman
7600
  value: 17.291293843893733
7601
+ task:
7602
  type: STS
7603
+ - dataset:
 
 
7604
  config: pl-en
7605
+ name: MTEB STS22 (pl-en)
7606
  revision: de9d86b3b84231dc21f76c7b7af1f28e2f57f6e3
7607
+ split: test
7608
+ type: mteb/sts22-crosslingual-sts
7609
  metrics:
7610
  - type: cosine_pearson
7611
  value: 9.647587208410648
 
7625
  value: 9.647587208410648
7626
  - type: spearman
7627
  value: 21.33739699413266
7628
+ task:
7629
  type: STS
7630
+ - dataset:
 
 
7631
  config: default
7632
+ name: MTEB STSBenchmark (default)
7633
  revision: b0fddb56ed78048fa8b90373c8a3cfc37b684831
7634
+ split: test
7635
+ type: mteb/stsbenchmark-sts
7636
  metrics:
7637
  - type: cosine_pearson
7638
  value: 77.2598255013409
 
7652
  value: 77.2598255013409
7653
  - type: spearman
7654
  value: 75.40519061413276
7655
+ task:
7656
+ type: STS
7657
+ - dataset:
 
 
7658
  config: default
7659
+ name: MTEB SciDocsRR (default)
7660
  revision: d3c5e1fc0b855ab6097bf1cda04dd73947d7caab
7661
+ split: test
7662
+ type: mteb/scidocs-reranking
7663
  metrics:
7664
  - type: main_score
7665
  value: 72.10127087089839
 
7679
  value: 74.90946175462605
7680
  - type: nAUC_mrr_std
7681
  value: 71.81332269641806
7682
+ task:
7683
+ type: Reranking
7684
+ - dataset:
 
 
7685
  config: default
7686
+ name: MTEB SciFact (default)
7687
  revision: 0228b52cf27578f30900b9e5271d331663a030d7
7688
+ split: test
7689
+ type: mteb/scifact
7690
  metrics:
7691
  - type: main_score
7692
  value: 50.63999999999999
 
7970
  value: 50.306
7971
  - type: recall_at_5
7972
  value: 57.443999999999996
7973
+ task:
7974
+ type: Retrieval
7975
+ - dataset:
 
 
7976
  config: default
7977
+ name: MTEB SprintDuplicateQuestions (default)
7978
  revision: d66bd1f72af766a5cc4b0ca5e00c162f89e8cc46
7979
+ split: test
7980
+ type: mteb/sprintduplicatequestions-pairclassification
7981
  metrics:
7982
  - type: cosine_accuracy
7983
  value: 99.71386138613862
 
8061
  value: 85.7727737973388
8062
  - type: similarity_recall
8063
  value: 83.8
8064
+ task:
8065
+ type: PairClassification
8066
+ - dataset:
 
 
8067
  config: default
8068
+ name: MTEB StackExchangeClustering (default)
8069
  revision: 6cbc1f7b2bc0622f2e39d2c77fa502909748c259
8070
+ split: test
8071
+ type: mteb/stackexchange-clustering
8072
  metrics:
8073
  - type: main_score
8074
  value: 48.18939518021159
 
8076
  value: 48.18939518021159
8077
  - type: v_measure_std
8078
  value: 4.6189444340187995
8079
+ task:
8080
  type: Clustering
8081
+ - dataset:
 
 
8082
  config: default
8083
+ name: MTEB StackExchangeClusteringP2P (default)
8084
  revision: 815ca46b2622cec33ccafc3735d572c266efdb44
8085
+ split: test
8086
+ type: mteb/stackexchange-clustering-p2p
8087
  metrics:
8088
  - type: main_score
8089
  value: 30.743938802421265
 
8091
  value: 30.743938802421265
8092
  - type: v_measure_std
8093
  value: 1.4645401677053824
8094
+ task:
8095
+ type: Clustering
8096
+ - dataset:
 
 
8097
  config: default
8098
+ name: MTEB StackOverflowDupQuestions (default)
8099
  revision: e185fbe320c72810689fc5848eb6114e1ef5ec69
8100
+ split: test
8101
+ type: mteb/stackoverflowdupquestions-reranking
8102
  metrics:
8103
  - type: main_score
8104
  value: 43.254152892780986
 
8118
  value: 14.093758435205075
8119
  - type: nAUC_mrr_std
8120
  value: 4.198791420159734
8121
+ task:
8122
+ type: Reranking
8123
+ - dataset:
 
 
8124
  config: default
8125
+ name: MTEB SummEval (default)
8126
  revision: cda12ad7615edc362dbf25a00fdd61d3b1eaf93c
8127
+ split: test
8128
+ type: mteb/summeval
8129
  metrics:
8130
  - type: cosine_pearson
8131
  value: 29.88360050203766
 
8141
  value: 29.88360050203766
8142
  - type: spearman
8143
  value: 29.275185932109494
8144
+ task:
8145
+ type: Summarization
8146
+ - dataset:
 
 
8147
  config: default
8148
+ name: MTEB TRECCOVID (default)
8149
  revision: bb9466bac8153a0349341eb1b22e06409e78ef4e
8150
+ split: test
8151
+ type: mteb/trec-covid
8152
  metrics:
8153
  - type: main_score
8154
  value: 45.747
 
8432
  value: 0.395
8433
  - type: recall_at_5
8434
  value: 0.628
8435
+ task:
8436
  type: Retrieval
8437
+ - dataset:
 
 
8438
  config: default
8439
+ name: MTEB Touche2020 (default)
8440
  revision: a34f9a33db75fa0cbb21bb5cfc3dae8dc8bec93f
8441
+ split: test
8442
+ type: mteb/touche2020
8443
  metrics:
8444
  - type: main_score
8445
  value: 18.618000000000002
 
8723
  value: 4.2540000000000004
8724
  - type: recall_at_5
8725
  value: 7.142999999999999
8726
+ task:
8727
+ type: Retrieval
8728
+ - dataset:
 
 
8729
  config: default
8730
+ name: MTEB ToxicConversationsClassification (default)
8731
  revision: edfaf9da55d3dd50d43143d90c1ac476895ae6de
8732
+ split: test
8733
+ type: mteb/toxic_conversations_50k
8734
  metrics:
8735
  - type: accuracy
8736
  value: 69.0625
 
8744
  value: 75.91632061778698
8745
  - type: main_score
8746
  value: 69.0625
8747
+ task:
8748
  type: Classification
8749
+ - dataset:
 
 
8750
  config: default
8751
+ name: MTEB TweetSentimentExtractionClassification (default)
8752
  revision: d604517c81ca91fe16a244d1248fc021f9ecee7a
8753
+ split: test
8754
+ type: mteb/tweet_sentiment_extraction
8755
  metrics:
8756
  - type: accuracy
8757
  value: 55.387662705149964
 
8761
  value: 55.01561915660653
8762
  - type: main_score
8763
  value: 55.387662705149964
8764
+ task:
8765
+ type: Classification
8766
+ - dataset:
 
 
8767
  config: default
8768
+ name: MTEB TwentyNewsgroupsClustering (default)
8769
  revision: 6125ec4e24fa026cec8a478383ee943acfbd5449
8770
+ split: test
8771
+ type: mteb/twentynewsgroups-clustering
8772
  metrics:
8773
  - type: main_score
8774
  value: 33.535908963951435
 
8776
  value: 33.535908963951435
8777
  - type: v_measure_std
8778
  value: 1.8862804680454297
8779
+ task:
8780
+ type: Clustering
8781
+ - dataset:
 
 
8782
  config: default
8783
+ name: MTEB TwitterSemEval2015 (default)
8784
  revision: 70970daeab8776df92f5ea462b6173c0b46fd2d1
8785
+ split: test
8786
+ type: mteb/twittersemeval2015-pairclassification
8787
  metrics:
8788
  - type: cosine_accuracy
8789
  value: 81.57000655659535
 
8867
  value: 47.93632075471698
8868
  - type: similarity_recall
8869
  value: 64.35356200527704
8870
+ task:
8871
  type: PairClassification
8872
+ - dataset:
 
 
8873
  config: default
8874
+ name: MTEB TwitterURLCorpus (default)
8875
  revision: 8b6510b0b1fa4e4c4f879467980e9be563ec1cdf
8876
+ split: test
8877
+ type: mteb/twitterurlcorpus-pairclassification
8878
  metrics:
8879
  - type: cosine_accuracy
8880
  value: 87.09977878682035
 
8958
  value: 70.7599466704091
8959
  - type: similarity_recall
8960
  value: 77.64089929165382
8961
+ task:
8962
+ type: PairClassification
8963
+ model_name: potion-base-8M
8964
+ tags:
8965
+ - embeddings
8966
+ - static-embeddings
8967
+ - mteb
8968
  ---
8969
 
8970
  # potion-base-8M Model Card
8971
 
8972
+ <div align="center">
8973
+ <img width="35%" alt="Model2Vec logo" src="https://raw.githubusercontent.com/MinishLab/model2vec/main/assets/images/logo_v2.png">
8974
+ </div>
8975
+
8976
+
8977
  This [Model2Vec](https://github.com/MinishLab/model2vec) model is pre-trained using [Tokenlearn](https://github.com/MinishLab/tokenlearn). It is a distilled version of the [baai/bge-base-en-v1.5](https://huggingface.co/baai/bge-base-en-v1.5) Sentence Transformer. It uses static embeddings, allowing text embeddings to be computed orders of magnitude faster on both GPU and CPU. It is designed for applications where computational resources are limited or where real-time performance is critical.
8978
 
8979