Omartificial-Intelligence-Space commited on
Commit
9ccbfed
·
verified ·
1 Parent(s): f072255

update readme.md

Browse files
Files changed (1) hide show
  1. README.md +174 -13
README.md CHANGED
@@ -36,45 +36,206 @@ model-index:
36
  name: MTEB MIRACLRetrievalHardNegatives (ar)
37
  revision: 95c8db7d4a6e9c1d8a60601afd63d553ae20a2eb
38
  split: dev
39
- type: mteb/miracl-hard-negatives
40
  metrics:
41
  - type: main_score
42
  value: 57.737
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
43
  task:
44
  type: Retrieval
45
  - dataset:
46
- config: ara-ara
47
- name: MTEB MLQARetrieval (ara-ara)
48
- revision: 397ed406c1a7902140303e7faf60fff35b58d285
49
  split: test
50
- type: facebook/mlqa
51
  metrics:
52
  - type: main_score
53
- value: 62.58
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
54
  task:
55
  type: Retrieval
56
  - dataset:
57
  config: ar
58
- name: MTEB MintakaRetrieval (ar)
59
- revision: efa78cc2f74bbcd21eff2261f9e13aebe40b814e
60
- split: test
61
- type: jinaai/mintakaqa
62
  metrics:
63
  - type: main_score
64
- value: 19.13
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
65
  task:
66
  type: Retrieval
67
  - dataset:
68
  config: default
69
- name: MTEB SadeemQuestionRetrieval (default)
70
  revision: 3cb0752b182e5d5d740df547748b06663c8e0bd9
71
  split: test
72
- type: sadeem-ai/sadeem-ar-eval-retrieval-questions
73
  metrics:
74
  - type: main_score
75
  value: 63.155
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
76
  task:
77
  type: Retrieval
 
78
  - dataset:
79
  config: ar-ar
80
  name: MTEB STS17 (ar-ar)
 
36
  name: MTEB MIRACLRetrievalHardNegatives (ar)
37
  revision: 95c8db7d4a6e9c1d8a60601afd63d553ae20a2eb
38
  split: dev
39
+ type: miracl/mmteb-miracl-hardnegatives
40
  metrics:
41
  - type: main_score
42
  value: 57.737
43
+ - type: map_at_1
44
+ value: 32.108
45
+ - type: map_at_3
46
+ value: 44.405
47
+ - type: map_at_5
48
+ value: 47.164
49
+ - type: map_at_10
50
+ value: 49.477
51
+ - type: ndcg_at_1
52
+ value: 48.7
53
+ - type: ndcg_at_3
54
+ value: 51.161
55
+ - type: ndcg_at_5
56
+ value: 53.923
57
+ - type: ndcg_at_10
58
+ value: 57.737
59
+ - type: recall_at_1
60
+ value: 32.108
61
+ - type: recall_at_3
62
+ value: 52.675
63
+ - type: recall_at_5
64
+ value: 60.709
65
+ - type: recall_at_10
66
+ value: 70.61
67
+ - type: precision_at_1
68
+ value: 48.7
69
+ - type: precision_at_3
70
+ value: 29.7
71
+ - type: precision_at_5
72
+ value: 21.34
73
+ - type: precision_at_10
74
+ value: 12.98
75
+ - type: mrr_at_1
76
+ value: 48.7
77
+ - type: mrr_at_3
78
+ value: 57.5333
79
+ - type: mrr_at_5
80
+ value: 59.1333
81
+ - type: mrr_at_10
82
+ value: 60.1163
83
  task:
84
  type: Retrieval
85
  - dataset:
86
+ config: ar
87
+ name: MTEB MintakaRetrieval (ar)
88
+ revision: efa78cc2f74bbcd21eff2261f9e13aebe40b814e
89
  split: test
90
+ type: mintaka/mmteb-mintaka
91
  metrics:
92
  - type: main_score
93
+ value: 19.13
94
+ - type: map_at_1
95
+ value: 11.212
96
+ - type: map_at_3
97
+ value: 14.904
98
+ - type: map_at_5
99
+ value: 15.719
100
+ - type: map_at_10
101
+ value: 16.375
102
+ - type: ndcg_at_1
103
+ value: 11.212
104
+ - type: ndcg_at_3
105
+ value: 16.08
106
+ - type: ndcg_at_5
107
+ value: 17.543
108
+ - type: ndcg_at_10
109
+ value: 19.13
110
+ - type: recall_at_1
111
+ value: 11.212
112
+ - type: recall_at_3
113
+ value: 19.473
114
+ - type: recall_at_5
115
+ value: 23.014
116
+ - type: recall_at_10
117
+ value: 27.916
118
+ - type: precision_at_1
119
+ value: 11.212
120
+ - type: precision_at_3
121
+ value: 6.491
122
+ - type: precision_at_5
123
+ value: 4.603
124
+ - type: precision_at_10
125
+ value: 2.792
126
+ - type: mrr_at_1
127
+ value: 11.212
128
+ - type: mrr_at_3
129
+ value: 14.9039
130
+ - type: mrr_at_5
131
+ value: 15.7187
132
+ - type: mrr_at_10
133
+ value: 16.3746
134
  task:
135
  type: Retrieval
136
  - dataset:
137
  config: ar
138
+ name: MTEB MLQARetrieval (ar)
139
+ revision: 397ed406c1a7902140303e7faf60fff35b58d285
140
+ split: validation
141
+ type: mlqa/mmteb-mlqa
142
  metrics:
143
  - type: main_score
144
+ value: 62.58
145
+ - type: map_at_1
146
+ value: 47.776
147
+ - type: map_at_3
148
+ value: 55.545
149
+ - type: map_at_5
150
+ value: 56.812
151
+ - type: map_at_10
152
+ value: 57.756
153
+ - type: ndcg_at_1
154
+ value: 47.776
155
+ - type: ndcg_at_3
156
+ value: 57.978
157
+ - type: ndcg_at_5
158
+ value: 60.276
159
+ - type: ndcg_at_10
160
+ value: 62.58
161
+ - type: recall_at_1
162
+ value: 47.776
163
+ - type: recall_at_3
164
+ value: 64.99
165
+ - type: recall_at_5
166
+ value: 70.6
167
+ - type: recall_at_10
168
+ value: 77.756
169
+ - type: precision_at_1
170
+ value: 47.776
171
+ - type: precision_at_3
172
+ value: 21.663
173
+ - type: precision_at_5
174
+ value: 14.12
175
+ - type: precision_at_10
176
+ value: 7.776
177
+ - type: mrr_at_1
178
+ value: 47.7756
179
+ - type: mrr_at_3
180
+ value: 55.5448
181
+ - type: mrr_at_5
182
+ value: 56.8117
183
+ - type: mrr_at_10
184
+ value: 57.7562
185
  task:
186
  type: Retrieval
187
  - dataset:
188
  config: default
189
+ name: MTEB SadeemQuestionRetrieval (ar)
190
  revision: 3cb0752b182e5d5d740df547748b06663c8e0bd9
191
  split: test
192
+ type: sadeem/mmteb-sadeem
193
  metrics:
194
  - type: main_score
195
  value: 63.155
196
+ - type: map_at_1
197
+ value: 28.674
198
+ - type: map_at_3
199
+ value: 52.21
200
+ - type: map_at_5
201
+ value: 53.052
202
+ - type: map_at_10
203
+ value: 53.498
204
+ - type: ndcg_at_1
205
+ value: 28.674
206
+ - type: ndcg_at_3
207
+ value: 60.604
208
+ - type: ndcg_at_5
209
+ value: 62.092
210
+ - type: ndcg_at_10
211
+ value: 63.155
212
+ - type: recall_at_1
213
+ value: 28.674
214
+ - type: recall_at_3
215
+ value: 85.112
216
+ - type: recall_at_5
217
+ value: 88.655
218
+ - type: recall_at_10
219
+ value: 91.91
220
+ - type: precision_at_1
221
+ value: 28.674
222
+ - type: precision_at_3
223
+ value: 28.371
224
+ - type: precision_at_5
225
+ value: 17.731
226
+ - type: precision_at_10
227
+ value: 9.191
228
+ - type: mrr_at_1
229
+ value: 25.371
230
+ - type: mrr_at_3
231
+ value: 50.2314
232
+ - type: mrr_at_5
233
+ value: 51.0212
234
+ - type: mrr_at_10
235
+ value: 51.4811
236
  task:
237
  type: Retrieval
238
+
239
  - dataset:
240
  config: ar-ar
241
  name: MTEB STS17 (ar-ar)