Dataset Viewer
model name
stringclasses 3
values | dataset
stringclasses 2
values | method
stringclasses 1
value | file name
stringclasses 1
value | submitter
stringclasses 4
values | MICRO precision
float64 0.84
0.96
| MICRO recall
float64 0.45
0.66
| MICRO f1
float64 0.61
0.74
| MACRO precision
float64 0.75
0.81
| MACRO recall
float64 0.31
0.63
| MACRO f1
float64 0.44
0.71
| detail result
dict |
---|---|---|---|---|---|---|---|---|---|---|---|
EleutherAI@pythia-1b | Setting3 | icl | answer.txt | sdf | 0.861082 | 0.552898 | 0.673405 | 0.814266 | 0.491727 | 0.613168 | {
"NAME": {
"precision": 0.8224498506188647,
"recall": 0.5612348915101208,
"f1": 0.6671860122911798,
"support": 13734
},
"AGE": {
"precision": 0.8147044212617983,
"recall": 0.401468788249694,
"f1": 0.5378812725483766,
"support": 4085
},
"CONTACT": {
"precision": 0.8852459016393442,
"recall": 0.5103969754253308,
"f1": 0.6474820143884891,
"support": 529
},
"ID": {
"precision": 0.87248322147651,
"recall": 0.7844395942996534,
"f1": 0.8261222282314765,
"support": 7789
},
"LOCATION": {
"precision": 0.8922374429223744,
"recall": 0.49504162142598623,
"f1": 0.6367783985102421,
"support": 13815
},
"DATE": {
"precision": 0.8811303704920727,
"recall": 0.5593158471085155,
"f1": 0.684274322169059,
"support": 19177
},
"PROFESSION": {
"precision": 0.5316091954022989,
"recall": 0.1301900070372977,
"f1": 0.2091577162238553,
"support": 1421
},
"MICRO_AVERAGE": {
"precision": 0.8610818179479925,
"recall": 0.55289843104872,
"f1": 0.673405143368635,
"support": 60550
},
"MACRO_AVERAGE": {
"precision": 0.8142657719733233,
"recall": 0.49172681786522837,
"f1": 0.613167823560949,
"support": 60550
}
} |
meta-llama@Llama-2-7b-hf | Setting1 | icl | answer.txt | test | 0.958962 | 0.447366 | 0.61011 | 0.748221 | 0.309891 | 0.438266 | {
"NAME": {
"precision": 0.9550425273390036,
"recall": 0.46338875132649454,
"f1": 0.6240076214671324,
"support": 8481
},
"AGE": {
"precision": 1,
"recall": 0.17857142857142858,
"f1": 0.30303030303030304,
"support": 140
},
"CONTACT": {
"precision": 0.4,
"recall": 0.2,
"f1": 0.26666666666666666,
"support": 10
},
"ID": {
"precision": 0.9278477884852977,
"recall": 0.41950620042453357,
"f1": 0.5777811971072472,
"support": 8951
},
"LOCATION": {
"precision": 0.9740314136125654,
"recall": 0.4717995536620004,
"f1": 0.6356864621061984,
"support": 9858
},
"DATE": {
"precision": 0.9806224310041104,
"recall": 0.4359744158726015,
"f1": 0.6035962772205657,
"support": 7661
},
"PROFESSION": {
"precision": 0,
"recall": 0,
"f1": 0,
"support": 0
},
"MICRO_AVERAGE": {
"precision": 0.9589618320610687,
"recall": 0.44736617190393435,
"f1": 0.6101095656228145,
"support": 35101
},
"MACRO_AVERAGE": {
"precision": 0.7482205943487109,
"recall": 0.30989147855100835,
"f1": 0.4382658363014653,
"support": 35101
}
} |
meta-llama@Llama-2-7b-hf | Setting3 | icl | answer.txt | rui | 0.852069 | 0.608621 | 0.710058 | 0.799693 | 0.566826 | 0.663419 | {
"PROFESSION": {
"precision": 0.5560747663551402,
"recall": 0.2512315270935961,
"f1": 0.3460979156568105,
"support": 1421
},
"LOCATION": {
"precision": 0.8683197463768116,
"recall": 0.5551212450235251,
"f1": 0.6772640967898618,
"support": 13815
},
"DATE": {
"precision": 0.8691779270778289,
"recall": 0.6004067372373155,
"f1": 0.7102146558105107,
"support": 19177
},
"CONTACT": {
"precision": 0.7916666666666666,
"recall": 0.610586011342155,
"f1": 0.6894343649946638,
"support": 529
},
"NAME": {
"precision": 0.8199010765202212,
"recall": 0.6155526430755789,
"f1": 0.7031815346225826,
"support": 13734
},
"AGE": {
"precision": 0.7964327258627375,
"recall": 0.5028151774785802,
"f1": 0.6164465786314526,
"support": 4085
},
"ID": {
"precision": 0.896279905960448,
"recall": 0.8320708691744768,
"f1": 0.8629826897470039,
"support": 7789
},
"MICRO_AVERAGE": {
"precision": 0.8520693641618498,
"recall": 0.6086209744013212,
"f1": 0.7100578034682082,
"support": 60550
},
"MACRO_AVERAGE": {
"precision": 0.7996932592599791,
"recall": 0.5668263157750325,
"f1": 0.6634185008068373,
"support": 60550
}
} |
Setting3 | icl | answer.txt | zhaorui | 0.840219 | 0.657688 | 0.737832 | 0.80437 | 0.634122 | 0.709172 | {
"LOCATION": {
"precision": 0.8358999509563512,
"recall": 0.6168657256605139,
"f1": 0.7098708871303623,
"support": 13815
},
"CONTACT": {
"precision": 0.8289156626506025,
"recall": 0.6502835538752363,
"f1": 0.728813559322034,
"support": 529
},
"DATE": {
"precision": 0.8623968675709691,
"recall": 0.6431662929551024,
"f1": 0.7368200961796948,
"support": 19177
},
"NAME": {
"precision": 0.8296005972377752,
"recall": 0.6472986748216106,
"f1": 0.7271983640081799,
"support": 13734
},
"ID": {
"precision": 0.8709163346613545,
"recall": 0.8419566054692516,
"f1": 0.8561916574188915,
"support": 7789
},
"AGE": {
"precision": 0.7862318840579711,
"recall": 0.6374541003671971,
"f1": 0.7040692172502365,
"support": 4085
},
"PROFESSION": {
"precision": 0.6166306695464363,
"recall": 0.40182969739619984,
"f1": 0.4865786109927567,
"support": 1421
},
"MICRO_AVERAGE": {
"precision": 0.8402185838467381,
"recall": 0.6576878612716763,
"f1": 0.7378318789024142,
"support": 60550
},
"MACRO_AVERAGE": {
"precision": 0.8043702809544941,
"recall": 0.634122092935016,
"f1": 0.7091715956399902,
"support": 60550
}
} |
README.md exists but content is empty.
- Downloads last month
- 44