leaderboard-score / data /train-[meta-llama@Llama-2-7b-hf][Setting3][icl][answer.txt].json
zhao rui
all icl and ft
3d63e1c
raw
history blame
2.02 kB
{
"model name": "meta-llama@Llama-2-7b-hf",
"dataset": "Setting3",
"method": "icl",
"file name": "answer.txt",
"submitter": "zhaorui",
"MICRO precision": 0.8520693641618498,
"MICRO recall": 0.6086209744013212,
"MICRO f1": 0.7100578034682082,
"MACRO precision": 0.7996932592599791,
"MACRO recall": 0.5668263157750325,
"MACRO f1": 0.6634185008068373,
"detail result": "{\n \"ID\": {\n \"precision\": 0.896279905960448,\n \"recall\": 0.8320708691744768,\n \"f1\": 0.8629826897470039,\n \"support\": 7789\n },\n \"CONTACT\": {\n \"precision\": 0.7916666666666666,\n \"recall\": 0.610586011342155,\n \"f1\": 0.6894343649946638,\n \"support\": 529\n },\n \"DATE\": {\n \"precision\": 0.8691779270778289,\n \"recall\": 0.6004067372373155,\n \"f1\": 0.7102146558105107,\n \"support\": 19177\n },\n \"PROFESSION\": {\n \"precision\": 0.5560747663551402,\n \"recall\": 0.2512315270935961,\n \"f1\": 0.3460979156568105,\n \"support\": 1421\n },\n \"AGE\": {\n \"precision\": 0.7964327258627375,\n \"recall\": 0.5028151774785802,\n \"f1\": 0.6164465786314526,\n \"support\": 4085\n },\n \"LOCATION\": {\n \"precision\": 0.8683197463768116,\n \"recall\": 0.5551212450235251,\n \"f1\": 0.6772640967898618,\n \"support\": 13815\n },\n \"NAME\": {\n \"precision\": 0.8199010765202212,\n \"recall\": 0.6155526430755789,\n \"f1\": 0.7031815346225826,\n \"support\": 13734\n },\n \"MICRO_AVERAGE\": {\n \"precision\": 0.8520693641618498,\n \"recall\": 0.6086209744013212,\n \"f1\": 0.7100578034682082,\n \"support\": 60550\n },\n \"MACRO_AVERAGE\": {\n \"precision\": 0.7996932592599791,\n \"recall\": 0.5668263157750325,\n \"f1\": 0.6634185008068373,\n \"support\": 60550\n }\n}"
}