Update README.md
Browse files
README.md
CHANGED
@@ -75,22 +75,21 @@ For reference, Warner et al.'s ModernBERT uses 1.72T tokens for stage 1, 250B to
|
|
75 |
|
76 |
## Evaluation
|
77 |
|
78 |
-
|
79 |
-
|
80 |
Evaluation code can be found at https://github.com/speed1313/bert-eval
|
81 |
|
82 |
-
| Model
|
83 |
-
|
84 |
-
| tohoku-nlp/bert-base-japanese-v3
|
85 |
-
| sbintuitions/modernbert-ja-130m
|
86 |
-
| sbintuitions/modernbert-ja-310m
|
87 |
-
| llm-jp-modernbert-base-v3-stage1-500k
|
88 |
-
| llm-jp-modernbert-base-v3-stage2-200k
|
89 |
-
| llm-jp-modernbert-base-v4-ja-stage1-100k
|
90 |
-
| llm-jp-modernbert-base-v4-ja-stage1-
|
91 |
-
| llm-jp-modernbert-base-v4-ja-stage1-
|
92 |
-
| llm-jp-modernbert-base-v4-ja-stage1-
|
93 |
-
| llm-jp-modernbert-base-v4-ja-
|
94 |
-
|
95 |
|
96 |
|
|
|
75 |
|
76 |
## Evaluation
|
77 |
|
78 |
+
JSTS, JNLI, and JCoLA from [JGLUE](https://aclanthology.org/2022.lrec-1.317/) were used.
|
|
|
79 |
Evaluation code can be found at https://github.com/speed1313/bert-eval
|
80 |
|
81 |
+
| Model | JSTS (pearson) | JNLI (accuracy) | JCoLA(accuracy) | Avg |
|
82 |
+
|-------------------------------------------------------|--------|--------|---------|--------------|
|
83 |
+
| tohoku-nlp/bert-base-japanese-v3 | 0.920 | 0.912 | 0.880 | 0.904 |
|
84 |
+
| sbintuitions/modernbert-ja-130m | 0.916 | 0.927 | 0.868 | 0.904 |
|
85 |
+
| sbintuitions/modernbert-ja-310m | 0.932 | 0.933 | 0.883 | 0.916 |
|
86 |
+
| speed/llm-jp-modernbert-base-v3-ja-stage1-500k | 0.925 | 0.917 | 0.856 | 0.899 |
|
87 |
+
| speed/llm-jp-modernbert-base-v3-ja-stage2-200k | 0.924 | 0.911 | 0.844 | 0.893 |
|
88 |
+
| speed/llm-jp-modernbert-base-v4-ja-stage1-100k | 0.921 | 0.918 | 0.861 | 0.900 |
|
89 |
+
| speed/llm-jp-modernbert-base-v4-ja-stage1-200k | 0.920 | 0.927 | 0.850 | 0.899 |
|
90 |
+
| speed/llm-jp-modernbert-base-v4-ja-stage1-300k | 0.920 | 0.919 | 0.852 | 0.897 |
|
91 |
+
| speed/llm-jp-modernbert-base-v4-ja-stage1-400k | 0.921 | 0.920 | 0.856 | 0.899 |
|
92 |
+
| speed/llm-jp-modernbert-base-v4-ja-stage1-500k | 0.921 | 0.919 | 0.845 | 0.895 |
|
93 |
+
| speed/llm-jp-modernbert-base-v4-ja-stage2-200k | 0.918 | 0.913 | 0.844 | 0.892 |
|
94 |
|
95 |
|