End of training
Browse files- README.md +139 -139
- model.safetensors +1 -1
README.md
CHANGED
@@ -21,21 +21,21 @@ should probably proofread and complete it, then remove this comment. -->
|
|
21 |
|
22 |
This model is a fine-tuned version of [nlpaueb/legal-bert-base-uncased](https://huggingface.co/nlpaueb/legal-bert-base-uncased) on the None dataset.
|
23 |
It achieves the following results on the evaluation set:
|
24 |
-
- Loss: 0.
|
25 |
-
- Accuracy: 0.
|
26 |
-
- Precision: 0.
|
27 |
-
- Recall: 0.
|
28 |
-
- F1: 0.
|
29 |
- Classification Report: precision recall f1-score support
|
30 |
|
31 |
LOC 0.94 0.96 0.95 1837
|
32 |
-
MISC 0.
|
33 |
-
ORG 0.
|
34 |
-
PER 0.
|
35 |
|
36 |
-
micro avg 0.92 0.93 0.
|
37 |
-
macro avg 0.91 0.
|
38 |
-
weighted avg 0.92 0.93 0.
|
39 |
|
40 |
|
41 |
## Model description
|
@@ -68,209 +68,209 @@ The following hyperparameters were used during training:
|
|
68 |
|
69 |
| Training Loss | Epoch | Step | Validation Loss | Accuracy | Precision | Recall | F1 | Classification Report |
|
70 |
|:-------------:|:------:|:----:|:---------------:|:--------:|:---------:|:------:|:------:|:--------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------:|
|
71 |
-
| 0.
|
72 |
|
73 |
-
LOC 0.
|
74 |
-
MISC 0.
|
75 |
-
ORG 0.81 0.
|
76 |
-
PER 0.
|
77 |
|
78 |
-
micro avg 0.
|
79 |
-
macro avg 0.
|
80 |
-
weighted avg 0.
|
81 |
|
|
82 |
-
| 0.
|
83 |
|
84 |
-
LOC 0.
|
85 |
-
MISC 0.89 0.
|
86 |
-
ORG 0.78 0.
|
87 |
-
PER 0.
|
88 |
|
89 |
-
micro avg 0.86 0.87 0.
|
90 |
macro avg 0.86 0.84 0.85 5942
|
91 |
weighted avg 0.86 0.87 0.86 5942
|
92 |
|
|
93 |
-
| 0.
|
94 |
|
95 |
-
LOC 0.
|
96 |
-
MISC 0.
|
97 |
-
ORG 0.
|
98 |
-
PER 0.96 0.
|
99 |
|
100 |
micro avg 0.89 0.88 0.88 5942
|
101 |
-
macro avg 0.
|
102 |
weighted avg 0.89 0.88 0.89 5942
|
103 |
|
|
104 |
-
| 0.
|
105 |
|
106 |
LOC 0.93 0.95 0.94 1837
|
107 |
-
MISC 0.
|
108 |
-
ORG 0.
|
109 |
-
PER 0.
|
110 |
-
|
111 |
-
micro avg 0.90 0.90 0.90 5942
|
112 |
-
macro avg 0.89 0.88 0.88 5942
|
113 |
-
weighted avg 0.90 0.90 0.90 5942
|
114 |
-
|
|
115 |
-
| 0.0491 | 1.3340 | 2500 | 0.0734 | 0.9814 | 0.9021 | 0.9088 | 0.9054 | precision recall f1-score support
|
116 |
-
|
117 |
-
LOC 0.92 0.95 0.93 1837
|
118 |
-
MISC 0.86 0.82 0.84 922
|
119 |
-
ORG 0.84 0.85 0.84 1341
|
120 |
-
PER 0.95 0.96 0.96 1842
|
121 |
|
122 |
micro avg 0.90 0.91 0.91 5942
|
123 |
macro avg 0.89 0.89 0.89 5942
|
124 |
weighted avg 0.90 0.91 0.91 5942
|
125 |
|
|
126 |
-
| 0.
|
127 |
|
128 |
-
LOC 0.
|
129 |
-
MISC 0.
|
130 |
-
ORG 0.
|
131 |
-
PER 0.
|
132 |
-
|
133 |
-
micro avg 0.87 0.90 0.88 5942
|
134 |
-
macro avg 0.86 0.89 0.87 5942
|
135 |
-
weighted avg 0.88 0.90 0.89 5942
|
136 |
-
|
|
137 |
-
| 0.0341 | 1.8677 | 3500 | 0.0777 | 0.9813 | 0.9072 | 0.9111 | 0.9092 | precision recall f1-score support
|
138 |
-
|
139 |
-
LOC 0.91 0.96 0.94 1837
|
140 |
-
MISC 0.87 0.84 0.85 922
|
141 |
-
ORG 0.86 0.83 0.85 1341
|
142 |
-
PER 0.95 0.96 0.95 1842
|
143 |
|
144 |
micro avg 0.91 0.91 0.91 5942
|
145 |
macro avg 0.90 0.90 0.90 5942
|
146 |
weighted avg 0.91 0.91 0.91 5942
|
147 |
|
|
148 |
-
| 0.
|
149 |
|
150 |
-
LOC 0.92 0.
|
151 |
-
MISC 0.
|
152 |
-
ORG 0.
|
153 |
-
PER 0.
|
154 |
|
155 |
-
micro avg 0.
|
156 |
-
macro avg 0.
|
157 |
-
weighted avg 0.
|
158 |
|
|
159 |
-
| 0.
|
160 |
|
161 |
-
LOC 0.
|
162 |
-
MISC 0.
|
163 |
-
ORG 0.
|
164 |
-
PER 0.
|
165 |
|
166 |
-
micro avg 0.
|
167 |
-
macro avg 0.
|
168 |
-
weighted avg 0.
|
169 |
|
|
170 |
-
| 0.
|
171 |
|
172 |
LOC 0.92 0.96 0.94 1837
|
173 |
-
MISC 0.
|
174 |
-
ORG 0.
|
175 |
-
PER 0.
|
176 |
|
177 |
-
micro avg 0.
|
178 |
-
macro avg 0.
|
179 |
-
weighted avg 0.
|
180 |
|
|
181 |
-
| 0.
|
182 |
|
183 |
-
LOC 0.
|
184 |
MISC 0.88 0.84 0.86 922
|
185 |
-
ORG 0.
|
186 |
-
PER 0.
|
187 |
|
188 |
-
micro avg 0.
|
189 |
-
macro avg 0.
|
190 |
-
weighted avg 0.
|
191 |
|
|
192 |
-
| 0.
|
193 |
|
194 |
-
LOC 0.
|
195 |
-
MISC 0.
|
196 |
ORG 0.86 0.87 0.87 1341
|
197 |
-
PER 0.94 0.
|
198 |
|
199 |
-
micro avg 0.
|
200 |
-
macro avg 0.91 0.
|
201 |
-
weighted avg 0.
|
202 |
|
|
203 |
-
| 0.
|
204 |
|
205 |
-
LOC 0.
|
206 |
-
MISC 0.89 0.86 0.
|
207 |
-
ORG 0.
|
208 |
-
PER 0.
|
209 |
|
210 |
-
micro avg 0.
|
211 |
-
macro avg 0.
|
212 |
-
weighted avg 0.
|
213 |
|
|
214 |
-
| 0.
|
215 |
|
216 |
-
LOC 0.
|
217 |
-
MISC 0.
|
218 |
-
ORG 0.
|
219 |
PER 0.95 0.97 0.96 1842
|
220 |
|
221 |
-
micro avg 0.
|
222 |
-
macro avg 0.
|
223 |
-
weighted avg 0.
|
224 |
|
|
225 |
-
| 0.
|
226 |
|
227 |
-
LOC 0.
|
228 |
-
MISC 0.
|
229 |
-
ORG 0.
|
230 |
-
PER 0.
|
231 |
|
232 |
-
micro avg 0.
|
233 |
-
macro avg 0.
|
234 |
-
weighted avg 0.
|
235 |
|
|
236 |
-
| 0.
|
237 |
|
238 |
LOC 0.94 0.96 0.95 1837
|
239 |
-
MISC 0.90 0.
|
240 |
-
ORG 0.
|
241 |
PER 0.95 0.96 0.96 1842
|
242 |
|
243 |
-
micro avg 0.92 0.
|
244 |
macro avg 0.92 0.91 0.91 5942
|
245 |
-
weighted avg 0.92 0.
|
246 |
|
|
247 |
-
| 0.
|
248 |
|
249 |
-
LOC 0.
|
250 |
-
MISC 0.89 0.86 0.
|
251 |
-
ORG 0.
|
252 |
PER 0.95 0.96 0.96 1842
|
253 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
254 |
micro avg 0.92 0.93 0.93 5942
|
255 |
macro avg 0.92 0.92 0.92 5942
|
256 |
weighted avg 0.92 0.93 0.93 5942
|
257 |
|
|
258 |
-
| 0.
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
259 |
|
260 |
LOC 0.94 0.96 0.95 1837
|
261 |
-
MISC 0.
|
262 |
-
ORG 0.
|
263 |
-
PER 0.
|
264 |
|
265 |
-
micro avg 0.92 0.93 0.
|
266 |
-
macro avg 0.91 0.
|
267 |
-
weighted avg 0.92 0.93 0.
|
268 |
|
|
269 |
|
270 |
|
271 |
### Framework versions
|
272 |
|
273 |
- Transformers 4.47.1
|
274 |
-
- Pytorch 2.
|
275 |
- Datasets 3.2.0
|
276 |
- Tokenizers 0.21.0
|
|
|
21 |
|
22 |
This model is a fine-tuned version of [nlpaueb/legal-bert-base-uncased](https://huggingface.co/nlpaueb/legal-bert-base-uncased) on the None dataset.
|
23 |
It achieves the following results on the evaluation set:
|
24 |
+
- Loss: 0.0887
|
25 |
+
- Accuracy: 0.9845
|
26 |
+
- Precision: 0.9232
|
27 |
+
- Recall: 0.9286
|
28 |
+
- F1: 0.9259
|
29 |
- Classification Report: precision recall f1-score support
|
30 |
|
31 |
LOC 0.94 0.96 0.95 1837
|
32 |
+
MISC 0.87 0.87 0.87 922
|
33 |
+
ORG 0.89 0.88 0.89 1341
|
34 |
+
PER 0.95 0.96 0.96 1842
|
35 |
|
36 |
+
micro avg 0.92 0.93 0.93 5942
|
37 |
+
macro avg 0.91 0.92 0.92 5942
|
38 |
+
weighted avg 0.92 0.93 0.93 5942
|
39 |
|
40 |
|
41 |
## Model description
|
|
|
68 |
|
69 |
| Training Loss | Epoch | Step | Validation Loss | Accuracy | Precision | Recall | F1 | Classification Report |
|
70 |
|:-------------:|:------:|:----:|:---------------:|:--------:|:---------:|:------:|:------:|:--------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------:|
|
71 |
+
| 0.1403 | 0.2668 | 500 | 0.1145 | 0.9685 | 0.8296 | 0.8334 | 0.8315 | precision recall f1-score support
|
72 |
|
73 |
+
LOC 0.85 0.93 0.89 1837
|
74 |
+
MISC 0.72 0.79 0.75 922
|
75 |
+
ORG 0.81 0.57 0.67 1341
|
76 |
+
PER 0.87 0.95 0.91 1842
|
77 |
|
78 |
+
micro avg 0.83 0.83 0.83 5942
|
79 |
+
macro avg 0.81 0.81 0.80 5942
|
80 |
+
weighted avg 0.83 0.83 0.82 5942
|
81 |
|
|
82 |
+
| 0.0978 | 0.5336 | 1000 | 0.1046 | 0.9734 | 0.8630 | 0.8679 | 0.8654 | precision recall f1-score support
|
83 |
|
84 |
+
LOC 0.85 0.94 0.89 1837
|
85 |
+
MISC 0.89 0.69 0.78 922
|
86 |
+
ORG 0.78 0.77 0.77 1341
|
87 |
+
PER 0.92 0.96 0.94 1842
|
88 |
|
89 |
+
micro avg 0.86 0.87 0.87 5942
|
90 |
macro avg 0.86 0.84 0.85 5942
|
91 |
weighted avg 0.86 0.87 0.86 5942
|
92 |
|
|
93 |
+
| 0.1052 | 0.8004 | 1500 | 0.0852 | 0.9770 | 0.8901 | 0.8782 | 0.8841 | precision recall f1-score support
|
94 |
|
95 |
+
LOC 0.94 0.91 0.92 1837
|
96 |
+
MISC 0.85 0.77 0.81 922
|
97 |
+
ORG 0.77 0.86 0.82 1341
|
98 |
+
PER 0.96 0.91 0.93 1842
|
99 |
|
100 |
micro avg 0.89 0.88 0.88 5942
|
101 |
+
macro avg 0.88 0.86 0.87 5942
|
102 |
weighted avg 0.89 0.88 0.89 5942
|
103 |
|
|
104 |
+
| 0.0532 | 1.0672 | 2000 | 0.0779 | 0.9810 | 0.9041 | 0.9091 | 0.9066 | precision recall f1-score support
|
105 |
|
106 |
LOC 0.93 0.95 0.94 1837
|
107 |
+
MISC 0.83 0.80 0.81 922
|
108 |
+
ORG 0.85 0.86 0.86 1341
|
109 |
+
PER 0.95 0.96 0.95 1842
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
110 |
|
111 |
micro avg 0.90 0.91 0.91 5942
|
112 |
macro avg 0.89 0.89 0.89 5942
|
113 |
weighted avg 0.90 0.91 0.91 5942
|
114 |
|
|
115 |
+
| 0.0507 | 1.3340 | 2500 | 0.0739 | 0.9819 | 0.9094 | 0.9073 | 0.9083 | precision recall f1-score support
|
116 |
|
117 |
+
LOC 0.95 0.93 0.94 1837
|
118 |
+
MISC 0.84 0.86 0.85 922
|
119 |
+
ORG 0.87 0.83 0.85 1341
|
120 |
+
PER 0.93 0.96 0.95 1842
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
121 |
|
122 |
micro avg 0.91 0.91 0.91 5942
|
123 |
macro avg 0.90 0.90 0.90 5942
|
124 |
weighted avg 0.91 0.91 0.91 5942
|
125 |
|
|
126 |
+
| 0.0451 | 1.6009 | 3000 | 0.0816 | 0.9791 | 0.8883 | 0.9022 | 0.8952 | precision recall f1-score support
|
127 |
|
128 |
+
LOC 0.92 0.94 0.93 1837
|
129 |
+
MISC 0.81 0.81 0.81 922
|
130 |
+
ORG 0.80 0.88 0.84 1341
|
131 |
+
PER 0.96 0.93 0.95 1842
|
132 |
|
133 |
+
micro avg 0.89 0.90 0.90 5942
|
134 |
+
macro avg 0.88 0.89 0.88 5942
|
135 |
+
weighted avg 0.89 0.90 0.90 5942
|
136 |
|
|
137 |
+
| 0.0397 | 1.8677 | 3500 | 0.0755 | 0.9812 | 0.9033 | 0.9135 | 0.9084 | precision recall f1-score support
|
138 |
|
139 |
+
LOC 0.90 0.96 0.93 1837
|
140 |
+
MISC 0.83 0.86 0.85 922
|
141 |
+
ORG 0.88 0.83 0.86 1341
|
142 |
+
PER 0.95 0.95 0.95 1842
|
143 |
|
144 |
+
micro avg 0.90 0.91 0.91 5942
|
145 |
+
macro avg 0.89 0.90 0.90 5942
|
146 |
+
weighted avg 0.90 0.91 0.91 5942
|
147 |
|
|
148 |
+
| 0.0211 | 2.1345 | 4000 | 0.0895 | 0.9814 | 0.9173 | 0.9130 | 0.9151 | precision recall f1-score support
|
149 |
|
150 |
LOC 0.92 0.96 0.94 1837
|
151 |
+
MISC 0.88 0.83 0.86 922
|
152 |
+
ORG 0.88 0.86 0.87 1341
|
153 |
+
PER 0.96 0.95 0.95 1842
|
154 |
|
155 |
+
micro avg 0.92 0.91 0.92 5942
|
156 |
+
macro avg 0.91 0.90 0.90 5942
|
157 |
+
weighted avg 0.92 0.91 0.91 5942
|
158 |
|
|
159 |
+
| 0.0224 | 2.4013 | 4500 | 0.0840 | 0.9815 | 0.9005 | 0.9110 | 0.9057 | precision recall f1-score support
|
160 |
|
161 |
+
LOC 0.92 0.95 0.93 1837
|
162 |
MISC 0.88 0.84 0.86 922
|
163 |
+
ORG 0.82 0.88 0.85 1341
|
164 |
+
PER 0.96 0.93 0.94 1842
|
165 |
|
166 |
+
micro avg 0.90 0.91 0.91 5942
|
167 |
+
macro avg 0.89 0.90 0.90 5942
|
168 |
+
weighted avg 0.90 0.91 0.91 5942
|
169 |
|
|
170 |
+
| 0.0285 | 2.6681 | 5000 | 0.0770 | 0.9823 | 0.9143 | 0.9157 | 0.9150 | precision recall f1-score support
|
171 |
|
172 |
+
LOC 0.93 0.94 0.94 1837
|
173 |
+
MISC 0.89 0.83 0.86 922
|
174 |
ORG 0.86 0.87 0.87 1341
|
175 |
+
PER 0.94 0.97 0.95 1842
|
176 |
|
177 |
+
micro avg 0.91 0.92 0.91 5942
|
178 |
+
macro avg 0.91 0.90 0.90 5942
|
179 |
+
weighted avg 0.91 0.92 0.91 5942
|
180 |
|
|
181 |
+
| 0.0256 | 2.9349 | 5500 | 0.0779 | 0.9840 | 0.9260 | 0.9244 | 0.9252 | precision recall f1-score support
|
182 |
|
183 |
+
LOC 0.94 0.95 0.94 1837
|
184 |
+
MISC 0.89 0.86 0.87 922
|
185 |
+
ORG 0.90 0.87 0.89 1341
|
186 |
+
PER 0.95 0.97 0.96 1842
|
187 |
|
188 |
+
micro avg 0.93 0.92 0.93 5942
|
189 |
+
macro avg 0.92 0.91 0.92 5942
|
190 |
+
weighted avg 0.93 0.92 0.92 5942
|
191 |
|
|
192 |
+
| 0.0135 | 3.2017 | 6000 | 0.0878 | 0.9831 | 0.9111 | 0.9229 | 0.9170 | precision recall f1-score support
|
193 |
|
194 |
+
LOC 0.93 0.95 0.94 1837
|
195 |
+
MISC 0.82 0.87 0.84 922
|
196 |
+
ORG 0.89 0.86 0.87 1341
|
197 |
PER 0.95 0.97 0.96 1842
|
198 |
|
199 |
+
micro avg 0.91 0.92 0.92 5942
|
200 |
+
macro avg 0.90 0.91 0.90 5942
|
201 |
+
weighted avg 0.91 0.92 0.92 5942
|
202 |
|
|
203 |
+
| 0.0112 | 3.4685 | 6500 | 0.0845 | 0.9835 | 0.9131 | 0.9265 | 0.9197 | precision recall f1-score support
|
204 |
|
205 |
+
LOC 0.93 0.96 0.94 1837
|
206 |
+
MISC 0.85 0.88 0.86 922
|
207 |
+
ORG 0.87 0.88 0.87 1341
|
208 |
+
PER 0.96 0.96 0.96 1842
|
209 |
|
210 |
+
micro avg 0.91 0.93 0.92 5942
|
211 |
+
macro avg 0.90 0.92 0.91 5942
|
212 |
+
weighted avg 0.91 0.93 0.92 5942
|
213 |
|
|
214 |
+
| 0.0119 | 3.7353 | 7000 | 0.0901 | 0.9837 | 0.9230 | 0.9219 | 0.9225 | precision recall f1-score support
|
215 |
|
216 |
LOC 0.94 0.96 0.95 1837
|
217 |
+
MISC 0.90 0.83 0.86 922
|
218 |
+
ORG 0.88 0.88 0.88 1341
|
219 |
PER 0.95 0.96 0.96 1842
|
220 |
|
221 |
+
micro avg 0.92 0.92 0.92 5942
|
222 |
macro avg 0.92 0.91 0.91 5942
|
223 |
+
weighted avg 0.92 0.92 0.92 5942
|
224 |
|
|
225 |
+
| 0.0132 | 4.0021 | 7500 | 0.0902 | 0.9843 | 0.9262 | 0.9248 | 0.9255 | precision recall f1-score support
|
226 |
|
227 |
+
LOC 0.93 0.96 0.95 1837
|
228 |
+
MISC 0.89 0.86 0.87 922
|
229 |
+
ORG 0.91 0.87 0.89 1341
|
230 |
PER 0.95 0.96 0.96 1842
|
231 |
|
232 |
+
micro avg 0.93 0.92 0.93 5942
|
233 |
+
macro avg 0.92 0.91 0.92 5942
|
234 |
+
weighted avg 0.93 0.92 0.93 5942
|
235 |
+
|
|
236 |
+
| 0.006 | 4.2689 | 8000 | 0.0914 | 0.9844 | 0.9233 | 0.9273 | 0.9253 | precision recall f1-score support
|
237 |
+
|
238 |
+
LOC 0.94 0.96 0.95 1837
|
239 |
+
MISC 0.88 0.86 0.87 922
|
240 |
+
ORG 0.88 0.89 0.88 1341
|
241 |
+
PER 0.96 0.96 0.96 1842
|
242 |
+
|
243 |
micro avg 0.92 0.93 0.93 5942
|
244 |
macro avg 0.92 0.92 0.92 5942
|
245 |
weighted avg 0.92 0.93 0.93 5942
|
246 |
|
|
247 |
+
| 0.005 | 4.5358 | 8500 | 0.0919 | 0.9846 | 0.9284 | 0.9268 | 0.9276 | precision recall f1-score support
|
248 |
+
|
249 |
+
LOC 0.95 0.96 0.95 1837
|
250 |
+
MISC 0.90 0.85 0.87 922
|
251 |
+
ORG 0.90 0.88 0.89 1341
|
252 |
+
PER 0.95 0.97 0.96 1842
|
253 |
+
|
254 |
+
micro avg 0.93 0.93 0.93 5942
|
255 |
+
macro avg 0.92 0.91 0.92 5942
|
256 |
+
weighted avg 0.93 0.93 0.93 5942
|
257 |
+
|
|
258 |
+
| 0.0062 | 4.8026 | 9000 | 0.0887 | 0.9845 | 0.9232 | 0.9286 | 0.9259 | precision recall f1-score support
|
259 |
|
260 |
LOC 0.94 0.96 0.95 1837
|
261 |
+
MISC 0.87 0.87 0.87 922
|
262 |
+
ORG 0.89 0.88 0.89 1341
|
263 |
+
PER 0.95 0.96 0.96 1842
|
264 |
|
265 |
+
micro avg 0.92 0.93 0.93 5942
|
266 |
+
macro avg 0.91 0.92 0.92 5942
|
267 |
+
weighted avg 0.92 0.93 0.93 5942
|
268 |
|
|
269 |
|
270 |
|
271 |
### Framework versions
|
272 |
|
273 |
- Transformers 4.47.1
|
274 |
+
- Pytorch 2.5.1+cpu
|
275 |
- Datasets 3.2.0
|
276 |
- Tokenizers 0.21.0
|
model.safetensors
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 435617620
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:165e72e887066034c0e8b42aa5da40b8eecf0035af54db60ea150c7221090277
|
3 |
size 435617620
|