raphael0202
commited on
Commit
•
8b7b797
1
Parent(s):
dea426b
Model save
Browse files- README.md +222 -341
- config.json +141 -141
- model.safetensors +1 -1
- runs/Dec23_15-55-22_gpu-l4-20241223-150914/events.out.tfevents.1734969342.gpu-l4-20241223-150914 +3 -0
- tokenizer_config.json +0 -7
- training_args.bin +3 -0
README.md
CHANGED
@@ -1,6 +1,5 @@
|
|
1 |
---
|
2 |
license: cc-by-nc-sa-4.0
|
3 |
-
library_name: transformers
|
4 |
base_model: microsoft/layoutlmv3-large
|
5 |
tags:
|
6 |
- generated_from_trainer
|
@@ -9,153 +8,35 @@ metrics:
|
|
9 |
- recall
|
10 |
- f1
|
11 |
- accuracy
|
12 |
-
datasets:
|
13 |
-
- openfoodfacts/nutrient-detection-layout
|
14 |
model-index:
|
15 |
- name: nutrition-extractor
|
16 |
results: []
|
17 |
---
|
18 |
|
|
|
|
|
|
|
19 |
# nutrition-extractor
|
20 |
|
21 |
-
This model is a fine-tuned version of [microsoft/layoutlmv3-large](https://huggingface.co/microsoft/layoutlmv3-large) on
|
22 |
-
[v5 of the nutrient extraction dataset](https://huggingface.co/datasets/openfoodfacts/nutrient-detection-layout).
|
23 |
It achieves the following results on the evaluation set:
|
24 |
-
-
|
25 |
-
-
|
26 |
-
-
|
27 |
-
-
|
28 |
-
-
|
29 |
-
|
30 |
-
|
31 |
-
|
32 |
-
|
33 |
-
|
34 |
-
|
35 |
-
|
36 |
-
|
37 |
-
|
38 |
-
|
39 |
-
- eval_CARBOHYDRATES_100G_recall: 0.9716
|
40 |
-
- eval_CARBOHYDRATES_100G_f1: 0.9580
|
41 |
-
- eval_CARBOHYDRATES_100G_number: 176
|
42 |
-
- eval_CARBOHYDRATES_SERVING_precision: 0.9242
|
43 |
-
- eval_CARBOHYDRATES_SERVING_recall: 0.8841
|
44 |
-
- eval_CARBOHYDRATES_SERVING_f1: 0.9037
|
45 |
-
- eval_CARBOHYDRATES_SERVING_number: 69
|
46 |
-
- eval_CHOLESTEROL_SERVING_precision: 1.0
|
47 |
-
- eval_CHOLESTEROL_SERVING_recall: 1.0
|
48 |
-
- eval_CHOLESTEROL_SERVING_f1: 1.0
|
49 |
-
- eval_CHOLESTEROL_SERVING_number: 7
|
50 |
-
- eval_ENERGY_KCAL_100G_precision: 0.9771
|
51 |
-
- eval_ENERGY_KCAL_100G_recall: 0.9884
|
52 |
-
- eval_ENERGY_KCAL_100G_f1: 0.9828
|
53 |
-
- eval_ENERGY_KCAL_100G_number: 173
|
54 |
-
- eval_ENERGY_KCAL_SERVING_precision: 0.8971
|
55 |
-
- eval_ENERGY_KCAL_SERVING_recall: 0.9385
|
56 |
-
- eval_ENERGY_KCAL_SERVING_f1: 0.9173
|
57 |
-
- eval_ENERGY_KCAL_SERVING_number: 65
|
58 |
-
- eval_ENERGY_KJ_100G_precision: 0.9670
|
59 |
-
- eval_ENERGY_KJ_100G_recall: 0.9617
|
60 |
-
- eval_ENERGY_KJ_100G_f1: 0.9644
|
61 |
-
- eval_ENERGY_KJ_100G_number: 183
|
62 |
-
- eval_ENERGY_KJ_SERVING_precision: 0.9194
|
63 |
-
- eval_ENERGY_KJ_SERVING_recall: 1.0
|
64 |
-
- eval_ENERGY_KJ_SERVING_f1: 0.9580
|
65 |
-
- eval_ENERGY_KJ_SERVING_number: 57
|
66 |
-
- eval_FAT_100G_precision: 0.9611
|
67 |
-
- eval_FAT_100G_recall: 0.9505
|
68 |
-
- eval_FAT_100G_f1: 0.9558
|
69 |
-
- eval_FAT_100G_number: 182
|
70 |
-
- eval_FAT_SERVING_precision: 0.9403
|
71 |
-
- eval_FAT_SERVING_recall: 0.9545
|
72 |
-
- eval_FAT_SERVING_f1: 0.9474
|
73 |
-
- eval_FAT_SERVING_number: 66
|
74 |
-
- eval_FIBER_100G_precision: 0.8966
|
75 |
-
- eval_FIBER_100G_recall: 0.9286
|
76 |
-
- eval_FIBER_100G_f1: 0.9123
|
77 |
-
- eval_FIBER_100G_number: 84
|
78 |
-
- eval_FIBER_SERVING_precision: 0.8654
|
79 |
-
- eval_FIBER_SERVING_recall: 0.9
|
80 |
-
- eval_FIBER_SERVING_f1: 0.8824
|
81 |
-
- eval_FIBER_SERVING_number: 50
|
82 |
-
- eval_IRON_SERVING_precision: 0.0
|
83 |
-
- eval_IRON_SERVING_recall: 0.0
|
84 |
-
- eval_IRON_SERVING_f1: 0.0
|
85 |
-
- eval_IRON_SERVING_number: 2
|
86 |
-
- eval_POTASSIUM_SERVING_precision: 0.8333
|
87 |
-
- eval_POTASSIUM_SERVING_recall: 1.0
|
88 |
-
- eval_POTASSIUM_SERVING_f1: 0.9091
|
89 |
-
- eval_POTASSIUM_SERVING_number: 5
|
90 |
-
- eval_PROTEINS_100G_precision: 0.9492
|
91 |
-
- eval_PROTEINS_100G_recall: 0.96
|
92 |
-
- eval_PROTEINS_100G_f1: 0.9545
|
93 |
-
- eval_PROTEINS_100G_number: 175
|
94 |
-
- eval_PROTEINS_SERVING_precision: 0.9375
|
95 |
-
- eval_PROTEINS_SERVING_recall: 0.9375
|
96 |
-
- eval_PROTEINS_SERVING_f1: 0.9375
|
97 |
-
- eval_PROTEINS_SERVING_number: 64
|
98 |
-
- eval_SALT_100G_precision: 0.9709
|
99 |
-
- eval_SALT_100G_recall: 0.9709
|
100 |
-
- eval_SALT_100G_f1: 0.9709
|
101 |
-
- eval_SALT_100G_number: 172
|
102 |
-
- eval_SALT_SERVING_precision: 0.9057
|
103 |
-
- eval_SALT_SERVING_recall: 0.96
|
104 |
-
- eval_SALT_SERVING_f1: 0.9320
|
105 |
-
- eval_SALT_SERVING_number: 50
|
106 |
-
- eval_SATURATED_FAT_100G_precision: 0.9497
|
107 |
-
- eval_SATURATED_FAT_100G_recall: 0.9659
|
108 |
-
- eval_SATURATED_FAT_100G_f1: 0.9577
|
109 |
-
- eval_SATURATED_FAT_100G_number: 176
|
110 |
-
- eval_SATURATED_FAT_SERVING_precision: 0.9672
|
111 |
-
- eval_SATURATED_FAT_SERVING_recall: 0.9516
|
112 |
-
- eval_SATURATED_FAT_SERVING_f1: 0.9593
|
113 |
-
- eval_SATURATED_FAT_SERVING_number: 62
|
114 |
-
- eval_SERVING_SIZE_precision: 0.9104
|
115 |
-
- eval_SERVING_SIZE_recall: 0.8841
|
116 |
-
- eval_SERVING_SIZE_f1: 0.8971
|
117 |
-
- eval_SERVING_SIZE_number: 69
|
118 |
-
- eval_SODIUM_100G_precision: 0.6667
|
119 |
-
- eval_SODIUM_100G_recall: 0.6667
|
120 |
-
- eval_SODIUM_100G_f1: 0.6667
|
121 |
-
- eval_SODIUM_100G_number: 3
|
122 |
-
- eval_SODIUM_SERVING_precision: 0.9286
|
123 |
-
- eval_SODIUM_SERVING_recall: 0.9286
|
124 |
-
- eval_SODIUM_SERVING_f1: 0.9286
|
125 |
-
- eval_SODIUM_SERVING_number: 14
|
126 |
-
- eval_SUGARS_100G_precision: 0.9368
|
127 |
-
- eval_SUGARS_100G_recall: 0.9477
|
128 |
-
- eval_SUGARS_100G_f1: 0.9422
|
129 |
-
- eval_SUGARS_100G_number: 172
|
130 |
-
- eval_SUGARS_SERVING_precision: 0.8730
|
131 |
-
- eval_SUGARS_SERVING_recall: 0.8594
|
132 |
-
- eval_SUGARS_SERVING_f1: 0.8661
|
133 |
-
- eval_SUGARS_SERVING_number: 64
|
134 |
-
- eval_TRANS_FAT_100G_precision: 0.0
|
135 |
-
- eval_TRANS_FAT_100G_recall: 0.0
|
136 |
-
- eval_TRANS_FAT_100G_f1: 0.0
|
137 |
-
- eval_TRANS_FAT_100G_number: 2
|
138 |
-
- eval_TRANS_FAT_SERVING_precision: 0.7143
|
139 |
-
- eval_TRANS_FAT_SERVING_recall: 1.0
|
140 |
-
- eval_TRANS_FAT_SERVING_f1: 0.8333
|
141 |
-
- eval_TRANS_FAT_SERVING_number: 5
|
142 |
-
- eval_VITAMIN_D_100G_precision: 0.0
|
143 |
-
- eval_VITAMIN_D_100G_recall: 0.0
|
144 |
-
- eval_VITAMIN_D_100G_f1: 0.0
|
145 |
-
- eval_VITAMIN_D_100G_number: 2
|
146 |
-
- eval_VITAMIN_D_SERVING_precision: 1.0
|
147 |
-
- eval_VITAMIN_D_SERVING_recall: 1.0
|
148 |
-
- eval_VITAMIN_D_SERVING_f1: 1.0
|
149 |
-
- eval_VITAMIN_D_SERVING_number: 2
|
150 |
-
- eval_overall_precision: 0.9400
|
151 |
-
- eval_overall_recall: 0.9478
|
152 |
-
- eval_overall_f1: 0.9439
|
153 |
-
- eval_overall_accuracy: 0.9892
|
154 |
-
- eval_runtime: 2159.9991
|
155 |
-
- eval_samples_per_second: 0.093
|
156 |
-
- eval_steps_per_second: 0.012
|
157 |
|
158 |
-
|
159 |
|
160 |
## Training procedure
|
161 |
|
@@ -177,211 +58,211 @@ The following hyperparameters were used during training:
|
|
177 |
|
178 |
| Training Loss | Epoch | Step | Validation Loss | Precision | Recall | F1 | Accuracy |
|
179 |
|:-------------:|:-------:|:----:|:---------------:|:---------:|:------:|:------:|:--------:|
|
180 |
-
| 1.
|
181 |
-
|
|
182 |
-
| 0.
|
183 |
-
| 0.
|
184 |
-
| 0.
|
185 |
-
| 0.
|
186 |
-
| 0.
|
187 |
-
| 0.
|
188 |
-
| 0.
|
189 |
-
| 0.
|
190 |
-
| 0.
|
191 |
-
| 0.
|
192 |
-
| 0.
|
193 |
-
| 0.
|
194 |
-
| 0.
|
195 |
-
| 0.
|
196 |
-
| 0.
|
197 |
-
| 0.
|
198 |
-
| 0.
|
199 |
-
| 0.
|
200 |
-
| 0.
|
201 |
-
| 0.
|
202 |
-
| 0.
|
203 |
-
| 0.
|
204 |
-
| 0.
|
205 |
-
| 0.1313 | 4.
|
206 |
-
| 0.
|
207 |
-
| 0.
|
208 |
-
| 0.
|
209 |
-
| 0.
|
210 |
-
| 0.
|
211 |
-
| 0.
|
212 |
-
| 0.
|
213 |
-
| 0.
|
214 |
-
| 0.
|
215 |
-
| 0.
|
216 |
-
| 0.
|
217 |
-
| 0.
|
218 |
-
| 0.
|
219 |
-
| 0.
|
220 |
-
| 0.
|
221 |
-
| 0.
|
222 |
-
| 0.
|
223 |
-
| 0.
|
224 |
-
| 0.
|
225 |
-
| 0.
|
226 |
-
| 0.
|
227 |
-
| 0.
|
228 |
-
| 0.
|
229 |
-
| 0.
|
230 |
-
| 0.
|
231 |
-
| 0.
|
232 |
-
| 0.
|
233 |
-
| 0.
|
234 |
-
| 0.0435 |
|
235 |
-
| 0.
|
236 |
-
| 0.
|
237 |
-
| 0.
|
238 |
-
| 0.
|
239 |
-
| 0.
|
240 |
-
| 0.
|
241 |
-
| 0.
|
242 |
-
| 0.
|
243 |
-
| 0.
|
244 |
-
| 0.
|
245 |
-
| 0.
|
246 |
-
| 0.
|
247 |
-
| 0.
|
248 |
-
| 0.
|
249 |
-
| 0.
|
250 |
-
| 0.
|
251 |
-
| 0.
|
252 |
-
| 0.
|
253 |
-
| 0.
|
254 |
-
| 0.
|
255 |
-
| 0.
|
256 |
-
| 0.
|
257 |
-
| 0.
|
258 |
-
| 0.
|
259 |
-
| 0.
|
260 |
-
| 0.
|
261 |
-
| 0.
|
262 |
-
| 0.
|
263 |
-
| 0.
|
264 |
-
| 0.
|
265 |
-
| 0.
|
266 |
-
| 0.
|
267 |
-
| 0.
|
268 |
-
| 0.
|
269 |
-
| 0.
|
270 |
-
| 0.
|
271 |
-
| 0.
|
272 |
-
| 0.
|
273 |
-
| 0.
|
274 |
-
| 0.
|
275 |
-
| 0.
|
276 |
-
| 0.
|
277 |
-
| 0.
|
278 |
-
| 0.
|
279 |
-
| 0.
|
280 |
-
| 0.
|
281 |
-
| 0.
|
282 |
-
| 0.
|
283 |
-
| 0.
|
284 |
-
| 0.
|
285 |
-
| 0.
|
286 |
-
| 0.
|
287 |
-
| 0.
|
288 |
-
| 0.
|
289 |
-
| 0.
|
290 |
-
| 0.
|
291 |
-
| 0.
|
292 |
-
| 0.
|
293 |
-
| 0.
|
294 |
-
| 0.
|
295 |
-
| 0.
|
296 |
-
| 0.
|
297 |
-
| 0.
|
298 |
-
| 0.
|
299 |
-
| 0.0126 |
|
300 |
-
| 0.
|
301 |
-
| 0.
|
302 |
-
| 0.
|
303 |
-
| 0.
|
304 |
-
| 0.
|
305 |
-
| 0.0122 |
|
306 |
-
| 0.
|
307 |
-
| 0.
|
308 |
-
| 0.
|
309 |
-
| 0.
|
310 |
-
| 0.
|
311 |
-
| 0.
|
312 |
-
| 0.
|
313 |
-
| 0.
|
314 |
-
| 0.
|
315 |
-
| 0.
|
316 |
-
| 0.
|
317 |
-
| 0.
|
318 |
-
| 0.
|
319 |
-
| 0.
|
320 |
-
| 0.
|
321 |
-
| 0.
|
322 |
-
| 0.
|
323 |
-
| 0.
|
324 |
-
| 0.
|
325 |
-
| 0.
|
326 |
-
| 0.
|
327 |
-
| 0.
|
328 |
-
| 0.0088 |
|
329 |
-
| 0.
|
330 |
-
| 0.
|
331 |
-
| 0.
|
332 |
-
| 0.
|
333 |
-
| 0.
|
334 |
-
| 0.
|
335 |
-
| 0.
|
336 |
-
| 0.
|
337 |
-
| 0.
|
338 |
-
| 0.
|
339 |
-
| 0.
|
340 |
-
| 0.
|
341 |
-
| 0.
|
342 |
-
| 0.
|
343 |
-
| 0.0073 |
|
344 |
-
| 0.
|
345 |
-
| 0.
|
346 |
-
| 0.
|
347 |
-
| 0.
|
348 |
-
| 0.
|
349 |
-
| 0.
|
350 |
-
| 0.
|
351 |
-
| 0.
|
352 |
-
| 0.
|
353 |
-
| 0.
|
354 |
-
| 0.
|
355 |
-
| 0.
|
356 |
-
| 0.
|
357 |
-
| 0.
|
358 |
-
| 0.
|
359 |
-
| 0.
|
360 |
-
| 0.
|
361 |
-
| 0.0064 |
|
362 |
-
| 0.0055 |
|
363 |
-
| 0.
|
364 |
-
| 0.
|
365 |
-
| 0.
|
366 |
-
| 0.
|
367 |
-
| 0.
|
368 |
-
| 0.
|
369 |
-
| 0.
|
370 |
-
| 0.
|
371 |
-
| 0.
|
372 |
-
| 0.
|
373 |
-
| 0.
|
374 |
-
| 0.
|
375 |
-
| 0.
|
376 |
-
| 0.
|
377 |
-
| 0.
|
378 |
-
| 0.
|
379 |
-
| 0.
|
380 |
|
381 |
|
382 |
### Framework versions
|
383 |
|
384 |
- Transformers 4.40.2
|
385 |
-
- Pytorch 2.
|
386 |
- Datasets 2.19.0
|
387 |
-
- Tokenizers 0.19.1
|
|
|
1 |
---
|
2 |
license: cc-by-nc-sa-4.0
|
|
|
3 |
base_model: microsoft/layoutlmv3-large
|
4 |
tags:
|
5 |
- generated_from_trainer
|
|
|
8 |
- recall
|
9 |
- f1
|
10 |
- accuracy
|
|
|
|
|
11 |
model-index:
|
12 |
- name: nutrition-extractor
|
13 |
results: []
|
14 |
---
|
15 |
|
16 |
+
<!-- This model card has been generated automatically according to the information the Trainer had access to. You
|
17 |
+
should probably proofread and complete it, then remove this comment. -->
|
18 |
+
|
19 |
# nutrition-extractor
|
20 |
|
21 |
+
This model is a fine-tuned version of [microsoft/layoutlmv3-large](https://huggingface.co/microsoft/layoutlmv3-large) on an unknown dataset.
|
|
|
22 |
It achieves the following results on the evaluation set:
|
23 |
+
- Loss: 0.0532
|
24 |
+
- Precision: 0.9536
|
25 |
+
- Recall: 0.9633
|
26 |
+
- F1: 0.9584
|
27 |
+
- Accuracy: 0.9916
|
28 |
+
|
29 |
+
## Model description
|
30 |
+
|
31 |
+
More information needed
|
32 |
+
|
33 |
+
## Intended uses & limitations
|
34 |
+
|
35 |
+
More information needed
|
36 |
+
|
37 |
+
## Training and evaluation data
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
38 |
|
39 |
+
More information needed
|
40 |
|
41 |
## Training procedure
|
42 |
|
|
|
58 |
|
59 |
| Training Loss | Epoch | Step | Validation Loss | Precision | Recall | F1 | Accuracy |
|
60 |
|:-------------:|:-------:|:----:|:---------------:|:---------:|:------:|:------:|:--------:|
|
61 |
+
| 1.9852 | 0.1664 | 15 | 1.1500 | 0.0 | 0.0 | 0.0 | 0.8101 |
|
62 |
+
| 1.0244 | 0.3329 | 30 | 0.8342 | 0.05 | 0.0074 | 0.0129 | 0.8123 |
|
63 |
+
| 0.7826 | 0.4993 | 45 | 0.6795 | 0.0789 | 0.1138 | 0.0932 | 0.8479 |
|
64 |
+
| 0.6767 | 0.6657 | 60 | 0.5963 | 0.1193 | 0.1644 | 0.1383 | 0.8578 |
|
65 |
+
| 0.6031 | 0.8322 | 75 | 0.5406 | 0.1671 | 0.2248 | 0.1917 | 0.8691 |
|
66 |
+
| 0.5756 | 0.9986 | 90 | 0.4935 | 0.2291 | 0.3112 | 0.2639 | 0.8778 |
|
67 |
+
| 0.5215 | 1.1650 | 105 | 0.4302 | 0.3267 | 0.3948 | 0.3575 | 0.8905 |
|
68 |
+
| 0.4782 | 1.3315 | 120 | 0.3782 | 0.3939 | 0.4654 | 0.4267 | 0.9020 |
|
69 |
+
| 0.4208 | 1.4979 | 135 | 0.3405 | 0.4027 | 0.5044 | 0.4478 | 0.9081 |
|
70 |
+
| 0.3532 | 1.6644 | 150 | 0.2930 | 0.4960 | 0.5820 | 0.5356 | 0.9252 |
|
71 |
+
| 0.3458 | 1.8308 | 165 | 0.2658 | 0.5155 | 0.6033 | 0.5560 | 0.9301 |
|
72 |
+
| 0.302 | 1.9972 | 180 | 0.2321 | 0.6112 | 0.7009 | 0.6530 | 0.9474 |
|
73 |
+
| 0.2655 | 2.1637 | 195 | 0.2093 | 0.6471 | 0.7264 | 0.6845 | 0.9520 |
|
74 |
+
| 0.2598 | 2.3301 | 210 | 0.1951 | 0.7013 | 0.7557 | 0.7275 | 0.9570 |
|
75 |
+
| 0.2364 | 2.4965 | 225 | 0.1794 | 0.7091 | 0.7743 | 0.7402 | 0.9590 |
|
76 |
+
| 0.2218 | 2.6630 | 240 | 0.1676 | 0.7216 | 0.7933 | 0.7558 | 0.9621 |
|
77 |
+
| 0.206 | 2.8294 | 255 | 0.1572 | 0.7436 | 0.8110 | 0.7758 | 0.9650 |
|
78 |
+
| 0.2053 | 2.9958 | 270 | 0.1580 | 0.7381 | 0.8114 | 0.7730 | 0.9640 |
|
79 |
+
| 0.1876 | 3.1623 | 285 | 0.1406 | 0.7738 | 0.8309 | 0.8013 | 0.9687 |
|
80 |
+
| 0.1602 | 3.3287 | 300 | 0.1420 | 0.7714 | 0.8277 | 0.7986 | 0.9671 |
|
81 |
+
| 0.1706 | 3.4951 | 315 | 0.1323 | 0.7933 | 0.8379 | 0.8150 | 0.9691 |
|
82 |
+
| 0.1585 | 3.6616 | 330 | 0.1313 | 0.8060 | 0.8551 | 0.8298 | 0.9700 |
|
83 |
+
| 0.1574 | 3.8280 | 345 | 0.1267 | 0.8129 | 0.8639 | 0.8376 | 0.9717 |
|
84 |
+
| 0.15 | 3.9945 | 360 | 0.1157 | 0.8336 | 0.8746 | 0.8536 | 0.9754 |
|
85 |
+
| 0.1192 | 4.1609 | 375 | 0.1120 | 0.8348 | 0.8709 | 0.8525 | 0.9741 |
|
86 |
+
| 0.1313 | 4.3273 | 390 | 0.1130 | 0.8395 | 0.8792 | 0.8589 | 0.9745 |
|
87 |
+
| 0.1179 | 4.4938 | 405 | 0.1093 | 0.8370 | 0.8871 | 0.8613 | 0.9755 |
|
88 |
+
| 0.1327 | 4.6602 | 420 | 0.1102 | 0.8400 | 0.8853 | 0.8621 | 0.9746 |
|
89 |
+
| 0.1323 | 4.8266 | 435 | 0.0997 | 0.8611 | 0.8987 | 0.8795 | 0.9782 |
|
90 |
+
| 0.1254 | 4.9931 | 450 | 0.0949 | 0.8499 | 0.8969 | 0.8728 | 0.9775 |
|
91 |
+
| 0.0999 | 5.1595 | 465 | 0.0847 | 0.8658 | 0.8992 | 0.8822 | 0.9797 |
|
92 |
+
| 0.1017 | 5.3259 | 480 | 0.0803 | 0.8747 | 0.9108 | 0.8924 | 0.9810 |
|
93 |
+
| 0.091 | 5.4924 | 495 | 0.0796 | 0.8784 | 0.9057 | 0.8918 | 0.9806 |
|
94 |
+
| 0.0979 | 5.6588 | 510 | 0.0943 | 0.8607 | 0.8950 | 0.8775 | 0.9773 |
|
95 |
+
| 0.1024 | 5.8252 | 525 | 0.0804 | 0.8710 | 0.9062 | 0.8882 | 0.9805 |
|
96 |
+
| 0.0952 | 5.9917 | 540 | 0.0787 | 0.8845 | 0.9178 | 0.9008 | 0.9816 |
|
97 |
+
| 0.0742 | 6.1581 | 555 | 0.0776 | 0.8918 | 0.9150 | 0.9033 | 0.9823 |
|
98 |
+
| 0.0764 | 6.3245 | 570 | 0.0721 | 0.9028 | 0.9187 | 0.9107 | 0.9837 |
|
99 |
+
| 0.0813 | 6.4910 | 585 | 0.0664 | 0.9065 | 0.9229 | 0.9146 | 0.9844 |
|
100 |
+
| 0.0791 | 6.6574 | 600 | 0.0642 | 0.9026 | 0.9252 | 0.9138 | 0.9848 |
|
101 |
+
| 0.0792 | 6.8239 | 615 | 0.0673 | 0.8964 | 0.9248 | 0.9104 | 0.9841 |
|
102 |
+
| 0.078 | 6.9903 | 630 | 0.0693 | 0.8938 | 0.9224 | 0.9079 | 0.9833 |
|
103 |
+
| 0.0678 | 7.1567 | 645 | 0.0672 | 0.9082 | 0.9327 | 0.9203 | 0.9852 |
|
104 |
+
| 0.0685 | 7.3232 | 660 | 0.0655 | 0.8926 | 0.9224 | 0.9073 | 0.9840 |
|
105 |
+
| 0.0555 | 7.4896 | 675 | 0.0615 | 0.9156 | 0.9271 | 0.9213 | 0.9856 |
|
106 |
+
| 0.07 | 7.6560 | 690 | 0.0587 | 0.9173 | 0.9373 | 0.9272 | 0.9868 |
|
107 |
+
| 0.065 | 7.8225 | 705 | 0.0558 | 0.9205 | 0.9405 | 0.9304 | 0.9875 |
|
108 |
+
| 0.0599 | 7.9889 | 720 | 0.0579 | 0.9253 | 0.9433 | 0.9342 | 0.9878 |
|
109 |
+
| 0.0571 | 8.1553 | 735 | 0.0593 | 0.9148 | 0.9331 | 0.9239 | 0.9866 |
|
110 |
+
| 0.0563 | 8.3218 | 750 | 0.0605 | 0.9152 | 0.9322 | 0.9236 | 0.9863 |
|
111 |
+
| 0.0602 | 8.4882 | 765 | 0.0581 | 0.9252 | 0.9308 | 0.9280 | 0.9863 |
|
112 |
+
| 0.0582 | 8.6546 | 780 | 0.0581 | 0.9206 | 0.9373 | 0.9289 | 0.9872 |
|
113 |
+
| 0.0514 | 8.8211 | 795 | 0.0557 | 0.9245 | 0.9382 | 0.9313 | 0.9873 |
|
114 |
+
| 0.0467 | 8.9875 | 810 | 0.0520 | 0.9291 | 0.9498 | 0.9394 | 0.9883 |
|
115 |
+
| 0.0435 | 9.1540 | 825 | 0.0526 | 0.9229 | 0.9447 | 0.9337 | 0.9880 |
|
116 |
+
| 0.0531 | 9.3204 | 840 | 0.0502 | 0.9249 | 0.9443 | 0.9345 | 0.9884 |
|
117 |
+
| 0.0502 | 9.4868 | 855 | 0.0545 | 0.9171 | 0.9452 | 0.9309 | 0.9874 |
|
118 |
+
| 0.0377 | 9.6533 | 870 | 0.0618 | 0.9077 | 0.9368 | 0.9221 | 0.9851 |
|
119 |
+
| 0.0416 | 9.8197 | 885 | 0.0549 | 0.9267 | 0.9392 | 0.9329 | 0.9881 |
|
120 |
+
| 0.044 | 9.9861 | 900 | 0.0529 | 0.9366 | 0.9475 | 0.9420 | 0.9884 |
|
121 |
+
| 0.0383 | 10.1526 | 915 | 0.0490 | 0.9332 | 0.9475 | 0.9403 | 0.9889 |
|
122 |
+
| 0.0454 | 10.3190 | 930 | 0.0507 | 0.9264 | 0.9471 | 0.9366 | 0.9885 |
|
123 |
+
| 0.0416 | 10.4854 | 945 | 0.0467 | 0.9364 | 0.9498 | 0.9430 | 0.9891 |
|
124 |
+
| 0.0403 | 10.6519 | 960 | 0.0499 | 0.9314 | 0.9457 | 0.9385 | 0.9886 |
|
125 |
+
| 0.0354 | 10.8183 | 975 | 0.0523 | 0.9258 | 0.9452 | 0.9354 | 0.9883 |
|
126 |
+
| 0.0338 | 10.9847 | 990 | 0.0521 | 0.9214 | 0.9424 | 0.9318 | 0.9880 |
|
127 |
+
| 0.0347 | 11.1512 | 1005 | 0.0539 | 0.9235 | 0.9475 | 0.9354 | 0.9880 |
|
128 |
+
| 0.0364 | 11.3176 | 1020 | 0.0560 | 0.9194 | 0.9480 | 0.9335 | 0.9871 |
|
129 |
+
| 0.0363 | 11.4840 | 1035 | 0.0509 | 0.9286 | 0.9480 | 0.9382 | 0.9889 |
|
130 |
+
| 0.0308 | 11.6505 | 1050 | 0.0498 | 0.9389 | 0.9484 | 0.9436 | 0.9893 |
|
131 |
+
| 0.032 | 11.8169 | 1065 | 0.0491 | 0.9364 | 0.9443 | 0.9403 | 0.9891 |
|
132 |
+
| 0.0331 | 11.9834 | 1080 | 0.0455 | 0.9373 | 0.9443 | 0.9408 | 0.9892 |
|
133 |
+
| 0.0301 | 12.1498 | 1095 | 0.0486 | 0.9359 | 0.9489 | 0.9423 | 0.9892 |
|
134 |
+
| 0.0308 | 12.3162 | 1110 | 0.0513 | 0.9325 | 0.9503 | 0.9413 | 0.9891 |
|
135 |
+
| 0.0253 | 12.4827 | 1125 | 0.0510 | 0.9296 | 0.9503 | 0.9398 | 0.9892 |
|
136 |
+
| 0.0301 | 12.6491 | 1140 | 0.0533 | 0.9308 | 0.9489 | 0.9397 | 0.9886 |
|
137 |
+
| 0.0328 | 12.8155 | 1155 | 0.0549 | 0.9287 | 0.9443 | 0.9364 | 0.9885 |
|
138 |
+
| 0.0298 | 12.9820 | 1170 | 0.0504 | 0.9402 | 0.9498 | 0.9450 | 0.9895 |
|
139 |
+
| 0.0256 | 13.1484 | 1185 | 0.0515 | 0.9354 | 0.9419 | 0.9387 | 0.9888 |
|
140 |
+
| 0.0313 | 13.3148 | 1200 | 0.0483 | 0.9418 | 0.9545 | 0.9481 | 0.9905 |
|
141 |
+
| 0.022 | 13.4813 | 1215 | 0.0463 | 0.9361 | 0.9531 | 0.9445 | 0.9899 |
|
142 |
+
| 0.0245 | 13.6477 | 1230 | 0.0494 | 0.9368 | 0.9494 | 0.9430 | 0.9893 |
|
143 |
+
| 0.0251 | 13.8141 | 1245 | 0.0493 | 0.9404 | 0.9531 | 0.9467 | 0.9898 |
|
144 |
+
| 0.0259 | 13.9806 | 1260 | 0.0511 | 0.9386 | 0.9522 | 0.9454 | 0.9895 |
|
145 |
+
| 0.03 | 14.1470 | 1275 | 0.0535 | 0.9344 | 0.9457 | 0.9400 | 0.9889 |
|
146 |
+
| 0.0192 | 14.3135 | 1290 | 0.0491 | 0.9428 | 0.9494 | 0.9461 | 0.9899 |
|
147 |
+
| 0.0267 | 14.4799 | 1305 | 0.0490 | 0.9457 | 0.9545 | 0.9501 | 0.9901 |
|
148 |
+
| 0.0241 | 14.6463 | 1320 | 0.0506 | 0.9435 | 0.9540 | 0.9487 | 0.9899 |
|
149 |
+
| 0.0211 | 14.8128 | 1335 | 0.0510 | 0.9444 | 0.9540 | 0.9492 | 0.9903 |
|
150 |
+
| 0.0171 | 14.9792 | 1350 | 0.0499 | 0.9405 | 0.9545 | 0.9474 | 0.9898 |
|
151 |
+
| 0.0226 | 15.1456 | 1365 | 0.0511 | 0.9366 | 0.9540 | 0.9452 | 0.9894 |
|
152 |
+
| 0.024 | 15.3121 | 1380 | 0.0484 | 0.9445 | 0.9559 | 0.9501 | 0.9899 |
|
153 |
+
| 0.018 | 15.4785 | 1395 | 0.0482 | 0.9469 | 0.9517 | 0.9493 | 0.9903 |
|
154 |
+
| 0.0191 | 15.6449 | 1410 | 0.0491 | 0.9442 | 0.9512 | 0.9477 | 0.9899 |
|
155 |
+
| 0.0203 | 15.8114 | 1425 | 0.0451 | 0.9510 | 0.9554 | 0.9532 | 0.9912 |
|
156 |
+
| 0.0198 | 15.9778 | 1440 | 0.0447 | 0.9497 | 0.9549 | 0.9523 | 0.9911 |
|
157 |
+
| 0.0167 | 16.1442 | 1455 | 0.0444 | 0.9487 | 0.9540 | 0.9514 | 0.9909 |
|
158 |
+
| 0.0178 | 16.3107 | 1470 | 0.0513 | 0.9386 | 0.9512 | 0.9449 | 0.9892 |
|
159 |
+
| 0.024 | 16.4771 | 1485 | 0.0502 | 0.9430 | 0.9536 | 0.9483 | 0.9899 |
|
160 |
+
| 0.0206 | 16.6436 | 1500 | 0.0459 | 0.9483 | 0.9545 | 0.9514 | 0.9908 |
|
161 |
+
| 0.0188 | 16.8100 | 1515 | 0.0469 | 0.9474 | 0.9540 | 0.9507 | 0.9906 |
|
162 |
+
| 0.016 | 16.9764 | 1530 | 0.0463 | 0.9468 | 0.9582 | 0.9524 | 0.9906 |
|
163 |
+
| 0.0161 | 17.1429 | 1545 | 0.0455 | 0.9516 | 0.9596 | 0.9556 | 0.9911 |
|
164 |
+
| 0.0135 | 17.3093 | 1560 | 0.0475 | 0.9524 | 0.9573 | 0.9548 | 0.9909 |
|
165 |
+
| 0.0148 | 17.4757 | 1575 | 0.0479 | 0.9440 | 0.9545 | 0.9492 | 0.9905 |
|
166 |
+
| 0.0173 | 17.6422 | 1590 | 0.0455 | 0.9539 | 0.9605 | 0.9572 | 0.9915 |
|
167 |
+
| 0.0173 | 17.8086 | 1605 | 0.0456 | 0.9475 | 0.9554 | 0.9514 | 0.9913 |
|
168 |
+
| 0.0185 | 17.9750 | 1620 | 0.0461 | 0.9498 | 0.9577 | 0.9537 | 0.9908 |
|
169 |
+
| 0.0153 | 18.1415 | 1635 | 0.0472 | 0.9491 | 0.9605 | 0.9548 | 0.9911 |
|
170 |
+
| 0.0148 | 18.3079 | 1650 | 0.0446 | 0.9507 | 0.9587 | 0.9547 | 0.9913 |
|
171 |
+
| 0.0136 | 18.4743 | 1665 | 0.0441 | 0.9486 | 0.9601 | 0.9543 | 0.9914 |
|
172 |
+
| 0.0185 | 18.6408 | 1680 | 0.0478 | 0.9528 | 0.9573 | 0.9551 | 0.9915 |
|
173 |
+
| 0.0147 | 18.8072 | 1695 | 0.0493 | 0.9515 | 0.9652 | 0.9583 | 0.9912 |
|
174 |
+
| 0.0156 | 18.9736 | 1710 | 0.0509 | 0.9440 | 0.9545 | 0.9492 | 0.9903 |
|
175 |
+
| 0.0113 | 19.1401 | 1725 | 0.0460 | 0.9559 | 0.9573 | 0.9566 | 0.9911 |
|
176 |
+
| 0.014 | 19.3065 | 1740 | 0.0493 | 0.9439 | 0.9526 | 0.9482 | 0.9905 |
|
177 |
+
| 0.0147 | 19.4730 | 1755 | 0.0498 | 0.9476 | 0.9568 | 0.9522 | 0.9906 |
|
178 |
+
| 0.0126 | 19.6394 | 1770 | 0.0493 | 0.9474 | 0.9531 | 0.9502 | 0.9906 |
|
179 |
+
| 0.0167 | 19.8058 | 1785 | 0.0491 | 0.9463 | 0.9577 | 0.9520 | 0.9904 |
|
180 |
+
| 0.0126 | 19.9723 | 1800 | 0.0474 | 0.9492 | 0.9540 | 0.9516 | 0.9908 |
|
181 |
+
| 0.0107 | 20.1387 | 1815 | 0.0462 | 0.9524 | 0.9577 | 0.9551 | 0.9914 |
|
182 |
+
| 0.0115 | 20.3051 | 1830 | 0.0481 | 0.9504 | 0.9614 | 0.9559 | 0.9911 |
|
183 |
+
| 0.0128 | 20.4716 | 1845 | 0.0486 | 0.9475 | 0.9563 | 0.9519 | 0.9907 |
|
184 |
+
| 0.0113 | 20.6380 | 1860 | 0.0491 | 0.9477 | 0.9591 | 0.9534 | 0.9910 |
|
185 |
+
| 0.0119 | 20.8044 | 1875 | 0.0514 | 0.9494 | 0.9503 | 0.9499 | 0.9901 |
|
186 |
+
| 0.0122 | 20.9709 | 1890 | 0.0480 | 0.9481 | 0.9591 | 0.9536 | 0.9911 |
|
187 |
+
| 0.0123 | 21.1373 | 1905 | 0.0477 | 0.9467 | 0.9577 | 0.9522 | 0.9909 |
|
188 |
+
| 0.0116 | 21.3037 | 1920 | 0.0486 | 0.9485 | 0.9582 | 0.9533 | 0.9910 |
|
189 |
+
| 0.0108 | 21.4702 | 1935 | 0.0488 | 0.9442 | 0.9582 | 0.9511 | 0.9905 |
|
190 |
+
| 0.0115 | 21.6366 | 1950 | 0.0472 | 0.9498 | 0.9587 | 0.9542 | 0.9913 |
|
191 |
+
| 0.0083 | 21.8031 | 1965 | 0.0476 | 0.9490 | 0.9596 | 0.9543 | 0.9911 |
|
192 |
+
| 0.0094 | 21.9695 | 1980 | 0.0475 | 0.9482 | 0.9605 | 0.9543 | 0.9909 |
|
193 |
+
| 0.0118 | 22.1359 | 1995 | 0.0492 | 0.9449 | 0.9554 | 0.9501 | 0.9904 |
|
194 |
+
| 0.01 | 22.3024 | 2010 | 0.0486 | 0.9492 | 0.9554 | 0.9523 | 0.9909 |
|
195 |
+
| 0.0114 | 22.4688 | 2025 | 0.0497 | 0.9502 | 0.9577 | 0.9540 | 0.9910 |
|
196 |
+
| 0.0091 | 22.6352 | 2040 | 0.0499 | 0.9503 | 0.9582 | 0.9542 | 0.9910 |
|
197 |
+
| 0.0077 | 22.8017 | 2055 | 0.0502 | 0.9513 | 0.9614 | 0.9563 | 0.9911 |
|
198 |
+
| 0.01 | 22.9681 | 2070 | 0.0513 | 0.9544 | 0.9628 | 0.9586 | 0.9913 |
|
199 |
+
| 0.0087 | 23.1345 | 2085 | 0.0485 | 0.9500 | 0.9610 | 0.9554 | 0.9912 |
|
200 |
+
| 0.0073 | 23.3010 | 2100 | 0.0485 | 0.9557 | 0.9628 | 0.9593 | 0.9917 |
|
201 |
+
| 0.0083 | 23.4674 | 2115 | 0.0485 | 0.9535 | 0.9610 | 0.9572 | 0.9913 |
|
202 |
+
| 0.0117 | 23.6338 | 2130 | 0.0479 | 0.9557 | 0.9624 | 0.9590 | 0.9916 |
|
203 |
+
| 0.0095 | 23.8003 | 2145 | 0.0508 | 0.9498 | 0.9587 | 0.9542 | 0.9911 |
|
204 |
+
| 0.009 | 23.9667 | 2160 | 0.0513 | 0.9492 | 0.9628 | 0.9560 | 0.9910 |
|
205 |
+
| 0.0077 | 24.1331 | 2175 | 0.0504 | 0.9553 | 0.9628 | 0.9591 | 0.9915 |
|
206 |
+
| 0.0087 | 24.2996 | 2190 | 0.0500 | 0.9521 | 0.9610 | 0.9565 | 0.9913 |
|
207 |
+
| 0.0068 | 24.4660 | 2205 | 0.0506 | 0.9539 | 0.9610 | 0.9574 | 0.9913 |
|
208 |
+
| 0.0094 | 24.6325 | 2220 | 0.0500 | 0.9507 | 0.9591 | 0.9549 | 0.9913 |
|
209 |
+
| 0.0088 | 24.7989 | 2235 | 0.0486 | 0.9508 | 0.9596 | 0.9552 | 0.9914 |
|
210 |
+
| 0.0089 | 24.9653 | 2250 | 0.0507 | 0.9508 | 0.9610 | 0.9559 | 0.9911 |
|
211 |
+
| 0.0063 | 25.1318 | 2265 | 0.0479 | 0.9561 | 0.9610 | 0.9585 | 0.9917 |
|
212 |
+
| 0.0058 | 25.2982 | 2280 | 0.0506 | 0.9526 | 0.9619 | 0.9572 | 0.9911 |
|
213 |
+
| 0.0102 | 25.4646 | 2295 | 0.0499 | 0.9526 | 0.9624 | 0.9575 | 0.9912 |
|
214 |
+
| 0.0079 | 25.6311 | 2310 | 0.0543 | 0.9469 | 0.9614 | 0.9541 | 0.9905 |
|
215 |
+
| 0.009 | 25.7975 | 2325 | 0.0498 | 0.9526 | 0.9619 | 0.9572 | 0.9915 |
|
216 |
+
| 0.0068 | 25.9639 | 2340 | 0.0511 | 0.9509 | 0.9619 | 0.9564 | 0.9911 |
|
217 |
+
| 0.007 | 26.1304 | 2355 | 0.0492 | 0.9527 | 0.9633 | 0.9580 | 0.9914 |
|
218 |
+
| 0.0086 | 26.2968 | 2370 | 0.0516 | 0.9500 | 0.9610 | 0.9554 | 0.9913 |
|
219 |
+
| 0.0078 | 26.4632 | 2385 | 0.0503 | 0.9504 | 0.9610 | 0.9557 | 0.9914 |
|
220 |
+
| 0.0067 | 26.6297 | 2400 | 0.0514 | 0.9527 | 0.9628 | 0.9577 | 0.9915 |
|
221 |
+
| 0.0059 | 26.7961 | 2415 | 0.0504 | 0.9549 | 0.9628 | 0.9588 | 0.9919 |
|
222 |
+
| 0.0089 | 26.9626 | 2430 | 0.0520 | 0.9517 | 0.9605 | 0.9561 | 0.9916 |
|
223 |
+
| 0.0059 | 27.1290 | 2445 | 0.0512 | 0.9522 | 0.9624 | 0.9573 | 0.9917 |
|
224 |
+
| 0.0073 | 27.2954 | 2460 | 0.0526 | 0.9530 | 0.9610 | 0.9570 | 0.9916 |
|
225 |
+
| 0.0065 | 27.4619 | 2475 | 0.0530 | 0.9527 | 0.9628 | 0.9577 | 0.9916 |
|
226 |
+
| 0.0064 | 27.6283 | 2490 | 0.0515 | 0.9535 | 0.9610 | 0.9572 | 0.9917 |
|
227 |
+
| 0.0072 | 27.7947 | 2505 | 0.0542 | 0.9482 | 0.9610 | 0.9546 | 0.9907 |
|
228 |
+
| 0.0066 | 27.9612 | 2520 | 0.0537 | 0.9491 | 0.9610 | 0.9550 | 0.9909 |
|
229 |
+
| 0.006 | 28.1276 | 2535 | 0.0518 | 0.9531 | 0.9628 | 0.9579 | 0.9915 |
|
230 |
+
| 0.0074 | 28.2940 | 2550 | 0.0523 | 0.9521 | 0.9610 | 0.9565 | 0.9914 |
|
231 |
+
| 0.0068 | 28.4605 | 2565 | 0.0534 | 0.9495 | 0.9614 | 0.9555 | 0.9913 |
|
232 |
+
| 0.0055 | 28.6269 | 2580 | 0.0521 | 0.9548 | 0.9619 | 0.9584 | 0.9917 |
|
233 |
+
| 0.0056 | 28.7933 | 2595 | 0.0526 | 0.9522 | 0.9614 | 0.9568 | 0.9913 |
|
234 |
+
| 0.0066 | 28.9598 | 2610 | 0.0527 | 0.9522 | 0.9619 | 0.9570 | 0.9913 |
|
235 |
+
| 0.0053 | 29.1262 | 2625 | 0.0533 | 0.9531 | 0.9628 | 0.9579 | 0.9913 |
|
236 |
+
| 0.0063 | 29.2926 | 2640 | 0.0520 | 0.9530 | 0.9610 | 0.9570 | 0.9913 |
|
237 |
+
| 0.0059 | 29.4591 | 2655 | 0.0533 | 0.9504 | 0.9605 | 0.9554 | 0.9910 |
|
238 |
+
| 0.0059 | 29.6255 | 2670 | 0.0532 | 0.9526 | 0.9619 | 0.9572 | 0.9912 |
|
239 |
+
| 0.0062 | 29.7920 | 2685 | 0.0516 | 0.9535 | 0.9624 | 0.9579 | 0.9917 |
|
240 |
+
| 0.0064 | 29.9584 | 2700 | 0.0515 | 0.9522 | 0.9624 | 0.9573 | 0.9915 |
|
241 |
+
| 0.0055 | 30.1248 | 2715 | 0.0513 | 0.9549 | 0.9633 | 0.9591 | 0.9917 |
|
242 |
+
| 0.0064 | 30.2913 | 2730 | 0.0524 | 0.9540 | 0.9628 | 0.9584 | 0.9916 |
|
243 |
+
| 0.0055 | 30.4577 | 2745 | 0.0530 | 0.9531 | 0.9633 | 0.9582 | 0.9915 |
|
244 |
+
| 0.0065 | 30.6241 | 2760 | 0.0528 | 0.9536 | 0.9642 | 0.9589 | 0.9917 |
|
245 |
+
| 0.0068 | 30.7906 | 2775 | 0.0530 | 0.9518 | 0.9633 | 0.9575 | 0.9916 |
|
246 |
+
| 0.0047 | 30.9570 | 2790 | 0.0545 | 0.9532 | 0.9647 | 0.9589 | 0.9916 |
|
247 |
+
| 0.0051 | 31.1234 | 2805 | 0.0534 | 0.9545 | 0.9647 | 0.9596 | 0.9917 |
|
248 |
+
| 0.0044 | 31.2899 | 2820 | 0.0532 | 0.9531 | 0.9633 | 0.9582 | 0.9914 |
|
249 |
+
| 0.0068 | 31.4563 | 2835 | 0.0532 | 0.9527 | 0.9633 | 0.9580 | 0.9913 |
|
250 |
+
| 0.0045 | 31.6227 | 2850 | 0.0531 | 0.9545 | 0.9638 | 0.9591 | 0.9915 |
|
251 |
+
| 0.0047 | 31.7892 | 2865 | 0.0530 | 0.9540 | 0.9633 | 0.9586 | 0.9916 |
|
252 |
+
| 0.0075 | 31.9556 | 2880 | 0.0533 | 0.9549 | 0.9638 | 0.9593 | 0.9916 |
|
253 |
+
| 0.0055 | 32.1221 | 2895 | 0.0525 | 0.9553 | 0.9638 | 0.9595 | 0.9917 |
|
254 |
+
| 0.006 | 32.2885 | 2910 | 0.0523 | 0.9553 | 0.9638 | 0.9595 | 0.9917 |
|
255 |
+
| 0.0062 | 32.4549 | 2925 | 0.0525 | 0.9544 | 0.9633 | 0.9589 | 0.9917 |
|
256 |
+
| 0.0059 | 32.6214 | 2940 | 0.0525 | 0.9549 | 0.9638 | 0.9593 | 0.9917 |
|
257 |
+
| 0.0058 | 32.7878 | 2955 | 0.0531 | 0.9549 | 0.9642 | 0.9596 | 0.9917 |
|
258 |
+
| 0.005 | 32.9542 | 2970 | 0.0533 | 0.9536 | 0.9633 | 0.9584 | 0.9916 |
|
259 |
+
| 0.007 | 33.1207 | 2985 | 0.0533 | 0.9536 | 0.9633 | 0.9584 | 0.9916 |
|
260 |
+
| 0.0047 | 33.2871 | 3000 | 0.0532 | 0.9536 | 0.9633 | 0.9584 | 0.9916 |
|
261 |
|
262 |
|
263 |
### Framework versions
|
264 |
|
265 |
- Transformers 4.40.2
|
266 |
+
- Pytorch 2.5.1
|
267 |
- Datasets 2.19.0
|
268 |
+
- Tokenizers 0.19.1
|
config.json
CHANGED
@@ -1,5 +1,5 @@
|
|
1 |
{
|
2 |
-
"_name_or_path": "
|
3 |
"architectures": [
|
4 |
"LayoutLMv3ForTokenClassification"
|
5 |
],
|
@@ -16,151 +16,151 @@
|
|
16 |
"hidden_size": 1024,
|
17 |
"id2label": {
|
18 |
"0": "O",
|
19 |
-
"1": "B-
|
20 |
-
"2": "I-
|
21 |
-
"3": "B-
|
22 |
-
"4": "I-
|
23 |
-
"5": "B-
|
24 |
-
"6": "I-
|
25 |
-
"7": "B-
|
26 |
-
"8": "I-
|
27 |
-
"9": "B-
|
28 |
-
"10": "I-
|
29 |
-
"11": "B-
|
30 |
-
"12": "I-
|
31 |
-
"13": "B-
|
32 |
-
"14": "I-
|
33 |
-
"15": "B-
|
34 |
-
"16": "I-
|
35 |
-
"17": "B-
|
36 |
-
"18": "I-
|
37 |
-
"19": "B-
|
38 |
-
"20": "I-
|
39 |
-
"21": "B-
|
40 |
-
"22": "I-
|
41 |
-
"23": "B-
|
42 |
-
"24": "I-
|
43 |
-
"25": "B-
|
44 |
-
"26": "I-
|
45 |
-
"27": "B-
|
46 |
-
"28": "I-
|
47 |
-
"29": "B-
|
48 |
-
"30": "I-
|
49 |
-
"31": "B-
|
50 |
-
"32": "I-
|
51 |
-
"33": "B-
|
52 |
-
"34": "I-
|
53 |
-
"35": "B-
|
54 |
-
"36": "I-
|
55 |
-
"37": "B-
|
56 |
-
"38": "I-
|
57 |
-
"39": "B-
|
58 |
-
"40": "I-
|
59 |
-
"41": "B-
|
60 |
-
"42": "I-
|
61 |
-
"43": "B-
|
62 |
-
"44": "I-
|
63 |
-
"45": "B-
|
64 |
-
"46": "I-
|
65 |
-
"47": "B-
|
66 |
-
"48": "I-
|
67 |
-
"49": "B-
|
68 |
-
"50": "I-
|
69 |
-
"51": "B-
|
70 |
-
"52": "I-
|
71 |
-
"53": "B-
|
72 |
-
"54": "I-
|
73 |
-
"55": "B-
|
74 |
-
"56": "I-
|
75 |
-
"57": "B-
|
76 |
-
"58": "I-
|
77 |
-
"59": "B-
|
78 |
-
"60": "I-
|
79 |
-
"61": "B-
|
80 |
-
"62": "I-
|
81 |
-
"63": "B-
|
82 |
-
"64": "I-
|
83 |
-
"65": "B-
|
84 |
-
"66": "I-
|
85 |
-
"67": "B-
|
86 |
-
"68": "I-
|
87 |
-
"69": "B-
|
88 |
-
"70": "I-
|
89 |
},
|
90 |
"initializer_range": 0.02,
|
91 |
"input_size": 224,
|
92 |
"intermediate_size": 4096,
|
93 |
"label2id": {
|
94 |
-
"B-ADDED_SUGARS_100G":
|
95 |
-
"B-ADDED_SUGARS_SERVING":
|
96 |
-
"B-CALCIUM_100G":
|
97 |
-
"B-CALCIUM_SERVING":
|
98 |
-
"B-CARBOHYDRATES_100G":
|
99 |
-
"B-CARBOHYDRATES_SERVING":
|
100 |
-
"B-CHOLESTEROL_100G":
|
101 |
-
"B-CHOLESTEROL_SERVING":
|
102 |
-
"B-ENERGY_KCAL_100G":
|
103 |
-
"B-ENERGY_KCAL_SERVING":
|
104 |
-
"B-ENERGY_KJ_100G":
|
105 |
-
"B-ENERGY_KJ_SERVING":
|
106 |
-
"B-FAT_100G":
|
107 |
-
"B-FAT_SERVING":
|
108 |
-
"B-FIBER_100G":
|
109 |
-
"B-FIBER_SERVING":
|
110 |
-
"B-IRON_100G":
|
111 |
-
"B-IRON_SERVING":
|
112 |
-
"B-POTASSIUM_100G":
|
113 |
-
"B-POTASSIUM_SERVING":
|
114 |
-
"B-PROTEINS_100G":
|
115 |
-
"B-PROTEINS_SERVING":
|
116 |
-
"B-SALT_100G":
|
117 |
-
"B-SALT_SERVING":
|
118 |
-
"B-SATURATED_FAT_100G":
|
119 |
-
"B-SATURATED_FAT_SERVING":
|
120 |
-
"B-SERVING_SIZE":
|
121 |
-
"B-SODIUM_100G":
|
122 |
-
"B-SODIUM_SERVING":
|
123 |
-
"B-SUGARS_100G":
|
124 |
-
"B-SUGARS_SERVING":
|
125 |
-
"B-TRANS_FAT_100G":
|
126 |
-
"B-TRANS_FAT_SERVING":
|
127 |
-
"B-VITAMIN_D_100G":
|
128 |
-
"B-VITAMIN_D_SERVING":
|
129 |
-
"I-ADDED_SUGARS_100G":
|
130 |
-
"I-ADDED_SUGARS_SERVING":
|
131 |
-
"I-CALCIUM_100G":
|
132 |
-
"I-CALCIUM_SERVING":
|
133 |
-
"I-CARBOHYDRATES_100G":
|
134 |
-
"I-CARBOHYDRATES_SERVING":
|
135 |
-
"I-CHOLESTEROL_100G":
|
136 |
-
"I-CHOLESTEROL_SERVING":
|
137 |
-
"I-ENERGY_KCAL_100G":
|
138 |
-
"I-ENERGY_KCAL_SERVING":
|
139 |
-
"I-ENERGY_KJ_100G":
|
140 |
-
"I-ENERGY_KJ_SERVING":
|
141 |
-
"I-FAT_100G":
|
142 |
-
"I-FAT_SERVING":
|
143 |
-
"I-FIBER_100G":
|
144 |
-
"I-FIBER_SERVING":
|
145 |
-
"I-IRON_100G":
|
146 |
-
"I-IRON_SERVING":
|
147 |
-
"I-POTASSIUM_100G":
|
148 |
-
"I-POTASSIUM_SERVING":
|
149 |
-
"I-PROTEINS_100G":
|
150 |
-
"I-PROTEINS_SERVING":
|
151 |
-
"I-SALT_100G":
|
152 |
-
"I-SALT_SERVING":
|
153 |
-
"I-SATURATED_FAT_100G":
|
154 |
-
"I-SATURATED_FAT_SERVING":
|
155 |
-
"I-SERVING_SIZE":
|
156 |
-
"I-SODIUM_100G":
|
157 |
-
"I-SODIUM_SERVING":
|
158 |
-
"I-SUGARS_100G":
|
159 |
-
"I-SUGARS_SERVING":
|
160 |
-
"I-TRANS_FAT_100G":
|
161 |
-
"I-TRANS_FAT_SERVING":
|
162 |
-
"I-VITAMIN_D_100G":
|
163 |
-
"I-VITAMIN_D_SERVING":
|
164 |
"O": 0
|
165 |
},
|
166 |
"layer_norm_eps": 1e-05,
|
|
|
1 |
{
|
2 |
+
"_name_or_path": "microsoft/layoutlmv3-large",
|
3 |
"architectures": [
|
4 |
"LayoutLMv3ForTokenClassification"
|
5 |
],
|
|
|
16 |
"hidden_size": 1024,
|
17 |
"id2label": {
|
18 |
"0": "O",
|
19 |
+
"1": "B-ENERGY_KJ_100G",
|
20 |
+
"2": "I-ENERGY_KJ_100G",
|
21 |
+
"3": "B-VITAMIN_D_SERVING",
|
22 |
+
"4": "I-VITAMIN_D_SERVING",
|
23 |
+
"5": "B-SODIUM_SERVING",
|
24 |
+
"6": "I-SODIUM_SERVING",
|
25 |
+
"7": "B-PROTEINS_SERVING",
|
26 |
+
"8": "I-PROTEINS_SERVING",
|
27 |
+
"9": "B-ADDED_SUGARS_SERVING",
|
28 |
+
"10": "I-ADDED_SUGARS_SERVING",
|
29 |
+
"11": "B-CALCIUM_SERVING",
|
30 |
+
"12": "I-CALCIUM_SERVING",
|
31 |
+
"13": "B-FAT_SERVING",
|
32 |
+
"14": "I-FAT_SERVING",
|
33 |
+
"15": "B-ENERGY_KJ_SERVING",
|
34 |
+
"16": "I-ENERGY_KJ_SERVING",
|
35 |
+
"17": "B-SUGARS_100G",
|
36 |
+
"18": "I-SUGARS_100G",
|
37 |
+
"19": "B-SATURATED_FAT_SERVING",
|
38 |
+
"20": "I-SATURATED_FAT_SERVING",
|
39 |
+
"21": "B-SERVING_SIZE",
|
40 |
+
"22": "I-SERVING_SIZE",
|
41 |
+
"23": "B-SALT_SERVING",
|
42 |
+
"24": "I-SALT_SERVING",
|
43 |
+
"25": "B-ENERGY_KCAL_SERVING",
|
44 |
+
"26": "I-ENERGY_KCAL_SERVING",
|
45 |
+
"27": "B-FAT_100G",
|
46 |
+
"28": "I-FAT_100G",
|
47 |
+
"29": "B-SUGARS_SERVING",
|
48 |
+
"30": "I-SUGARS_SERVING",
|
49 |
+
"31": "B-FIBER_SERVING",
|
50 |
+
"32": "I-FIBER_SERVING",
|
51 |
+
"33": "B-TRANS_FAT_SERVING",
|
52 |
+
"34": "I-TRANS_FAT_SERVING",
|
53 |
+
"35": "B-POTASSIUM_SERVING",
|
54 |
+
"36": "I-POTASSIUM_SERVING",
|
55 |
+
"37": "B-CARBOHYDRATES_100G",
|
56 |
+
"38": "I-CARBOHYDRATES_100G",
|
57 |
+
"39": "B-POTASSIUM_100G",
|
58 |
+
"40": "I-POTASSIUM_100G",
|
59 |
+
"41": "B-IRON_SERVING",
|
60 |
+
"42": "I-IRON_SERVING",
|
61 |
+
"43": "B-CHOLESTEROL_100G",
|
62 |
+
"44": "I-CHOLESTEROL_100G",
|
63 |
+
"45": "B-TRANS_FAT_100G",
|
64 |
+
"46": "I-TRANS_FAT_100G",
|
65 |
+
"47": "B-ADDED_SUGARS_100G",
|
66 |
+
"48": "I-ADDED_SUGARS_100G",
|
67 |
+
"49": "B-FIBER_100G",
|
68 |
+
"50": "I-FIBER_100G",
|
69 |
+
"51": "B-CALCIUM_100G",
|
70 |
+
"52": "I-CALCIUM_100G",
|
71 |
+
"53": "B-SODIUM_100G",
|
72 |
+
"54": "I-SODIUM_100G",
|
73 |
+
"55": "B-ENERGY_KCAL_100G",
|
74 |
+
"56": "I-ENERGY_KCAL_100G",
|
75 |
+
"57": "B-CHOLESTEROL_SERVING",
|
76 |
+
"58": "I-CHOLESTEROL_SERVING",
|
77 |
+
"59": "B-CARBOHYDRATES_SERVING",
|
78 |
+
"60": "I-CARBOHYDRATES_SERVING",
|
79 |
+
"61": "B-SALT_100G",
|
80 |
+
"62": "I-SALT_100G",
|
81 |
+
"63": "B-VITAMIN_D_100G",
|
82 |
+
"64": "I-VITAMIN_D_100G",
|
83 |
+
"65": "B-SATURATED_FAT_100G",
|
84 |
+
"66": "I-SATURATED_FAT_100G",
|
85 |
+
"67": "B-PROTEINS_100G",
|
86 |
+
"68": "I-PROTEINS_100G",
|
87 |
+
"69": "B-IRON_100G",
|
88 |
+
"70": "I-IRON_100G"
|
89 |
},
|
90 |
"initializer_range": 0.02,
|
91 |
"input_size": 224,
|
92 |
"intermediate_size": 4096,
|
93 |
"label2id": {
|
94 |
+
"B-ADDED_SUGARS_100G": 47,
|
95 |
+
"B-ADDED_SUGARS_SERVING": 9,
|
96 |
+
"B-CALCIUM_100G": 51,
|
97 |
+
"B-CALCIUM_SERVING": 11,
|
98 |
+
"B-CARBOHYDRATES_100G": 37,
|
99 |
+
"B-CARBOHYDRATES_SERVING": 59,
|
100 |
+
"B-CHOLESTEROL_100G": 43,
|
101 |
+
"B-CHOLESTEROL_SERVING": 57,
|
102 |
+
"B-ENERGY_KCAL_100G": 55,
|
103 |
+
"B-ENERGY_KCAL_SERVING": 25,
|
104 |
+
"B-ENERGY_KJ_100G": 1,
|
105 |
+
"B-ENERGY_KJ_SERVING": 15,
|
106 |
+
"B-FAT_100G": 27,
|
107 |
+
"B-FAT_SERVING": 13,
|
108 |
+
"B-FIBER_100G": 49,
|
109 |
+
"B-FIBER_SERVING": 31,
|
110 |
+
"B-IRON_100G": 69,
|
111 |
+
"B-IRON_SERVING": 41,
|
112 |
+
"B-POTASSIUM_100G": 39,
|
113 |
+
"B-POTASSIUM_SERVING": 35,
|
114 |
+
"B-PROTEINS_100G": 67,
|
115 |
+
"B-PROTEINS_SERVING": 7,
|
116 |
+
"B-SALT_100G": 61,
|
117 |
+
"B-SALT_SERVING": 23,
|
118 |
+
"B-SATURATED_FAT_100G": 65,
|
119 |
+
"B-SATURATED_FAT_SERVING": 19,
|
120 |
+
"B-SERVING_SIZE": 21,
|
121 |
+
"B-SODIUM_100G": 53,
|
122 |
+
"B-SODIUM_SERVING": 5,
|
123 |
+
"B-SUGARS_100G": 17,
|
124 |
+
"B-SUGARS_SERVING": 29,
|
125 |
+
"B-TRANS_FAT_100G": 45,
|
126 |
+
"B-TRANS_FAT_SERVING": 33,
|
127 |
+
"B-VITAMIN_D_100G": 63,
|
128 |
+
"B-VITAMIN_D_SERVING": 3,
|
129 |
+
"I-ADDED_SUGARS_100G": 48,
|
130 |
+
"I-ADDED_SUGARS_SERVING": 10,
|
131 |
+
"I-CALCIUM_100G": 52,
|
132 |
+
"I-CALCIUM_SERVING": 12,
|
133 |
+
"I-CARBOHYDRATES_100G": 38,
|
134 |
+
"I-CARBOHYDRATES_SERVING": 60,
|
135 |
+
"I-CHOLESTEROL_100G": 44,
|
136 |
+
"I-CHOLESTEROL_SERVING": 58,
|
137 |
+
"I-ENERGY_KCAL_100G": 56,
|
138 |
+
"I-ENERGY_KCAL_SERVING": 26,
|
139 |
+
"I-ENERGY_KJ_100G": 2,
|
140 |
+
"I-ENERGY_KJ_SERVING": 16,
|
141 |
+
"I-FAT_100G": 28,
|
142 |
+
"I-FAT_SERVING": 14,
|
143 |
+
"I-FIBER_100G": 50,
|
144 |
+
"I-FIBER_SERVING": 32,
|
145 |
+
"I-IRON_100G": 70,
|
146 |
+
"I-IRON_SERVING": 42,
|
147 |
+
"I-POTASSIUM_100G": 40,
|
148 |
+
"I-POTASSIUM_SERVING": 36,
|
149 |
+
"I-PROTEINS_100G": 68,
|
150 |
+
"I-PROTEINS_SERVING": 8,
|
151 |
+
"I-SALT_100G": 62,
|
152 |
+
"I-SALT_SERVING": 24,
|
153 |
+
"I-SATURATED_FAT_100G": 66,
|
154 |
+
"I-SATURATED_FAT_SERVING": 20,
|
155 |
+
"I-SERVING_SIZE": 22,
|
156 |
+
"I-SODIUM_100G": 54,
|
157 |
+
"I-SODIUM_SERVING": 6,
|
158 |
+
"I-SUGARS_100G": 18,
|
159 |
+
"I-SUGARS_SERVING": 30,
|
160 |
+
"I-TRANS_FAT_100G": 46,
|
161 |
+
"I-TRANS_FAT_SERVING": 34,
|
162 |
+
"I-VITAMIN_D_100G": 64,
|
163 |
+
"I-VITAMIN_D_SERVING": 4,
|
164 |
"O": 0
|
165 |
},
|
166 |
"layer_norm_eps": 1e-05,
|
model.safetensors
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 1428562028
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:3f81a1228baf9eb80cebba31650adbaa161dd31dfa4ca315dc5802150cacc8ca
|
3 |
size 1428562028
|
runs/Dec23_15-55-22_gpu-l4-20241223-150914/events.out.tfevents.1734969342.gpu-l4-20241223-150914
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:eba48d44e1eb47cf14ea0f18c9a345ff8c4b8090ea3be4543aced2d39f7acd79
|
3 |
+
size 146043
|
tokenizer_config.json
CHANGED
@@ -55,10 +55,8 @@
|
|
55 |
"eos_token": "</s>",
|
56 |
"errors": "replace",
|
57 |
"mask_token": "<mask>",
|
58 |
-
"max_length": 512,
|
59 |
"model_max_length": 512,
|
60 |
"only_label_first_subword": true,
|
61 |
-
"pad_to_multiple_of": null,
|
62 |
"pad_token": "<pad>",
|
63 |
"pad_token_box": [
|
64 |
0,
|
@@ -67,8 +65,6 @@
|
|
67 |
0
|
68 |
],
|
69 |
"pad_token_label": -100,
|
70 |
-
"pad_token_type_id": 0,
|
71 |
-
"padding_side": "right",
|
72 |
"processor_class": "LayoutLMv3Processor",
|
73 |
"sep_token": "</s>",
|
74 |
"sep_token_box": [
|
@@ -77,11 +73,8 @@
|
|
77 |
0,
|
78 |
0
|
79 |
],
|
80 |
-
"stride": 0,
|
81 |
"tokenizer_class": "LayoutLMv3Tokenizer",
|
82 |
"trim_offsets": true,
|
83 |
-
"truncation_side": "right",
|
84 |
-
"truncation_strategy": "longest_first",
|
85 |
"unk_token": "<unk>",
|
86 |
"use_fast": true
|
87 |
}
|
|
|
55 |
"eos_token": "</s>",
|
56 |
"errors": "replace",
|
57 |
"mask_token": "<mask>",
|
|
|
58 |
"model_max_length": 512,
|
59 |
"only_label_first_subword": true,
|
|
|
60 |
"pad_token": "<pad>",
|
61 |
"pad_token_box": [
|
62 |
0,
|
|
|
65 |
0
|
66 |
],
|
67 |
"pad_token_label": -100,
|
|
|
|
|
68 |
"processor_class": "LayoutLMv3Processor",
|
69 |
"sep_token": "</s>",
|
70 |
"sep_token_box": [
|
|
|
73 |
0,
|
74 |
0
|
75 |
],
|
|
|
76 |
"tokenizer_class": "LayoutLMv3Tokenizer",
|
77 |
"trim_offsets": true,
|
|
|
|
|
78 |
"unk_token": "<unk>",
|
79 |
"use_fast": true
|
80 |
}
|
training_args.bin
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:db4952162d7c9a6e327afbfaaba51b2281e5e881119c9dca7a2fe5136e0449fe
|
3 |
+
size 5048
|