hjctty commited on
Commit
77a242d
1 Parent(s): 307d807

Model save

Browse files
README.md CHANGED
@@ -18,9 +18,9 @@ should probably proofread and complete it, then remove this comment. -->
18
 
19
  This model is a fine-tuned version of [sentence-transformers/all-distilroberta-v1](https://huggingface.co/sentence-transformers/all-distilroberta-v1) on the None dataset.
20
  It achieves the following results on the evaluation set:
21
- - Loss: 1.0505
22
- - Accuracy: 1.0
23
- - F1: 1.0
24
 
25
  ## Model description
26
 
@@ -52,16 +52,16 @@ The following hyperparameters were used during training:
52
 
53
  | Training Loss | Epoch | Step | Validation Loss | Accuracy | F1 |
54
  |:-------------:|:-----:|:----:|:---------------:|:--------:|:------:|
55
- | 4.8137 | 1.0 | 30 | 4.2195 | 0.5 | 0.4046 |
56
- | 3.8117 | 2.0 | 60 | 3.3029 | 0.7667 | 0.7270 |
57
- | 3.0598 | 3.0 | 90 | 2.6420 | 0.8333 | 0.8137 |
58
- | 2.4888 | 4.0 | 120 | 2.1405 | 0.8667 | 0.8470 |
59
- | 2.0583 | 5.0 | 150 | 1.7464 | 0.9 | 0.8825 |
60
- | 1.6867 | 6.0 | 180 | 1.4774 | 0.9667 | 0.9644 |
61
- | 1.4817 | 7.0 | 210 | 1.2812 | 0.9667 | 0.9644 |
62
- | 1.2876 | 8.0 | 240 | 1.1481 | 1.0 | 1.0 |
63
- | 1.1833 | 9.0 | 270 | 1.0764 | 1.0 | 1.0 |
64
- | 1.1549 | 10.0 | 300 | 1.0505 | 1.0 | 1.0 |
65
 
66
 
67
  ### Framework versions
 
18
 
19
  This model is a fine-tuned version of [sentence-transformers/all-distilroberta-v1](https://huggingface.co/sentence-transformers/all-distilroberta-v1) on the None dataset.
20
  It achieves the following results on the evaluation set:
21
+ - Loss: 1.0841
22
+ - Accuracy: 0.9333
23
+ - F1: 0.9289
24
 
25
  ## Model description
26
 
 
52
 
53
  | Training Loss | Epoch | Step | Validation Loss | Accuracy | F1 |
54
  |:-------------:|:-----:|:----:|:---------------:|:--------:|:------:|
55
+ | 5.0075 | 1.0 | 30 | 4.3968 | 0.5 | 0.44 |
56
+ | 3.9817 | 2.0 | 60 | 3.4466 | 0.8333 | 0.8114 |
57
+ | 3.248 | 3.0 | 90 | 2.7491 | 0.8667 | 0.8470 |
58
+ | 2.6848 | 4.0 | 120 | 2.2106 | 0.8667 | 0.8470 |
59
+ | 2.2175 | 5.0 | 150 | 1.8139 | 0.9 | 0.8978 |
60
+ | 1.8263 | 6.0 | 180 | 1.5195 | 0.9 | 0.8978 |
61
+ | 1.6005 | 7.0 | 210 | 1.3236 | 0.9333 | 0.9289 |
62
+ | 1.4191 | 8.0 | 240 | 1.1857 | 0.9333 | 0.9289 |
63
+ | 1.3119 | 9.0 | 270 | 1.1092 | 0.9333 | 0.9289 |
64
+ | 1.2529 | 10.0 | 300 | 1.0841 | 0.9333 | 0.9289 |
65
 
66
 
67
  ### Framework versions
config.json CHANGED
@@ -190,7 +190,58 @@
190
  "175": "LABEL_175",
191
  "176": "LABEL_176",
192
  "177": "LABEL_177",
193
- "178": "LABEL_178"
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
194
  },
195
  "initializer_range": 0.02,
196
  "intermediate_size": 3072,
@@ -284,12 +335,63 @@
284
  "LABEL_176": 176,
285
  "LABEL_177": 177,
286
  "LABEL_178": 178,
 
287
  "LABEL_18": 18,
 
 
 
 
 
 
 
 
 
 
288
  "LABEL_19": 19,
 
 
 
 
 
 
 
 
 
 
289
  "LABEL_2": 2,
290
  "LABEL_20": 20,
 
 
 
 
 
 
 
 
 
 
291
  "LABEL_21": 21,
 
 
 
 
 
 
 
 
 
 
292
  "LABEL_22": 22,
 
 
 
 
 
 
 
 
 
 
293
  "LABEL_23": 23,
294
  "LABEL_24": 24,
295
  "LABEL_25": 25,
 
190
  "175": "LABEL_175",
191
  "176": "LABEL_176",
192
  "177": "LABEL_177",
193
+ "178": "LABEL_178",
194
+ "179": "LABEL_179",
195
+ "180": "LABEL_180",
196
+ "181": "LABEL_181",
197
+ "182": "LABEL_182",
198
+ "183": "LABEL_183",
199
+ "184": "LABEL_184",
200
+ "185": "LABEL_185",
201
+ "186": "LABEL_186",
202
+ "187": "LABEL_187",
203
+ "188": "LABEL_188",
204
+ "189": "LABEL_189",
205
+ "190": "LABEL_190",
206
+ "191": "LABEL_191",
207
+ "192": "LABEL_192",
208
+ "193": "LABEL_193",
209
+ "194": "LABEL_194",
210
+ "195": "LABEL_195",
211
+ "196": "LABEL_196",
212
+ "197": "LABEL_197",
213
+ "198": "LABEL_198",
214
+ "199": "LABEL_199",
215
+ "200": "LABEL_200",
216
+ "201": "LABEL_201",
217
+ "202": "LABEL_202",
218
+ "203": "LABEL_203",
219
+ "204": "LABEL_204",
220
+ "205": "LABEL_205",
221
+ "206": "LABEL_206",
222
+ "207": "LABEL_207",
223
+ "208": "LABEL_208",
224
+ "209": "LABEL_209",
225
+ "210": "LABEL_210",
226
+ "211": "LABEL_211",
227
+ "212": "LABEL_212",
228
+ "213": "LABEL_213",
229
+ "214": "LABEL_214",
230
+ "215": "LABEL_215",
231
+ "216": "LABEL_216",
232
+ "217": "LABEL_217",
233
+ "218": "LABEL_218",
234
+ "219": "LABEL_219",
235
+ "220": "LABEL_220",
236
+ "221": "LABEL_221",
237
+ "222": "LABEL_222",
238
+ "223": "LABEL_223",
239
+ "224": "LABEL_224",
240
+ "225": "LABEL_225",
241
+ "226": "LABEL_226",
242
+ "227": "LABEL_227",
243
+ "228": "LABEL_228",
244
+ "229": "LABEL_229"
245
  },
246
  "initializer_range": 0.02,
247
  "intermediate_size": 3072,
 
335
  "LABEL_176": 176,
336
  "LABEL_177": 177,
337
  "LABEL_178": 178,
338
+ "LABEL_179": 179,
339
  "LABEL_18": 18,
340
+ "LABEL_180": 180,
341
+ "LABEL_181": 181,
342
+ "LABEL_182": 182,
343
+ "LABEL_183": 183,
344
+ "LABEL_184": 184,
345
+ "LABEL_185": 185,
346
+ "LABEL_186": 186,
347
+ "LABEL_187": 187,
348
+ "LABEL_188": 188,
349
+ "LABEL_189": 189,
350
  "LABEL_19": 19,
351
+ "LABEL_190": 190,
352
+ "LABEL_191": 191,
353
+ "LABEL_192": 192,
354
+ "LABEL_193": 193,
355
+ "LABEL_194": 194,
356
+ "LABEL_195": 195,
357
+ "LABEL_196": 196,
358
+ "LABEL_197": 197,
359
+ "LABEL_198": 198,
360
+ "LABEL_199": 199,
361
  "LABEL_2": 2,
362
  "LABEL_20": 20,
363
+ "LABEL_200": 200,
364
+ "LABEL_201": 201,
365
+ "LABEL_202": 202,
366
+ "LABEL_203": 203,
367
+ "LABEL_204": 204,
368
+ "LABEL_205": 205,
369
+ "LABEL_206": 206,
370
+ "LABEL_207": 207,
371
+ "LABEL_208": 208,
372
+ "LABEL_209": 209,
373
  "LABEL_21": 21,
374
+ "LABEL_210": 210,
375
+ "LABEL_211": 211,
376
+ "LABEL_212": 212,
377
+ "LABEL_213": 213,
378
+ "LABEL_214": 214,
379
+ "LABEL_215": 215,
380
+ "LABEL_216": 216,
381
+ "LABEL_217": 217,
382
+ "LABEL_218": 218,
383
+ "LABEL_219": 219,
384
  "LABEL_22": 22,
385
+ "LABEL_220": 220,
386
+ "LABEL_221": 221,
387
+ "LABEL_222": 222,
388
+ "LABEL_223": 223,
389
+ "LABEL_224": 224,
390
+ "LABEL_225": 225,
391
+ "LABEL_226": 226,
392
+ "LABEL_227": 227,
393
+ "LABEL_228": 228,
394
+ "LABEL_229": 229,
395
  "LABEL_23": 23,
396
  "LABEL_24": 24,
397
  "LABEL_25": 25,
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:013b51ccecd39c3bd3cb3db7449762b4555b0be35d3b51cf4f5b69c0e0b7dec2
3
- size 329036732
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:50a76763f7513054b4fbe321dbb0755d5758d9d98be27e9c416dc8ee61796572
3
+ size 329193608
runs/Apr02_22-19-28_fsm-hjason-text-test-t4/events.out.tfevents.1712096373.fsm-hjason-text-test-t4 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:af07cf7b7924657a40c099a21e381c007bf5e80e13592e9a8eca213e739744d3
3
+ size 20939
runs/Apr02_22-19-28_fsm-hjason-text-test-t4/events.out.tfevents.1712096872.fsm-hjason-text-test-t4 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ed73a09c4ed5b5754b23b39e91ad23746bf179c222b718e8cebb9b6067d9de84
3
+ size 15447
runs/Apr02_22-28-49_fsm-hjason-text-test-t4/events.out.tfevents.1712096934.fsm-hjason-text-test-t4 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8399c750d5dee008287cbfc1019bed4b7d871b9fa3913dac6bcf0b8a0c44af23
3
+ size 20939
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:6329be7baaaeefee665abe38784efabd4feb0b0fc7bffa9aab85fb58e0631d15
3
  size 4920
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ff00a4a5fb5441a99ee408fdf9084c7e5163850d54ea142786b654eec1261868
3
  size 4920