PaulD commited on
Commit
0696d05
·
verified ·
1 Parent(s): 86c8ba0

End of training

Browse files
adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:84271dee3b3d53351c0ba8b484aeca2de76aee8a340e11871c8ba0b701b8b666
3
  size 27297544
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f7b11766ec08896f64ee4937ebfa4d3eb2adda32ab24eac2de94ad815b7eb92c
3
  size 27297544
metrics.jsonl CHANGED
@@ -30,3 +30,9 @@
30
  {"epoch": 4.0, "precision": 0.9999999500000026, "recall": 0.4999999875000003, "fold": 0}
31
  {"epoch": 4.9411764705882355, "precision": 0.9999999500000026, "recall": 0.4999999875000003, "fold": 0}
32
  {"epoch": 5.647058823529412, "precision": 0.9999999000000099, "recall": 0.24999999375000015, "fold": 0}
 
 
 
 
 
 
 
30
  {"epoch": 4.0, "precision": 0.9999999500000026, "recall": 0.4999999875000003, "fold": 0}
31
  {"epoch": 4.9411764705882355, "precision": 0.9999999500000026, "recall": 0.4999999875000003, "fold": 0}
32
  {"epoch": 5.647058823529412, "precision": 0.9999999000000099, "recall": 0.24999999375000015, "fold": 0}
33
+ {"epoch": 0.9411764705882353, "precision": 0.4999999937500001, "recall": 0.9999999750000006, "fold": 0}
34
+ {"epoch": 1.9607843137254903, "precision": 0.6666666555555557, "recall": 0.9999999750000006, "fold": 0}
35
+ {"epoch": 2.980392156862745, "precision": 0.9999999666666678, "recall": 0.7499999812500004, "fold": 0}
36
+ {"epoch": 4.0, "precision": 0.9999999500000026, "recall": 0.4999999875000003, "fold": 0}
37
+ {"epoch": 4.9411764705882355, "precision": 0.9999999500000026, "recall": 0.4999999875000003, "fold": 0}
38
+ {"epoch": 5.647058823529412, "precision": 0.9999999666666678, "recall": 0.7499999812500004, "fold": 0}
metrics_epoch_0.9411764705882353_fold_0_lr_1e-05_seed_9012_weight_1.5.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"epoch": 0.9411764705882353, "precision": 0.4999999937500001, "recall": 0.9999999750000006, "fold": 0}
metrics_epoch_1.9607843137254903_fold_0_lr_1e-05_seed_9012_weight_1.5.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"epoch": 1.9607843137254903, "precision": 0.6666666555555557, "recall": 0.9999999750000006, "fold": 0}
metrics_epoch_2.980392156862745_fold_0_lr_1e-05_seed_9012_weight_1.5.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"epoch": 2.980392156862745, "precision": 0.9999999666666678, "recall": 0.7499999812500004, "fold": 0}
metrics_epoch_4.0_fold_0_lr_1e-05_seed_9012_weight_1.5.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"epoch": 4.0, "precision": 0.9999999500000026, "recall": 0.4999999875000003, "fold": 0}
metrics_epoch_4.9411764705882355_fold_0_lr_1e-05_seed_9012_weight_1.5.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"epoch": 4.9411764705882355, "precision": 0.9999999500000026, "recall": 0.4999999875000003, "fold": 0}
metrics_epoch_5.647058823529412_fold_0_lr_1e-05_seed_9012_weight_1.5.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"epoch": 5.647058823529412, "precision": 0.9999999666666678, "recall": 0.7499999812500004, "fold": 0}
results_epoch_0.9411764705882353_fold_0_lr_1e-05_seed_9012_weight_1.5.json ADDED
The diff for this file is too large to render. See raw diff
 
results_epoch_1.9607843137254903_fold_0_lr_1e-05_seed_9012_weight_1.5.json ADDED
The diff for this file is too large to render. See raw diff
 
results_epoch_2.980392156862745_fold_0_lr_1e-05_seed_9012_weight_1.5.json ADDED
The diff for this file is too large to render. See raw diff
 
results_epoch_4.0_fold_0_lr_1e-05_seed_9012_weight_1.5.json ADDED
The diff for this file is too large to render. See raw diff
 
results_epoch_4.9411764705882355_fold_0_lr_1e-05_seed_9012_weight_1.5.json ADDED
The diff for this file is too large to render. See raw diff
 
results_epoch_5.647058823529412_fold_0_lr_1e-05_seed_9012_weight_1.5.json ADDED
The diff for this file is too large to render. See raw diff
 
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:be742e23842dd3c340c21c2791e9513464778733bf2c64b177f7e57a55ebe433
3
  size 5368
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b881753a9a12110010367823330ea6dea9e687b2d02fbb12b709b8b7a3a69aa8
3
  size 5368