tomtseng commited on
Commit
42f5374
·
verified ·
1 Parent(s): b9cb157

Model save

Browse files
README.md CHANGED
@@ -14,6 +14,8 @@ should probably proofread and complete it, then remove this comment. -->
14
  # robust_llm_r2d2_tom-debug_pythia-14m_lora
15
 
16
  This model is a fine-tuned version of [EleutherAI/pythia-14m](https://huggingface.co/EleutherAI/pythia-14m) on an unknown dataset.
 
 
17
 
18
  ## Model description
19
 
@@ -41,6 +43,13 @@ The following hyperparameters were used during training:
41
  - lr_scheduler_type: cosine
42
  - training_steps: 1
43
 
 
 
 
 
 
 
 
44
  ### Framework versions
45
 
46
  - Transformers 4.37.2
 
14
  # robust_llm_r2d2_tom-debug_pythia-14m_lora
15
 
16
  This model is a fine-tuned version of [EleutherAI/pythia-14m](https://huggingface.co/EleutherAI/pythia-14m) on an unknown dataset.
17
+ It achieves the following results on the evaluation set:
18
+ - Loss: 6.2327
19
 
20
  ## Model description
21
 
 
43
  - lr_scheduler_type: cosine
44
  - training_steps: 1
45
 
46
+ ### Training results
47
+
48
+ | Training Loss | Epoch | Step | Validation Loss |
49
+ |:-------------:|:-----:|:----:|:---------------:|
50
+ | 9.125 | 0.0 | 1 | 6.2329 |
51
+
52
+
53
  ### Framework versions
54
 
55
  - Transformers 4.37.2
adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:b6657d938dee44bb28d765adcb19d3a0c428e055ca6ceab9ed26c3a22b160691
3
  size 198216
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b3d46417fab125b3c2e4fab028940f6a27ec11dff0d062f4d87ef1d710e57395
3
  size 198216
all_results.json ADDED
@@ -0,0 +1,13 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "epoch": 0.0,
3
+ "eval_loss": 6.23274040222168,
4
+ "eval_runtime": 52.0876,
5
+ "eval_samples": 23110,
6
+ "eval_samples_per_second": 443.676,
7
+ "eval_steps_per_second": 55.464,
8
+ "train_loss": 9.125,
9
+ "train_runtime": 57.0654,
10
+ "train_samples": 207865,
11
+ "train_samples_per_second": 0.07,
12
+ "train_steps_per_second": 0.018
13
+ }
eval_results.json ADDED
@@ -0,0 +1,8 @@
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "epoch": 0.0,
3
+ "eval_loss": 6.23274040222168,
4
+ "eval_runtime": 52.0876,
5
+ "eval_samples": 23110,
6
+ "eval_samples_per_second": 443.676,
7
+ "eval_steps_per_second": 55.464
8
+ }
runs/Feb05_19-16-43_r2d2-devbox-6vt8q/events.out.tfevents.1738783026.r2d2-devbox-6vt8q.6908.0 CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:59e3b7f43701f40e1f7616cd19288c53505543fd63577e8075ae31aa797ace12
3
- size 4608
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f0744892dea0535e4e419a44eb564261de427a2f8ace0c67a8521e65da59e526
3
+ size 5376
runs/Feb05_19-16-43_r2d2-devbox-6vt8q/events.out.tfevents.1738783135.r2d2-devbox-6vt8q.6908.1 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ca59dceda4459acdf537c83649d3357e85e94f0562a8821a4efab636b865572b
3
+ size 354
train_results.json ADDED
@@ -0,0 +1,8 @@
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "epoch": 0.0,
3
+ "train_loss": 9.125,
4
+ "train_runtime": 57.0654,
5
+ "train_samples": 207865,
6
+ "train_samples_per_second": 0.07,
7
+ "train_steps_per_second": 0.018
8
+ }
trainer_state.json ADDED
@@ -0,0 +1,44 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "best_metric": null,
3
+ "best_model_checkpoint": null,
4
+ "epoch": 1.9242981122635517e-05,
5
+ "eval_steps": 500,
6
+ "global_step": 1,
7
+ "is_hyper_param_search": false,
8
+ "is_local_process_zero": true,
9
+ "is_world_process_zero": true,
10
+ "log_history": [
11
+ {
12
+ "epoch": 0.0,
13
+ "learning_rate": 0.0,
14
+ "loss": 9.125,
15
+ "step": 1
16
+ },
17
+ {
18
+ "epoch": 0.0,
19
+ "eval_loss": 6.232901573181152,
20
+ "eval_runtime": 52.8258,
21
+ "eval_samples_per_second": 437.476,
22
+ "eval_steps_per_second": 54.689,
23
+ "step": 1
24
+ },
25
+ {
26
+ "epoch": 0.0,
27
+ "step": 1,
28
+ "total_flos": 379802615808.0,
29
+ "train_loss": 9.125,
30
+ "train_runtime": 57.0654,
31
+ "train_samples_per_second": 0.07,
32
+ "train_steps_per_second": 0.018
33
+ }
34
+ ],
35
+ "logging_steps": 5,
36
+ "max_steps": 1,
37
+ "num_input_tokens_seen": 0,
38
+ "num_train_epochs": 1,
39
+ "save_steps": 500,
40
+ "total_flos": 379802615808.0,
41
+ "train_batch_size": null,
42
+ "trial_name": null,
43
+ "trial_params": null
44
+ }