DorinSht commited on
Commit
f63a06e
·
verified ·
1 Parent(s): 4cb2959

End of training

Browse files
README.md CHANGED
@@ -3,6 +3,8 @@ license: apache-2.0
3
  base_model: JackFram/llama-68m
4
  tags:
5
  - generated_from_trainer
 
 
6
  model-index:
7
  - name: recreate_llama_68M_vanilla
8
  results: []
@@ -13,7 +15,10 @@ should probably proofread and complete it, then remove this comment. -->
13
 
14
  # recreate_llama_68M_vanilla
15
 
16
- This model is a fine-tuned version of [JackFram/llama-68m](https://huggingface.co/JackFram/llama-68m) on an unknown dataset.
 
 
 
17
 
18
  ## Model description
19
 
 
3
  base_model: JackFram/llama-68m
4
  tags:
5
  - generated_from_trainer
6
+ metrics:
7
+ - accuracy
8
  model-index:
9
  - name: recreate_llama_68M_vanilla
10
  results: []
 
15
 
16
  # recreate_llama_68M_vanilla
17
 
18
+ This model is a fine-tuned version of [JackFram/llama-68m](https://huggingface.co/JackFram/llama-68m) on the anon8231489123/ShareGPT_Vicuna_unfiltered/ShareGPT_V3_unfiltered_cleaned_split.json dataset.
19
+ It achieves the following results on the evaluation set:
20
+ - Loss: 4.3086
21
+ - Accuracy: 0.4590
22
 
23
  ## Model description
24
 
all_results.json CHANGED
@@ -1,16 +1,16 @@
1
  {
2
  "epoch": 3.0,
3
- "eval_accuracy": 0.3512457254518808,
4
- "eval_loss": 9.549426078796387,
5
- "eval_runtime": 2.7507,
6
  "eval_samples": 10,
7
- "eval_samples_per_second": 3.635,
8
- "eval_steps_per_second": 0.364,
9
- "perplexity": 14036.636436049854,
10
  "total_flos": 1601895923712000.0,
11
- "train_loss": 11.46164576212565,
12
- "train_runtime": 150.2617,
13
  "train_samples": 1000,
14
- "train_samples_per_second": 19.965,
15
- "train_steps_per_second": 0.639
16
  }
 
1
  {
2
  "epoch": 3.0,
3
+ "eval_accuracy": 0.45896433805569126,
4
+ "eval_loss": 4.308589935302734,
5
+ "eval_runtime": 2.8119,
6
  "eval_samples": 10,
7
+ "eval_samples_per_second": 3.556,
8
+ "eval_steps_per_second": 0.356,
9
+ "perplexity": 74.33559700443098,
10
  "total_flos": 1601895923712000.0,
11
+ "train_loss": 4.819753979879712,
12
+ "train_runtime": 153.0154,
13
  "train_samples": 1000,
14
+ "train_samples_per_second": 19.606,
15
+ "train_steps_per_second": 1.647
16
  }
args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:7565f26b6b3335b4c74c77cbb2af2017cb5dcb6983320e6a9a66a11f8e0ac811
3
- size 5908
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:da0612c3c9d86d6249df50bee087bc8118d35203b4807ae34f25b74705525d44
3
+ size 6036
eval_results.json CHANGED
@@ -1,10 +1,10 @@
1
  {
2
  "epoch": 3.0,
3
- "eval_accuracy": 0.3512457254518808,
4
- "eval_loss": 9.549426078796387,
5
- "eval_runtime": 2.7507,
6
  "eval_samples": 10,
7
- "eval_samples_per_second": 3.635,
8
- "eval_steps_per_second": 0.364,
9
- "perplexity": 14036.636436049854
10
  }
 
1
  {
2
  "epoch": 3.0,
3
+ "eval_accuracy": 0.45896433805569126,
4
+ "eval_loss": 4.308589935302734,
5
+ "eval_runtime": 2.8119,
6
  "eval_samples": 10,
7
+ "eval_samples_per_second": 3.556,
8
+ "eval_steps_per_second": 0.356,
9
+ "perplexity": 74.33559700443098
10
  }
events.out.tfevents.1716992485.isl-gpu27.3581638.1 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:85ffc476d8337347b72ca9cfa202beea19e8d9c599d44b60da44f727e63d6072
3
+ size 411
train_results.json CHANGED
@@ -1,9 +1,9 @@
1
  {
2
  "epoch": 3.0,
3
  "total_flos": 1601895923712000.0,
4
- "train_loss": 11.46164576212565,
5
- "train_runtime": 150.2617,
6
  "train_samples": 1000,
7
- "train_samples_per_second": 19.965,
8
- "train_steps_per_second": 0.639
9
  }
 
1
  {
2
  "epoch": 3.0,
3
  "total_flos": 1601895923712000.0,
4
+ "train_loss": 4.819753979879712,
5
+ "train_runtime": 153.0154,
6
  "train_samples": 1000,
7
+ "train_samples_per_second": 19.606,
8
+ "train_steps_per_second": 1.647
9
  }
trainer_state.json CHANGED
@@ -2,108 +2,27 @@
2
  "best_metric": null,
3
  "best_model_checkpoint": null,
4
  "epoch": 3.0,
5
- "eval_steps": 10,
6
- "global_step": 96,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
10
  "log_history": [
11
- {
12
- "epoch": 0.3125,
13
- "eval_accuracy": 0.367562286272594,
14
- "eval_loss": 7.9370436668396,
15
- "eval_runtime": 2.8708,
16
- "eval_samples_per_second": 3.483,
17
- "eval_steps_per_second": 0.348,
18
- "step": 10
19
- },
20
- {
21
- "epoch": 0.625,
22
- "eval_accuracy": 0.34782608695652173,
23
- "eval_loss": 8.680798530578613,
24
- "eval_runtime": 2.66,
25
- "eval_samples_per_second": 3.759,
26
- "eval_steps_per_second": 0.376,
27
- "step": 20
28
- },
29
- {
30
- "epoch": 0.9375,
31
- "eval_accuracy": 0.10293111871030777,
32
- "eval_loss": 10.979837417602539,
33
- "eval_runtime": 2.7406,
34
- "eval_samples_per_second": 3.649,
35
- "eval_steps_per_second": 0.365,
36
- "step": 30
37
- },
38
- {
39
- "epoch": 1.25,
40
- "eval_accuracy": 0.24934049829018076,
41
- "eval_loss": 10.302330017089844,
42
- "eval_runtime": 2.7648,
43
- "eval_samples_per_second": 3.617,
44
- "eval_steps_per_second": 0.362,
45
- "step": 40
46
- },
47
- {
48
- "epoch": 1.5625,
49
- "eval_accuracy": 0.35012212994626285,
50
- "eval_loss": 9.768780708312988,
51
- "eval_runtime": 2.7184,
52
- "eval_samples_per_second": 3.679,
53
- "eval_steps_per_second": 0.368,
54
- "step": 50
55
- },
56
- {
57
- "epoch": 1.875,
58
- "eval_accuracy": 0.35100146555935513,
59
- "eval_loss": 9.619012832641602,
60
- "eval_runtime": 2.8506,
61
- "eval_samples_per_second": 3.508,
62
- "eval_steps_per_second": 0.351,
63
- "step": 60
64
- },
65
- {
66
- "epoch": 2.1875,
67
- "eval_accuracy": 0.35095261358085,
68
- "eval_loss": 9.561655044555664,
69
- "eval_runtime": 2.7142,
70
- "eval_samples_per_second": 3.684,
71
- "eval_steps_per_second": 0.368,
72
- "step": 70
73
- },
74
- {
75
- "epoch": 2.5,
76
- "eval_accuracy": 0.35109916951636544,
77
- "eval_loss": 9.547042846679688,
78
- "eval_runtime": 2.8217,
79
- "eval_samples_per_second": 3.544,
80
- "eval_steps_per_second": 0.354,
81
- "step": 80
82
- },
83
- {
84
- "epoch": 2.8125,
85
- "eval_accuracy": 0.35109916951636544,
86
- "eval_loss": 9.548730850219727,
87
- "eval_runtime": 2.742,
88
- "eval_samples_per_second": 3.647,
89
- "eval_steps_per_second": 0.365,
90
- "step": 90
91
- },
92
  {
93
  "epoch": 3.0,
94
- "step": 96,
95
  "total_flos": 1601895923712000.0,
96
- "train_loss": 11.46164576212565,
97
- "train_runtime": 150.2617,
98
- "train_samples_per_second": 19.965,
99
- "train_steps_per_second": 0.639
100
  }
101
  ],
102
- "logging_steps": 100,
103
- "max_steps": 96,
104
  "num_input_tokens_seen": 0,
105
  "num_train_epochs": 3,
106
- "save_steps": 100,
107
  "stateful_callbacks": {
108
  "TrainerControl": {
109
  "args": {
@@ -117,7 +36,7 @@
117
  }
118
  },
119
  "total_flos": 1601895923712000.0,
120
- "train_batch_size": 32,
121
  "trial_name": null,
122
  "trial_params": null
123
  }
 
2
  "best_metric": null,
3
  "best_model_checkpoint": null,
4
  "epoch": 3.0,
5
+ "eval_steps": 1000,
6
+ "global_step": 252,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
10
  "log_history": [
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
11
  {
12
  "epoch": 3.0,
13
+ "step": 252,
14
  "total_flos": 1601895923712000.0,
15
+ "train_loss": 4.819753979879712,
16
+ "train_runtime": 153.0154,
17
+ "train_samples_per_second": 19.606,
18
+ "train_steps_per_second": 1.647
19
  }
20
  ],
21
+ "logging_steps": 500,
22
+ "max_steps": 252,
23
  "num_input_tokens_seen": 0,
24
  "num_train_epochs": 3,
25
+ "save_steps": 10000,
26
  "stateful_callbacks": {
27
  "TrainerControl": {
28
  "args": {
 
36
  }
37
  },
38
  "total_flos": 1601895923712000.0,
39
+ "train_batch_size": 12,
40
  "trial_name": null,
41
  "trial_params": null
42
  }