TinyPixel commited on
Commit
6eac6ad
·
1 Parent(s): e28ce5d

Upload folder using huggingface_hub

Browse files
adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:c11e80c2626b7e725c36b5d59e4313d238413de8eebfdb5b1d6b6e28ab8c714b
3
  size 113271504
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:aea01fe0dd2af8bf4b4d060ddad5d472a16722fc394fcb4b38e130025fbdd9a6
3
  size 113271504
optimizer.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:87d135c01136a11bcef2ed383e09bc32d52fff97352f14798fe440aa0238e742
3
  size 226609018
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a6dfd7167bc860806832aeae1caf4b2e1b91ea75533b6f21aba040240609424c
3
  size 226609018
rng_state.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:b3698f87a076ace386d58713681a24c766465261225a60a74e4631003e3b501c
3
  size 14244
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8be5a9156a087de7426b7455c1cfa58b09ad84e7aaa003d8794e888b3cdd82ad
3
  size 14244
scheduler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:4372b8bc129cb832d61f89de539110be4b375eeb5fe693144c4dd880623a9aaf
3
  size 1064
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3bf1e902c3114bb5312c2c178dc4d942be3a06cb5931cabbb4ddb7748d0f53ee
3
  size 1064
trainer_state.json CHANGED
@@ -1,164 +1,218 @@
1
  {
2
  "best_metric": null,
3
  "best_model_checkpoint": null,
4
- "epoch": 0.9846153846153847,
5
  "eval_steps": 500,
6
- "global_step": 48,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
10
  "log_history": [
11
  {
12
- "epoch": 0.04,
13
  "learning_rate": 2e-05,
14
- "loss": 1.6235,
15
  "step": 2
16
  },
17
  {
18
- "epoch": 0.08,
19
  "learning_rate": 2e-05,
20
- "loss": 1.878,
21
  "step": 4
22
  },
23
  {
24
- "epoch": 0.12,
25
  "learning_rate": 2e-05,
26
- "loss": 1.8871,
27
  "step": 6
28
  },
29
  {
30
- "epoch": 0.16,
31
  "learning_rate": 2e-05,
32
- "loss": 1.9954,
33
  "step": 8
34
  },
35
  {
36
- "epoch": 0.21,
37
  "learning_rate": 2e-05,
38
- "loss": 2.0806,
39
  "step": 10
40
  },
41
  {
42
- "epoch": 0.25,
43
  "learning_rate": 2e-05,
44
- "loss": 2.0435,
45
  "step": 12
46
  },
47
  {
48
- "epoch": 0.29,
49
  "learning_rate": 2e-05,
50
- "loss": 1.6826,
51
  "step": 14
52
  },
53
  {
54
- "epoch": 0.33,
55
  "learning_rate": 2e-05,
56
- "loss": 1.5779,
57
  "step": 16
58
  },
59
  {
60
- "epoch": 0.37,
61
  "learning_rate": 2e-05,
62
- "loss": 1.9058,
63
  "step": 18
64
  },
65
  {
66
- "epoch": 0.41,
67
  "learning_rate": 2e-05,
68
- "loss": 1.6972,
69
  "step": 20
70
  },
71
  {
72
- "epoch": 0.45,
73
  "learning_rate": 2e-05,
74
- "loss": 1.9398,
75
  "step": 22
76
  },
77
  {
78
- "epoch": 0.49,
79
  "learning_rate": 2e-05,
80
- "loss": 1.9236,
81
  "step": 24
82
  },
83
  {
84
- "epoch": 0.53,
85
  "learning_rate": 2e-05,
86
- "loss": 1.6062,
87
  "step": 26
88
  },
89
  {
90
- "epoch": 0.57,
91
  "learning_rate": 2e-05,
92
- "loss": 1.7559,
93
  "step": 28
94
  },
95
  {
96
- "epoch": 0.62,
97
  "learning_rate": 2e-05,
98
- "loss": 1.9474,
99
  "step": 30
100
  },
101
  {
102
- "epoch": 0.66,
103
  "learning_rate": 2e-05,
104
- "loss": 1.7649,
105
  "step": 32
106
  },
107
  {
108
- "epoch": 0.7,
109
  "learning_rate": 2e-05,
110
- "loss": 1.7762,
111
  "step": 34
112
  },
113
  {
114
- "epoch": 0.74,
115
  "learning_rate": 2e-05,
116
- "loss": 2.011,
117
  "step": 36
118
  },
119
  {
120
- "epoch": 0.78,
121
  "learning_rate": 2e-05,
122
- "loss": 1.6693,
123
  "step": 38
124
  },
125
  {
126
- "epoch": 0.82,
127
  "learning_rate": 2e-05,
128
- "loss": 1.8561,
129
  "step": 40
130
  },
131
  {
132
- "epoch": 0.86,
133
  "learning_rate": 2e-05,
134
- "loss": 1.9452,
135
  "step": 42
136
  },
137
  {
138
- "epoch": 0.9,
139
  "learning_rate": 2e-05,
140
- "loss": 1.9725,
141
  "step": 44
142
  },
143
  {
144
- "epoch": 0.94,
145
  "learning_rate": 2e-05,
146
- "loss": 1.8454,
147
  "step": 46
148
  },
149
  {
150
- "epoch": 0.98,
151
  "learning_rate": 2e-05,
152
- "loss": 1.8836,
153
  "step": 48
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
154
  }
155
  ],
156
  "logging_steps": 2,
157
- "max_steps": 144,
158
  "num_input_tokens_seen": 0,
159
- "num_train_epochs": 3,
160
  "save_steps": 500,
161
- "total_flos": 1.901574003725107e+16,
162
  "train_batch_size": 1,
163
  "trial_name": null,
164
  "trial_params": null
 
1
  {
2
  "best_metric": null,
3
  "best_model_checkpoint": null,
4
+ "epoch": 0.9990680335507922,
5
  "eval_steps": 500,
6
+ "global_step": 67,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
10
  "log_history": [
11
  {
12
+ "epoch": 0.03,
13
  "learning_rate": 2e-05,
14
+ "loss": 2.0684,
15
  "step": 2
16
  },
17
  {
18
+ "epoch": 0.06,
19
  "learning_rate": 2e-05,
20
+ "loss": 1.2351,
21
  "step": 4
22
  },
23
  {
24
+ "epoch": 0.09,
25
  "learning_rate": 2e-05,
26
+ "loss": 1.0218,
27
  "step": 6
28
  },
29
  {
30
+ "epoch": 0.12,
31
  "learning_rate": 2e-05,
32
+ "loss": 0.944,
33
  "step": 8
34
  },
35
  {
36
+ "epoch": 0.15,
37
  "learning_rate": 2e-05,
38
+ "loss": 1.0419,
39
  "step": 10
40
  },
41
  {
42
+ "epoch": 0.18,
43
  "learning_rate": 2e-05,
44
+ "loss": 1.0514,
45
  "step": 12
46
  },
47
  {
48
+ "epoch": 0.21,
49
  "learning_rate": 2e-05,
50
+ "loss": 1.4716,
51
  "step": 14
52
  },
53
  {
54
+ "epoch": 0.24,
55
  "learning_rate": 2e-05,
56
+ "loss": 1.9171,
57
  "step": 16
58
  },
59
  {
60
+ "epoch": 0.27,
61
  "learning_rate": 2e-05,
62
+ "loss": 1.8928,
63
  "step": 18
64
  },
65
  {
66
+ "epoch": 0.3,
67
  "learning_rate": 2e-05,
68
+ "loss": 1.0043,
69
  "step": 20
70
  },
71
  {
72
+ "epoch": 0.33,
73
  "learning_rate": 2e-05,
74
+ "loss": 0.9942,
75
  "step": 22
76
  },
77
  {
78
+ "epoch": 0.36,
79
  "learning_rate": 2e-05,
80
+ "loss": 1.0396,
81
  "step": 24
82
  },
83
  {
84
+ "epoch": 0.39,
85
  "learning_rate": 2e-05,
86
+ "loss": 0.7804,
87
  "step": 26
88
  },
89
  {
90
+ "epoch": 0.42,
91
  "learning_rate": 2e-05,
92
+ "loss": 1.0986,
93
  "step": 28
94
  },
95
  {
96
+ "epoch": 0.45,
97
  "learning_rate": 2e-05,
98
+ "loss": 1.3254,
99
  "step": 30
100
  },
101
  {
102
+ "epoch": 0.48,
103
  "learning_rate": 2e-05,
104
+ "loss": 1.4412,
105
  "step": 32
106
  },
107
  {
108
+ "epoch": 0.51,
109
  "learning_rate": 2e-05,
110
+ "loss": 1.533,
111
  "step": 34
112
  },
113
  {
114
+ "epoch": 0.54,
115
  "learning_rate": 2e-05,
116
+ "loss": 1.2709,
117
  "step": 36
118
  },
119
  {
120
+ "epoch": 0.57,
121
  "learning_rate": 2e-05,
122
+ "loss": 0.8853,
123
  "step": 38
124
  },
125
  {
126
+ "epoch": 0.6,
127
  "learning_rate": 2e-05,
128
+ "loss": 1.0464,
129
  "step": 40
130
  },
131
  {
132
+ "epoch": 0.63,
133
  "learning_rate": 2e-05,
134
+ "loss": 0.8004,
135
  "step": 42
136
  },
137
  {
138
+ "epoch": 0.66,
139
  "learning_rate": 2e-05,
140
+ "loss": 0.8526,
141
  "step": 44
142
  },
143
  {
144
+ "epoch": 0.69,
145
  "learning_rate": 2e-05,
146
+ "loss": 1.1846,
147
  "step": 46
148
  },
149
  {
150
+ "epoch": 0.72,
151
  "learning_rate": 2e-05,
152
+ "loss": 1.3307,
153
  "step": 48
154
+ },
155
+ {
156
+ "epoch": 0.75,
157
+ "learning_rate": 2e-05,
158
+ "loss": 1.9532,
159
+ "step": 50
160
+ },
161
+ {
162
+ "epoch": 0.78,
163
+ "learning_rate": 2e-05,
164
+ "loss": 0.9404,
165
+ "step": 52
166
+ },
167
+ {
168
+ "epoch": 0.81,
169
+ "learning_rate": 2e-05,
170
+ "loss": 0.9675,
171
+ "step": 54
172
+ },
173
+ {
174
+ "epoch": 0.84,
175
+ "learning_rate": 2e-05,
176
+ "loss": 0.8579,
177
+ "step": 56
178
+ },
179
+ {
180
+ "epoch": 0.86,
181
+ "learning_rate": 2e-05,
182
+ "loss": 0.9398,
183
+ "step": 58
184
+ },
185
+ {
186
+ "epoch": 0.89,
187
+ "learning_rate": 2e-05,
188
+ "loss": 1.045,
189
+ "step": 60
190
+ },
191
+ {
192
+ "epoch": 0.92,
193
+ "learning_rate": 2e-05,
194
+ "loss": 1.1489,
195
+ "step": 62
196
+ },
197
+ {
198
+ "epoch": 0.95,
199
+ "learning_rate": 2e-05,
200
+ "loss": 1.1677,
201
+ "step": 64
202
+ },
203
+ {
204
+ "epoch": 0.98,
205
+ "learning_rate": 2e-05,
206
+ "loss": 1.1349,
207
+ "step": 66
208
  }
209
  ],
210
  "logging_steps": 2,
211
+ "max_steps": 67,
212
  "num_input_tokens_seen": 0,
213
+ "num_train_epochs": 1,
214
  "save_steps": 500,
215
+ "total_flos": 2.539821361058611e+16,
216
  "train_batch_size": 1,
217
  "trial_name": null,
218
  "trial_params": null
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:1142229c66bf3f832461258a15f379deffb2b5691470e753d093d27a4adb154c
3
  size 4728
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:97d5f0e37cf4b722c6518f8cfa86a35259d2a47813b4394ade22d5e08cffdbf4
3
  size 4728