chansung commited on
Commit
d17bcbc
1 Parent(s): 0238403

Model save

Browse files
Files changed (4) hide show
  1. README.md +4 -4
  2. all_results.json +4 -9
  3. train_results.json +4 -4
  4. trainer_state.json +60 -60
README.md CHANGED
@@ -20,7 +20,7 @@ should probably proofread and complete it, then remove this comment. -->
20
 
21
  This model is a fine-tuned version of [mistralai/Mistral-7B-v0.3](https://huggingface.co/mistralai/Mistral-7B-v0.3) on the generator dataset.
22
  It achieves the following results on the evaluation set:
23
- - Loss: 1.8814
24
 
25
  ## Model description
26
 
@@ -48,7 +48,7 @@ The following hyperparameters were used during training:
48
  - gradient_accumulation_steps: 2
49
  - total_train_batch_size: 224
50
  - total_eval_batch_size: 112
51
- - optimizer: Use OptimizerNames.ADAMW_TORCH with betas=(0.9,0.999) and epsilon=1e-08 and optimizer_args=No additional optimizer arguments
52
  - lr_scheduler_type: cosine
53
  - lr_scheduler_warmup_ratio: 0.1
54
  - num_epochs: 1
@@ -57,13 +57,13 @@ The following hyperparameters were used during training:
57
 
58
  | Training Loss | Epoch | Step | Validation Loss |
59
  |:-------------:|:------:|:----:|:---------------:|
60
- | 1.1865 | 0.9970 | 168 | 1.8814 |
61
 
62
 
63
  ### Framework versions
64
 
65
  - PEFT 0.13.2
66
  - Transformers 4.46.2
67
- - Pytorch 2.3.1+cu121
68
  - Datasets 3.1.0
69
  - Tokenizers 0.20.3
 
20
 
21
  This model is a fine-tuned version of [mistralai/Mistral-7B-v0.3](https://huggingface.co/mistralai/Mistral-7B-v0.3) on the generator dataset.
22
  It achieves the following results on the evaluation set:
23
+ - Loss: 1.8820
24
 
25
  ## Model description
26
 
 
48
  - gradient_accumulation_steps: 2
49
  - total_train_batch_size: 224
50
  - total_eval_batch_size: 112
51
+ - optimizer: Use adamw_torch with betas=(0.9,0.999) and epsilon=1e-08 and optimizer_args=No additional optimizer arguments
52
  - lr_scheduler_type: cosine
53
  - lr_scheduler_warmup_ratio: 0.1
54
  - num_epochs: 1
 
57
 
58
  | Training Loss | Epoch | Step | Validation Loss |
59
  |:-------------:|:------:|:----:|:---------------:|
60
+ | 1.1865 | 0.9970 | 168 | 1.8820 |
61
 
62
 
63
  ### Framework versions
64
 
65
  - PEFT 0.13.2
66
  - Transformers 4.46.2
67
+ - Pytorch 2.5.1+cu124
68
  - Datasets 3.1.0
69
  - Tokenizers 0.20.3
all_results.json CHANGED
@@ -1,14 +1,9 @@
1
  {
2
  "epoch": 0.9970326409495549,
3
- "eval_loss": 1.8814080953598022,
4
- "eval_runtime": 0.6678,
5
- "eval_samples": 15,
6
- "eval_samples_per_second": 20.965,
7
- "eval_steps_per_second": 1.498,
8
  "total_flos": 8.227900459906499e+17,
9
- "train_loss": 1.2773614100047521,
10
- "train_runtime": 661.1188,
11
  "train_samples": 111440,
12
- "train_samples_per_second": 57.022,
13
- "train_steps_per_second": 0.254
14
  }
 
1
  {
2
  "epoch": 0.9970326409495549,
 
 
 
 
 
3
  "total_flos": 8.227900459906499e+17,
4
+ "train_loss": 1.2774092711153484,
5
+ "train_runtime": 653.4451,
6
  "train_samples": 111440,
7
+ "train_samples_per_second": 57.691,
8
+ "train_steps_per_second": 0.257
9
  }
train_results.json CHANGED
@@ -1,9 +1,9 @@
1
  {
2
  "epoch": 0.9970326409495549,
3
  "total_flos": 8.227900459906499e+17,
4
- "train_loss": 1.2773614100047521,
5
- "train_runtime": 661.1188,
6
  "train_samples": 111440,
7
- "train_samples_per_second": 57.022,
8
- "train_steps_per_second": 0.254
9
  }
 
1
  {
2
  "epoch": 0.9970326409495549,
3
  "total_flos": 8.227900459906499e+17,
4
+ "train_loss": 1.2774092711153484,
5
+ "train_runtime": 653.4451,
6
  "train_samples": 111440,
7
+ "train_samples_per_second": 57.691,
8
+ "train_steps_per_second": 0.257
9
  }
trainer_state.json CHANGED
@@ -10,258 +10,258 @@
10
  "log_history": [
11
  {
12
  "epoch": 0.005934718100890208,
13
- "grad_norm": 3.172546863555908,
14
  "learning_rate": 1.1764705882352942e-05,
15
  "loss": 1.7951,
16
  "step": 1
17
  },
18
  {
19
  "epoch": 0.02967359050445104,
20
- "grad_norm": 3.243614435195923,
21
  "learning_rate": 5.882352941176471e-05,
22
  "loss": 1.8001,
23
  "step": 5
24
  },
25
  {
26
  "epoch": 0.05934718100890208,
27
- "grad_norm": 2.237574577331543,
28
  "learning_rate": 0.00011764705882352942,
29
  "loss": 1.7245,
30
  "step": 10
31
  },
32
  {
33
  "epoch": 0.08902077151335312,
34
- "grad_norm": 2.2595160007476807,
35
  "learning_rate": 0.00017647058823529413,
36
- "loss": 1.6176,
37
  "step": 15
38
  },
39
  {
40
  "epoch": 0.11869436201780416,
41
- "grad_norm": 2.2810211181640625,
42
  "learning_rate": 0.00019980527694749952,
43
- "loss": 1.4882,
44
  "step": 20
45
  },
46
  {
47
  "epoch": 0.14836795252225518,
48
- "grad_norm": 2.004383087158203,
49
  "learning_rate": 0.00019861804788521493,
50
  "loss": 1.3673,
51
  "step": 25
52
  },
53
  {
54
  "epoch": 0.17804154302670624,
55
- "grad_norm": 1.0031766891479492,
56
  "learning_rate": 0.00019636458959356316,
57
- "loss": 1.3098,
58
  "step": 30
59
  },
60
  {
61
  "epoch": 0.20771513353115728,
62
- "grad_norm": 0.7017138004302979,
63
  "learning_rate": 0.00019306926579854821,
64
  "loss": 1.2812,
65
  "step": 35
66
  },
67
  {
68
  "epoch": 0.23738872403560832,
69
- "grad_norm": 0.7703685164451599,
70
  "learning_rate": 0.00018876770456851877,
71
- "loss": 1.265,
72
  "step": 40
73
  },
74
  {
75
  "epoch": 0.26706231454005935,
76
- "grad_norm": 0.7199541926383972,
77
  "learning_rate": 0.00018350641311400812,
78
- "loss": 1.2555,
79
  "step": 45
80
  },
81
  {
82
  "epoch": 0.29673590504451036,
83
- "grad_norm": 0.7666878700256348,
84
  "learning_rate": 0.0001773422749654988,
85
- "loss": 1.2433,
86
  "step": 50
87
  },
88
  {
89
  "epoch": 0.3264094955489614,
90
- "grad_norm": 1.0030404329299927,
91
  "learning_rate": 0.00017034193496547902,
92
- "loss": 1.2443,
93
  "step": 55
94
  },
95
  {
96
  "epoch": 0.3560830860534125,
97
- "grad_norm": 0.6837704181671143,
98
  "learning_rate": 0.00016258107872407375,
99
- "loss": 1.2282,
100
  "step": 60
101
  },
102
  {
103
  "epoch": 0.3857566765578635,
104
- "grad_norm": 0.6520164608955383,
105
  "learning_rate": 0.00015414361432856475,
106
- "loss": 1.2225,
107
  "step": 65
108
  },
109
  {
110
  "epoch": 0.41543026706231456,
111
- "grad_norm": 0.6578584909439087,
112
  "learning_rate": 0.00014512076515391375,
113
- "loss": 1.2227,
114
  "step": 70
115
  },
116
  {
117
  "epoch": 0.44510385756676557,
118
- "grad_norm": 0.7506173849105835,
119
  "learning_rate": 0.00013561008358255468,
120
- "loss": 1.2281,
121
  "step": 75
122
  },
123
  {
124
  "epoch": 0.47477744807121663,
125
- "grad_norm": 0.6809776425361633,
126
  "learning_rate": 0.0001257143962968246,
127
- "loss": 1.2078,
128
  "step": 80
129
  },
130
  {
131
  "epoch": 0.5044510385756676,
132
- "grad_norm": 0.6472010016441345,
133
  "learning_rate": 0.00011554069254722051,
134
- "loss": 1.2185,
135
  "step": 85
136
  },
137
  {
138
  "epoch": 0.5341246290801187,
139
- "grad_norm": 0.6473718881607056,
140
  "learning_rate": 0.00010519896741619803,
141
  "loss": 1.2216,
142
  "step": 90
143
  },
144
  {
145
  "epoch": 0.5637982195845698,
146
- "grad_norm": 0.6487621068954468,
147
  "learning_rate": 9.480103258380198e-05,
148
  "loss": 1.2095,
149
  "step": 95
150
  },
151
  {
152
  "epoch": 0.5934718100890207,
153
- "grad_norm": 0.7169174551963806,
154
  "learning_rate": 8.445930745277953e-05,
155
- "loss": 1.2196,
156
  "step": 100
157
  },
158
  {
159
  "epoch": 0.6231454005934718,
160
- "grad_norm": 0.6917266249656677,
161
  "learning_rate": 7.428560370317542e-05,
162
  "loss": 1.2122,
163
  "step": 105
164
  },
165
  {
166
  "epoch": 0.6528189910979229,
167
- "grad_norm": 0.6533765196800232,
168
  "learning_rate": 6.43899164174453e-05,
169
  "loss": 1.1987,
170
  "step": 110
171
  },
172
  {
173
  "epoch": 0.6824925816023739,
174
- "grad_norm": 0.740526556968689,
175
  "learning_rate": 5.487923484608629e-05,
176
  "loss": 1.2059,
177
  "step": 115
178
  },
179
  {
180
  "epoch": 0.712166172106825,
181
- "grad_norm": 0.7236852645874023,
182
  "learning_rate": 4.585638567143529e-05,
183
- "loss": 1.2049,
184
  "step": 120
185
  },
186
  {
187
  "epoch": 0.7418397626112759,
188
- "grad_norm": 0.6368817090988159,
189
  "learning_rate": 3.741892127592625e-05,
190
  "loss": 1.2139,
191
  "step": 125
192
  },
193
  {
194
  "epoch": 0.771513353115727,
195
- "grad_norm": 0.6370018124580383,
196
  "learning_rate": 2.9658065034520978e-05,
197
- "loss": 1.2009,
198
  "step": 130
199
  },
200
  {
201
  "epoch": 0.8011869436201781,
202
- "grad_norm": 0.691111147403717,
203
  "learning_rate": 2.265772503450122e-05,
204
- "loss": 1.1977,
205
  "step": 135
206
  },
207
  {
208
  "epoch": 0.8308605341246291,
209
- "grad_norm": 0.6910406947135925,
210
  "learning_rate": 1.649358688599191e-05,
211
  "loss": 1.2049,
212
  "step": 140
213
  },
214
  {
215
  "epoch": 0.8605341246290801,
216
- "grad_norm": 0.669152557849884,
217
  "learning_rate": 1.1232295431481222e-05,
218
  "loss": 1.2018,
219
  "step": 145
220
  },
221
  {
222
  "epoch": 0.8902077151335311,
223
- "grad_norm": 0.6983711123466492,
224
  "learning_rate": 6.930734201451816e-06,
225
- "loss": 1.2049,
226
  "step": 150
227
  },
228
  {
229
  "epoch": 0.9198813056379822,
230
- "grad_norm": 0.7779198288917542,
231
  "learning_rate": 3.6354104064368566e-06,
232
  "loss": 1.1999,
233
  "step": 155
234
  },
235
  {
236
  "epoch": 0.9495548961424333,
237
- "grad_norm": 0.6764826774597168,
238
  "learning_rate": 1.3819521147851123e-06,
239
  "loss": 1.1995,
240
  "step": 160
241
  },
242
  {
243
  "epoch": 0.9792284866468842,
244
- "grad_norm": 0.6475749611854553,
245
  "learning_rate": 1.947230525005006e-07,
246
  "loss": 1.1865,
247
  "step": 165
248
  },
249
  {
250
  "epoch": 0.9970326409495549,
251
- "eval_loss": 1.8814080953598022,
252
- "eval_runtime": 0.6632,
253
- "eval_samples_per_second": 21.108,
254
- "eval_steps_per_second": 1.508,
255
  "step": 168
256
  },
257
  {
258
  "epoch": 0.9970326409495549,
259
  "step": 168,
260
  "total_flos": 8.227900459906499e+17,
261
- "train_loss": 1.2773614100047521,
262
- "train_runtime": 661.1188,
263
- "train_samples_per_second": 57.022,
264
- "train_steps_per_second": 0.254
265
  }
266
  ],
267
  "logging_steps": 5,
 
10
  "log_history": [
11
  {
12
  "epoch": 0.005934718100890208,
13
+ "grad_norm": 3.172736644744873,
14
  "learning_rate": 1.1764705882352942e-05,
15
  "loss": 1.7951,
16
  "step": 1
17
  },
18
  {
19
  "epoch": 0.02967359050445104,
20
+ "grad_norm": 3.245225429534912,
21
  "learning_rate": 5.882352941176471e-05,
22
  "loss": 1.8001,
23
  "step": 5
24
  },
25
  {
26
  "epoch": 0.05934718100890208,
27
+ "grad_norm": 2.2378523349761963,
28
  "learning_rate": 0.00011764705882352942,
29
  "loss": 1.7245,
30
  "step": 10
31
  },
32
  {
33
  "epoch": 0.08902077151335312,
34
+ "grad_norm": 2.2833542823791504,
35
  "learning_rate": 0.00017647058823529413,
36
+ "loss": 1.6177,
37
  "step": 15
38
  },
39
  {
40
  "epoch": 0.11869436201780416,
41
+ "grad_norm": 2.2818639278411865,
42
  "learning_rate": 0.00019980527694749952,
43
+ "loss": 1.4883,
44
  "step": 20
45
  },
46
  {
47
  "epoch": 0.14836795252225518,
48
+ "grad_norm": 2.0385029315948486,
49
  "learning_rate": 0.00019861804788521493,
50
  "loss": 1.3673,
51
  "step": 25
52
  },
53
  {
54
  "epoch": 0.17804154302670624,
55
+ "grad_norm": 1.0247116088867188,
56
  "learning_rate": 0.00019636458959356316,
57
+ "loss": 1.3097,
58
  "step": 30
59
  },
60
  {
61
  "epoch": 0.20771513353115728,
62
+ "grad_norm": 0.694072961807251,
63
  "learning_rate": 0.00019306926579854821,
64
  "loss": 1.2812,
65
  "step": 35
66
  },
67
  {
68
  "epoch": 0.23738872403560832,
69
+ "grad_norm": 0.7701340913772583,
70
  "learning_rate": 0.00018876770456851877,
71
+ "loss": 1.2651,
72
  "step": 40
73
  },
74
  {
75
  "epoch": 0.26706231454005935,
76
+ "grad_norm": 0.7059149742126465,
77
  "learning_rate": 0.00018350641311400812,
78
+ "loss": 1.2556,
79
  "step": 45
80
  },
81
  {
82
  "epoch": 0.29673590504451036,
83
+ "grad_norm": 0.8174459338188171,
84
  "learning_rate": 0.0001773422749654988,
85
+ "loss": 1.2434,
86
  "step": 50
87
  },
88
  {
89
  "epoch": 0.3264094955489614,
90
+ "grad_norm": 0.999190628528595,
91
  "learning_rate": 0.00017034193496547902,
92
+ "loss": 1.2444,
93
  "step": 55
94
  },
95
  {
96
  "epoch": 0.3560830860534125,
97
+ "grad_norm": 0.6954330801963806,
98
  "learning_rate": 0.00016258107872407375,
99
+ "loss": 1.2284,
100
  "step": 60
101
  },
102
  {
103
  "epoch": 0.3857566765578635,
104
+ "grad_norm": 0.64048171043396,
105
  "learning_rate": 0.00015414361432856475,
106
+ "loss": 1.2226,
107
  "step": 65
108
  },
109
  {
110
  "epoch": 0.41543026706231456,
111
+ "grad_norm": 0.6632175445556641,
112
  "learning_rate": 0.00014512076515391375,
113
+ "loss": 1.2228,
114
  "step": 70
115
  },
116
  {
117
  "epoch": 0.44510385756676557,
118
+ "grad_norm": 0.7728683948516846,
119
  "learning_rate": 0.00013561008358255468,
120
+ "loss": 1.2283,
121
  "step": 75
122
  },
123
  {
124
  "epoch": 0.47477744807121663,
125
+ "grad_norm": 0.6928794384002686,
126
  "learning_rate": 0.0001257143962968246,
127
+ "loss": 1.2079,
128
  "step": 80
129
  },
130
  {
131
  "epoch": 0.5044510385756676,
132
+ "grad_norm": 0.6514697074890137,
133
  "learning_rate": 0.00011554069254722051,
134
+ "loss": 1.2187,
135
  "step": 85
136
  },
137
  {
138
  "epoch": 0.5341246290801187,
139
+ "grad_norm": 0.6465914249420166,
140
  "learning_rate": 0.00010519896741619803,
141
  "loss": 1.2216,
142
  "step": 90
143
  },
144
  {
145
  "epoch": 0.5637982195845698,
146
+ "grad_norm": 0.7518061399459839,
147
  "learning_rate": 9.480103258380198e-05,
148
  "loss": 1.2095,
149
  "step": 95
150
  },
151
  {
152
  "epoch": 0.5934718100890207,
153
+ "grad_norm": 0.7165743112564087,
154
  "learning_rate": 8.445930745277953e-05,
155
+ "loss": 1.2197,
156
  "step": 100
157
  },
158
  {
159
  "epoch": 0.6231454005934718,
160
+ "grad_norm": 0.6865619421005249,
161
  "learning_rate": 7.428560370317542e-05,
162
  "loss": 1.2122,
163
  "step": 105
164
  },
165
  {
166
  "epoch": 0.6528189910979229,
167
+ "grad_norm": 0.6512064933776855,
168
  "learning_rate": 6.43899164174453e-05,
169
  "loss": 1.1987,
170
  "step": 110
171
  },
172
  {
173
  "epoch": 0.6824925816023739,
174
+ "grad_norm": 0.720739483833313,
175
  "learning_rate": 5.487923484608629e-05,
176
  "loss": 1.2059,
177
  "step": 115
178
  },
179
  {
180
  "epoch": 0.712166172106825,
181
+ "grad_norm": 0.7217025756835938,
182
  "learning_rate": 4.585638567143529e-05,
183
+ "loss": 1.205,
184
  "step": 120
185
  },
186
  {
187
  "epoch": 0.7418397626112759,
188
+ "grad_norm": 0.6354172229766846,
189
  "learning_rate": 3.741892127592625e-05,
190
  "loss": 1.2139,
191
  "step": 125
192
  },
193
  {
194
  "epoch": 0.771513353115727,
195
+ "grad_norm": 0.6358697414398193,
196
  "learning_rate": 2.9658065034520978e-05,
197
+ "loss": 1.2008,
198
  "step": 130
199
  },
200
  {
201
  "epoch": 0.8011869436201781,
202
+ "grad_norm": 0.6895249485969543,
203
  "learning_rate": 2.265772503450122e-05,
204
+ "loss": 1.1978,
205
  "step": 135
206
  },
207
  {
208
  "epoch": 0.8308605341246291,
209
+ "grad_norm": 0.6897445321083069,
210
  "learning_rate": 1.649358688599191e-05,
211
  "loss": 1.2049,
212
  "step": 140
213
  },
214
  {
215
  "epoch": 0.8605341246290801,
216
+ "grad_norm": 0.6667594909667969,
217
  "learning_rate": 1.1232295431481222e-05,
218
  "loss": 1.2018,
219
  "step": 145
220
  },
221
  {
222
  "epoch": 0.8902077151335311,
223
+ "grad_norm": 0.6951190829277039,
224
  "learning_rate": 6.930734201451816e-06,
225
+ "loss": 1.205,
226
  "step": 150
227
  },
228
  {
229
  "epoch": 0.9198813056379822,
230
+ "grad_norm": 0.7803878784179688,
231
  "learning_rate": 3.6354104064368566e-06,
232
  "loss": 1.1999,
233
  "step": 155
234
  },
235
  {
236
  "epoch": 0.9495548961424333,
237
+ "grad_norm": 0.6693819761276245,
238
  "learning_rate": 1.3819521147851123e-06,
239
  "loss": 1.1995,
240
  "step": 160
241
  },
242
  {
243
  "epoch": 0.9792284866468842,
244
+ "grad_norm": 0.6466339230537415,
245
  "learning_rate": 1.947230525005006e-07,
246
  "loss": 1.1865,
247
  "step": 165
248
  },
249
  {
250
  "epoch": 0.9970326409495549,
251
+ "eval_loss": 1.8819891214370728,
252
+ "eval_runtime": 0.6391,
253
+ "eval_samples_per_second": 21.907,
254
+ "eval_steps_per_second": 1.565,
255
  "step": 168
256
  },
257
  {
258
  "epoch": 0.9970326409495549,
259
  "step": 168,
260
  "total_flos": 8.227900459906499e+17,
261
+ "train_loss": 1.2774092711153484,
262
+ "train_runtime": 653.4451,
263
+ "train_samples_per_second": 57.691,
264
+ "train_steps_per_second": 0.257
265
  }
266
  ],
267
  "logging_steps": 5,