Remove training files
Browse files- logs/events.out.tfevents.1725961564.c86e03eac353.886.0 +0 -3
- optimizer.pt +0 -3
- rng_state_0.pth +0 -3
- rng_state_1.pth +0 -3
- rng_state_2.pth +0 -3
- rng_state_3.pth +0 -3
- rng_state_4.pth +0 -3
- rng_state_5.pth +0 -3
- rng_state_6.pth +0 -3
- rng_state_7.pth +0 -3
- scheduler.pt +0 -3
- trainer_state.json +0 -390
- training_args.bin +0 -3
logs/events.out.tfevents.1725961564.c86e03eac353.886.0
DELETED
@@ -1,3 +0,0 @@
|
|
1 |
-
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:a22f472e8079144fb7f3b9032cd8c6bc50bed7a52453db55b8ed240b7a81bf18
|
3 |
-
size 16821
|
|
|
|
|
|
|
|
optimizer.pt
DELETED
@@ -1,3 +0,0 @@
|
|
1 |
-
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:9eca8ba7daf4cfad01aa960ce5e5900cace805d2913405312eba6b556e273502
|
3 |
-
size 433692228
|
|
|
|
|
|
|
|
rng_state_0.pth
DELETED
@@ -1,3 +0,0 @@
|
|
1 |
-
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:5b6f5d3a1061a6066a95653e8d8dd8997e47d823dd851080507b8787203d9df3
|
3 |
-
size 15984
|
|
|
|
|
|
|
|
rng_state_1.pth
DELETED
@@ -1,3 +0,0 @@
|
|
1 |
-
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:971c7e6f3b0bc10d3e811e81bc5e5e13688c886b61a1da3c639c4387a599a08b
|
3 |
-
size 15984
|
|
|
|
|
|
|
|
rng_state_2.pth
DELETED
@@ -1,3 +0,0 @@
|
|
1 |
-
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:4ddddb34f552c6f88968ed8c602e2c9de7cd93fba6996d5e5b30d5f5263af88e
|
3 |
-
size 15984
|
|
|
|
|
|
|
|
rng_state_3.pth
DELETED
@@ -1,3 +0,0 @@
|
|
1 |
-
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:09eb55758e60c09207f3081d1b600bddd519c73e854eeda299eb8a1ba7087666
|
3 |
-
size 15984
|
|
|
|
|
|
|
|
rng_state_4.pth
DELETED
@@ -1,3 +0,0 @@
|
|
1 |
-
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:e24e95f48b88d38a5adc9ba2caa8364985de870a8dc4acdb53e30ec22f7cb4f5
|
3 |
-
size 15984
|
|
|
|
|
|
|
|
rng_state_5.pth
DELETED
@@ -1,3 +0,0 @@
|
|
1 |
-
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:4388c4677e069fc340a89e10d58edf267a382bea5706a2fa85e976573f2f0e2c
|
3 |
-
size 15984
|
|
|
|
|
|
|
|
rng_state_6.pth
DELETED
@@ -1,3 +0,0 @@
|
|
1 |
-
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:7071119655e00ebd01e0082a2d8904b051d419536d25011c279d5cf6c6ffff6b
|
3 |
-
size 15984
|
|
|
|
|
|
|
|
rng_state_7.pth
DELETED
@@ -1,3 +0,0 @@
|
|
1 |
-
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:1ceaa15d91a4aaa5689e525096b9c166a7c9a8dd3387ff5d9ae716b1aa9e037c
|
3 |
-
size 15984
|
|
|
|
|
|
|
|
scheduler.pt
DELETED
@@ -1,3 +0,0 @@
|
|
1 |
-
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:014652d37c59e7882147b45e7b55f83eaeecc90884b3492993164177d583c3f1
|
3 |
-
size 1064
|
|
|
|
|
|
|
|
trainer_state.json
DELETED
@@ -1,390 +0,0 @@
|
|
1 |
-
{
|
2 |
-
"best_metric": null,
|
3 |
-
"best_model_checkpoint": null,
|
4 |
-
"epoch": 0.9959579628132579,
|
5 |
-
"eval_steps": 500,
|
6 |
-
"global_step": 154,
|
7 |
-
"is_hyper_param_search": false,
|
8 |
-
"is_local_process_zero": true,
|
9 |
-
"is_world_process_zero": true,
|
10 |
-
"log_history": [
|
11 |
-
{
|
12 |
-
"epoch": 0.019401778496362168,
|
13 |
-
"grad_norm": 139.02731323242188,
|
14 |
-
"learning_rate": 1.9610389610389612e-05,
|
15 |
-
"loss": 1.8718,
|
16 |
-
"step": 3
|
17 |
-
},
|
18 |
-
{
|
19 |
-
"epoch": 0.038803556992724336,
|
20 |
-
"grad_norm": 31.607357025146484,
|
21 |
-
"learning_rate": 1.9220779220779222e-05,
|
22 |
-
"loss": 1.6642,
|
23 |
-
"step": 6
|
24 |
-
},
|
25 |
-
{
|
26 |
-
"epoch": 0.0582053354890865,
|
27 |
-
"grad_norm": 118.545166015625,
|
28 |
-
"learning_rate": 1.8831168831168833e-05,
|
29 |
-
"loss": 1.5334,
|
30 |
-
"step": 9
|
31 |
-
},
|
32 |
-
{
|
33 |
-
"epoch": 0.07760711398544867,
|
34 |
-
"grad_norm": 180.7499542236328,
|
35 |
-
"learning_rate": 1.8441558441558443e-05,
|
36 |
-
"loss": 1.396,
|
37 |
-
"step": 12
|
38 |
-
},
|
39 |
-
{
|
40 |
-
"epoch": 0.09700889248181083,
|
41 |
-
"grad_norm": 199.1714324951172,
|
42 |
-
"learning_rate": 1.8051948051948053e-05,
|
43 |
-
"loss": 1.3652,
|
44 |
-
"step": 15
|
45 |
-
},
|
46 |
-
{
|
47 |
-
"epoch": 0.116410670978173,
|
48 |
-
"grad_norm": 136.60414123535156,
|
49 |
-
"learning_rate": 1.7662337662337664e-05,
|
50 |
-
"loss": 1.3432,
|
51 |
-
"step": 18
|
52 |
-
},
|
53 |
-
{
|
54 |
-
"epoch": 0.13581244947453516,
|
55 |
-
"grad_norm": 82.5650634765625,
|
56 |
-
"learning_rate": 1.7272727272727274e-05,
|
57 |
-
"loss": 1.2614,
|
58 |
-
"step": 21
|
59 |
-
},
|
60 |
-
{
|
61 |
-
"epoch": 0.15521422797089734,
|
62 |
-
"grad_norm": 18.96686553955078,
|
63 |
-
"learning_rate": 1.6883116883116884e-05,
|
64 |
-
"loss": 1.2303,
|
65 |
-
"step": 24
|
66 |
-
},
|
67 |
-
{
|
68 |
-
"epoch": 0.1746160064672595,
|
69 |
-
"grad_norm": 7.933801174163818,
|
70 |
-
"learning_rate": 1.6493506493506495e-05,
|
71 |
-
"loss": 1.1984,
|
72 |
-
"step": 27
|
73 |
-
},
|
74 |
-
{
|
75 |
-
"epoch": 0.19401778496362165,
|
76 |
-
"grad_norm": 2.686699390411377,
|
77 |
-
"learning_rate": 1.6103896103896105e-05,
|
78 |
-
"loss": 1.1016,
|
79 |
-
"step": 30
|
80 |
-
},
|
81 |
-
{
|
82 |
-
"epoch": 0.21341956345998384,
|
83 |
-
"grad_norm": 1.455581545829773,
|
84 |
-
"learning_rate": 1.5714285714285715e-05,
|
85 |
-
"loss": 1.0671,
|
86 |
-
"step": 33
|
87 |
-
},
|
88 |
-
{
|
89 |
-
"epoch": 0.232821341956346,
|
90 |
-
"grad_norm": 0.5924062132835388,
|
91 |
-
"learning_rate": 1.5324675324675326e-05,
|
92 |
-
"loss": 1.012,
|
93 |
-
"step": 36
|
94 |
-
},
|
95 |
-
{
|
96 |
-
"epoch": 0.25222312045270817,
|
97 |
-
"grad_norm": 0.3087107837200165,
|
98 |
-
"learning_rate": 1.4935064935064936e-05,
|
99 |
-
"loss": 0.9758,
|
100 |
-
"step": 39
|
101 |
-
},
|
102 |
-
{
|
103 |
-
"epoch": 0.2716248989490703,
|
104 |
-
"grad_norm": 0.2992459535598755,
|
105 |
-
"learning_rate": 1.4545454545454546e-05,
|
106 |
-
"loss": 0.9262,
|
107 |
-
"step": 42
|
108 |
-
},
|
109 |
-
{
|
110 |
-
"epoch": 0.2910266774454325,
|
111 |
-
"grad_norm": 0.2895904779434204,
|
112 |
-
"learning_rate": 1.4155844155844157e-05,
|
113 |
-
"loss": 0.8271,
|
114 |
-
"step": 45
|
115 |
-
},
|
116 |
-
{
|
117 |
-
"epoch": 0.3104284559417947,
|
118 |
-
"grad_norm": 0.2948096692562103,
|
119 |
-
"learning_rate": 1.3766233766233767e-05,
|
120 |
-
"loss": 0.7895,
|
121 |
-
"step": 48
|
122 |
-
},
|
123 |
-
{
|
124 |
-
"epoch": 0.32983023443815684,
|
125 |
-
"grad_norm": 0.31464704871177673,
|
126 |
-
"learning_rate": 1.3376623376623377e-05,
|
127 |
-
"loss": 0.7299,
|
128 |
-
"step": 51
|
129 |
-
},
|
130 |
-
{
|
131 |
-
"epoch": 0.349232012934519,
|
132 |
-
"grad_norm": 0.3038002550601959,
|
133 |
-
"learning_rate": 1.2987012987012988e-05,
|
134 |
-
"loss": 0.6857,
|
135 |
-
"step": 54
|
136 |
-
},
|
137 |
-
{
|
138 |
-
"epoch": 0.36863379143088115,
|
139 |
-
"grad_norm": 0.33729803562164307,
|
140 |
-
"learning_rate": 1.25974025974026e-05,
|
141 |
-
"loss": 0.5946,
|
142 |
-
"step": 57
|
143 |
-
},
|
144 |
-
{
|
145 |
-
"epoch": 0.3880355699272433,
|
146 |
-
"grad_norm": 0.39213827252388,
|
147 |
-
"learning_rate": 1.2207792207792208e-05,
|
148 |
-
"loss": 0.5636,
|
149 |
-
"step": 60
|
150 |
-
},
|
151 |
-
{
|
152 |
-
"epoch": 0.4074373484236055,
|
153 |
-
"grad_norm": 0.3482286334037781,
|
154 |
-
"learning_rate": 1.181818181818182e-05,
|
155 |
-
"loss": 0.5094,
|
156 |
-
"step": 63
|
157 |
-
},
|
158 |
-
{
|
159 |
-
"epoch": 0.42683912691996767,
|
160 |
-
"grad_norm": 0.3112964630126953,
|
161 |
-
"learning_rate": 1.1428571428571429e-05,
|
162 |
-
"loss": 0.4541,
|
163 |
-
"step": 66
|
164 |
-
},
|
165 |
-
{
|
166 |
-
"epoch": 0.4462409054163298,
|
167 |
-
"grad_norm": 0.26819908618927,
|
168 |
-
"learning_rate": 1.1038961038961041e-05,
|
169 |
-
"loss": 0.4181,
|
170 |
-
"step": 69
|
171 |
-
},
|
172 |
-
{
|
173 |
-
"epoch": 0.465642683912692,
|
174 |
-
"grad_norm": 0.28413137793540955,
|
175 |
-
"learning_rate": 1.064935064935065e-05,
|
176 |
-
"loss": 0.4095,
|
177 |
-
"step": 72
|
178 |
-
},
|
179 |
-
{
|
180 |
-
"epoch": 0.4850444624090542,
|
181 |
-
"grad_norm": 0.3022381365299225,
|
182 |
-
"learning_rate": 1.025974025974026e-05,
|
183 |
-
"loss": 0.3623,
|
184 |
-
"step": 75
|
185 |
-
},
|
186 |
-
{
|
187 |
-
"epoch": 0.5044462409054163,
|
188 |
-
"grad_norm": 0.29346349835395813,
|
189 |
-
"learning_rate": 9.87012987012987e-06,
|
190 |
-
"loss": 0.3334,
|
191 |
-
"step": 78
|
192 |
-
},
|
193 |
-
{
|
194 |
-
"epoch": 0.5238480194017785,
|
195 |
-
"grad_norm": 0.2659854292869568,
|
196 |
-
"learning_rate": 9.48051948051948e-06,
|
197 |
-
"loss": 0.3115,
|
198 |
-
"step": 81
|
199 |
-
},
|
200 |
-
{
|
201 |
-
"epoch": 0.5432497978981407,
|
202 |
-
"grad_norm": 0.23122940957546234,
|
203 |
-
"learning_rate": 9.090909090909091e-06,
|
204 |
-
"loss": 0.2817,
|
205 |
-
"step": 84
|
206 |
-
},
|
207 |
-
{
|
208 |
-
"epoch": 0.5626515763945028,
|
209 |
-
"grad_norm": 0.2369256317615509,
|
210 |
-
"learning_rate": 8.701298701298701e-06,
|
211 |
-
"loss": 0.2809,
|
212 |
-
"step": 87
|
213 |
-
},
|
214 |
-
{
|
215 |
-
"epoch": 0.582053354890865,
|
216 |
-
"grad_norm": 0.2082873433828354,
|
217 |
-
"learning_rate": 8.311688311688313e-06,
|
218 |
-
"loss": 0.2455,
|
219 |
-
"step": 90
|
220 |
-
},
|
221 |
-
{
|
222 |
-
"epoch": 0.6014551333872271,
|
223 |
-
"grad_norm": 0.21645894646644592,
|
224 |
-
"learning_rate": 7.922077922077924e-06,
|
225 |
-
"loss": 0.2503,
|
226 |
-
"step": 93
|
227 |
-
},
|
228 |
-
{
|
229 |
-
"epoch": 0.6208569118835894,
|
230 |
-
"grad_norm": 0.19337739050388336,
|
231 |
-
"learning_rate": 7.532467532467533e-06,
|
232 |
-
"loss": 0.2286,
|
233 |
-
"step": 96
|
234 |
-
},
|
235 |
-
{
|
236 |
-
"epoch": 0.6402586903799515,
|
237 |
-
"grad_norm": 0.1808944046497345,
|
238 |
-
"learning_rate": 7.1428571428571436e-06,
|
239 |
-
"loss": 0.2401,
|
240 |
-
"step": 99
|
241 |
-
},
|
242 |
-
{
|
243 |
-
"epoch": 0.6596604688763137,
|
244 |
-
"grad_norm": 0.1630856841802597,
|
245 |
-
"learning_rate": 6.753246753246754e-06,
|
246 |
-
"loss": 0.2251,
|
247 |
-
"step": 102
|
248 |
-
},
|
249 |
-
{
|
250 |
-
"epoch": 0.6790622473726758,
|
251 |
-
"grad_norm": 0.16326990723609924,
|
252 |
-
"learning_rate": 6.363636363636364e-06,
|
253 |
-
"loss": 0.2291,
|
254 |
-
"step": 105
|
255 |
-
},
|
256 |
-
{
|
257 |
-
"epoch": 0.698464025869038,
|
258 |
-
"grad_norm": 0.16061735153198242,
|
259 |
-
"learning_rate": 5.9740259740259746e-06,
|
260 |
-
"loss": 0.2331,
|
261 |
-
"step": 108
|
262 |
-
},
|
263 |
-
{
|
264 |
-
"epoch": 0.7178658043654002,
|
265 |
-
"grad_norm": 0.17352429032325745,
|
266 |
-
"learning_rate": 5.584415584415585e-06,
|
267 |
-
"loss": 0.2149,
|
268 |
-
"step": 111
|
269 |
-
},
|
270 |
-
{
|
271 |
-
"epoch": 0.7372675828617623,
|
272 |
-
"grad_norm": 0.17043530941009521,
|
273 |
-
"learning_rate": 5.194805194805194e-06,
|
274 |
-
"loss": 0.2187,
|
275 |
-
"step": 114
|
276 |
-
},
|
277 |
-
{
|
278 |
-
"epoch": 0.7566693613581245,
|
279 |
-
"grad_norm": 0.16479559242725372,
|
280 |
-
"learning_rate": 4.805194805194806e-06,
|
281 |
-
"loss": 0.2218,
|
282 |
-
"step": 117
|
283 |
-
},
|
284 |
-
{
|
285 |
-
"epoch": 0.7760711398544866,
|
286 |
-
"grad_norm": 0.17882439494132996,
|
287 |
-
"learning_rate": 4.415584415584416e-06,
|
288 |
-
"loss": 0.205,
|
289 |
-
"step": 120
|
290 |
-
},
|
291 |
-
{
|
292 |
-
"epoch": 0.7954729183508489,
|
293 |
-
"grad_norm": 0.1911778748035431,
|
294 |
-
"learning_rate": 4.025974025974026e-06,
|
295 |
-
"loss": 0.2172,
|
296 |
-
"step": 123
|
297 |
-
},
|
298 |
-
{
|
299 |
-
"epoch": 0.814874696847211,
|
300 |
-
"grad_norm": 0.17751498520374298,
|
301 |
-
"learning_rate": 3.6363636363636366e-06,
|
302 |
-
"loss": 0.2096,
|
303 |
-
"step": 126
|
304 |
-
},
|
305 |
-
{
|
306 |
-
"epoch": 0.8342764753435732,
|
307 |
-
"grad_norm": 0.1702156662940979,
|
308 |
-
"learning_rate": 3.246753246753247e-06,
|
309 |
-
"loss": 0.1911,
|
310 |
-
"step": 129
|
311 |
-
},
|
312 |
-
{
|
313 |
-
"epoch": 0.8536782538399353,
|
314 |
-
"grad_norm": 0.1764981597661972,
|
315 |
-
"learning_rate": 2.8571428571428573e-06,
|
316 |
-
"loss": 0.2103,
|
317 |
-
"step": 132
|
318 |
-
},
|
319 |
-
{
|
320 |
-
"epoch": 0.8730800323362975,
|
321 |
-
"grad_norm": 0.1592799872159958,
|
322 |
-
"learning_rate": 2.4675324675324676e-06,
|
323 |
-
"loss": 0.2053,
|
324 |
-
"step": 135
|
325 |
-
},
|
326 |
-
{
|
327 |
-
"epoch": 0.8924818108326596,
|
328 |
-
"grad_norm": 0.21512138843536377,
|
329 |
-
"learning_rate": 2.0779220779220784e-06,
|
330 |
-
"loss": 0.2197,
|
331 |
-
"step": 138
|
332 |
-
},
|
333 |
-
{
|
334 |
-
"epoch": 0.9118835893290218,
|
335 |
-
"grad_norm": 0.17707495391368866,
|
336 |
-
"learning_rate": 1.6883116883116885e-06,
|
337 |
-
"loss": 0.2051,
|
338 |
-
"step": 141
|
339 |
-
},
|
340 |
-
{
|
341 |
-
"epoch": 0.931285367825384,
|
342 |
-
"grad_norm": 0.1585138887166977,
|
343 |
-
"learning_rate": 1.2987012987012986e-06,
|
344 |
-
"loss": 0.1984,
|
345 |
-
"step": 144
|
346 |
-
},
|
347 |
-
{
|
348 |
-
"epoch": 0.9506871463217461,
|
349 |
-
"grad_norm": 0.15231232345104218,
|
350 |
-
"learning_rate": 9.090909090909091e-07,
|
351 |
-
"loss": 0.1774,
|
352 |
-
"step": 147
|
353 |
-
},
|
354 |
-
{
|
355 |
-
"epoch": 0.9700889248181084,
|
356 |
-
"grad_norm": 0.15338800847530365,
|
357 |
-
"learning_rate": 5.194805194805196e-07,
|
358 |
-
"loss": 0.2046,
|
359 |
-
"step": 150
|
360 |
-
},
|
361 |
-
{
|
362 |
-
"epoch": 0.9894907033144705,
|
363 |
-
"grad_norm": 0.16579587757587433,
|
364 |
-
"learning_rate": 1.298701298701299e-07,
|
365 |
-
"loss": 0.1871,
|
366 |
-
"step": 153
|
367 |
-
}
|
368 |
-
],
|
369 |
-
"logging_steps": 3,
|
370 |
-
"max_steps": 154,
|
371 |
-
"num_input_tokens_seen": 0,
|
372 |
-
"num_train_epochs": 1,
|
373 |
-
"save_steps": 10,
|
374 |
-
"stateful_callbacks": {
|
375 |
-
"TrainerControl": {
|
376 |
-
"args": {
|
377 |
-
"should_epoch_stop": false,
|
378 |
-
"should_evaluate": false,
|
379 |
-
"should_log": false,
|
380 |
-
"should_save": true,
|
381 |
-
"should_training_stop": true
|
382 |
-
},
|
383 |
-
"attributes": {}
|
384 |
-
}
|
385 |
-
},
|
386 |
-
"total_flos": 2.5705942959542764e+18,
|
387 |
-
"train_batch_size": 4,
|
388 |
-
"trial_name": null,
|
389 |
-
"trial_params": null
|
390 |
-
}
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
training_args.bin
DELETED
@@ -1,3 +0,0 @@
|
|
1 |
-
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:976ee61ae2df453812c83d50614dca91367b247a7a87691c610c1d3ced20ce84
|
3 |
-
size 5368
|
|
|
|
|
|
|
|