SF-Foundation commited on
Commit
f2c8538
·
verified ·
1 Parent(s): 07dce51

Upload folder using huggingface_hub

Browse files
.gitattributes CHANGED
@@ -33,3 +33,4 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
 
 
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
36
+ tokenizer.json filter=lfs diff=lfs merge=lfs -text
added_tokens.json ADDED
@@ -0,0 +1,1019 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "</think>": 151668,
3
+ "</tool_call>": 151658,
4
+ "</tool_response>": 151666,
5
+ "<think>": 151667,
6
+ "<tool_call>": 151657,
7
+ "<tool_response>": 151665,
8
+ "<|box_end|>": 151649,
9
+ "<|box_start|>": 151648,
10
+ "<|endoftext|>": 151643,
11
+ "<|file_sep|>": 151664,
12
+ "<|fim_middle|>": 151660,
13
+ "<|fim_pad|>": 151662,
14
+ "<|fim_prefix|>": 151659,
15
+ "<|fim_suffix|>": 151661,
16
+ "<|im_end|>": 151645,
17
+ "<|im_start|>": 151644,
18
+ "<|image_pad|>": 151655,
19
+ "<|object_ref_end|>": 151647,
20
+ "<|object_ref_start|>": 151646,
21
+ "<|quad_end|>": 151651,
22
+ "<|quad_start|>": 151650,
23
+ "<|repo_name|>": 151663,
24
+ "<|video_pad|>": 151656,
25
+ "<|vision_end|>": 151653,
26
+ "<|vision_pad|>": 151654,
27
+ "<|vision_start|>": 151652,
28
+ "[control_1000]": 152659,
29
+ "[control_100]": 151759,
30
+ "[control_101]": 151760,
31
+ "[control_102]": 151761,
32
+ "[control_103]": 151762,
33
+ "[control_104]": 151763,
34
+ "[control_105]": 151764,
35
+ "[control_106]": 151765,
36
+ "[control_107]": 151766,
37
+ "[control_108]": 151767,
38
+ "[control_109]": 151768,
39
+ "[control_10]": 151669,
40
+ "[control_110]": 151769,
41
+ "[control_111]": 151770,
42
+ "[control_112]": 151771,
43
+ "[control_113]": 151772,
44
+ "[control_114]": 151773,
45
+ "[control_115]": 151774,
46
+ "[control_116]": 151775,
47
+ "[control_117]": 151776,
48
+ "[control_118]": 151777,
49
+ "[control_119]": 151778,
50
+ "[control_11]": 151670,
51
+ "[control_120]": 151779,
52
+ "[control_121]": 151780,
53
+ "[control_122]": 151781,
54
+ "[control_123]": 151782,
55
+ "[control_124]": 151783,
56
+ "[control_125]": 151784,
57
+ "[control_126]": 151785,
58
+ "[control_127]": 151786,
59
+ "[control_128]": 151787,
60
+ "[control_129]": 151788,
61
+ "[control_12]": 151671,
62
+ "[control_130]": 151789,
63
+ "[control_131]": 151790,
64
+ "[control_132]": 151791,
65
+ "[control_133]": 151792,
66
+ "[control_134]": 151793,
67
+ "[control_135]": 151794,
68
+ "[control_136]": 151795,
69
+ "[control_137]": 151796,
70
+ "[control_138]": 151797,
71
+ "[control_139]": 151798,
72
+ "[control_13]": 151672,
73
+ "[control_140]": 151799,
74
+ "[control_141]": 151800,
75
+ "[control_142]": 151801,
76
+ "[control_143]": 151802,
77
+ "[control_144]": 151803,
78
+ "[control_145]": 151804,
79
+ "[control_146]": 151805,
80
+ "[control_147]": 151806,
81
+ "[control_148]": 151807,
82
+ "[control_149]": 151808,
83
+ "[control_14]": 151673,
84
+ "[control_150]": 151809,
85
+ "[control_151]": 151810,
86
+ "[control_152]": 151811,
87
+ "[control_153]": 151812,
88
+ "[control_154]": 151813,
89
+ "[control_155]": 151814,
90
+ "[control_156]": 151815,
91
+ "[control_157]": 151816,
92
+ "[control_158]": 151817,
93
+ "[control_159]": 151818,
94
+ "[control_15]": 151674,
95
+ "[control_160]": 151819,
96
+ "[control_161]": 151820,
97
+ "[control_162]": 151821,
98
+ "[control_163]": 151822,
99
+ "[control_164]": 151823,
100
+ "[control_165]": 151824,
101
+ "[control_166]": 151825,
102
+ "[control_167]": 151826,
103
+ "[control_168]": 151827,
104
+ "[control_169]": 151828,
105
+ "[control_16]": 151675,
106
+ "[control_170]": 151829,
107
+ "[control_171]": 151830,
108
+ "[control_172]": 151831,
109
+ "[control_173]": 151832,
110
+ "[control_174]": 151833,
111
+ "[control_175]": 151834,
112
+ "[control_176]": 151835,
113
+ "[control_177]": 151836,
114
+ "[control_178]": 151837,
115
+ "[control_179]": 151838,
116
+ "[control_17]": 151676,
117
+ "[control_180]": 151839,
118
+ "[control_181]": 151840,
119
+ "[control_182]": 151841,
120
+ "[control_183]": 151842,
121
+ "[control_184]": 151843,
122
+ "[control_185]": 151844,
123
+ "[control_186]": 151845,
124
+ "[control_187]": 151846,
125
+ "[control_188]": 151847,
126
+ "[control_189]": 151848,
127
+ "[control_18]": 151677,
128
+ "[control_190]": 151849,
129
+ "[control_191]": 151850,
130
+ "[control_192]": 151851,
131
+ "[control_193]": 151852,
132
+ "[control_194]": 151853,
133
+ "[control_195]": 151854,
134
+ "[control_196]": 151855,
135
+ "[control_197]": 151856,
136
+ "[control_198]": 151857,
137
+ "[control_199]": 151858,
138
+ "[control_19]": 151678,
139
+ "[control_200]": 151859,
140
+ "[control_201]": 151860,
141
+ "[control_202]": 151861,
142
+ "[control_203]": 151862,
143
+ "[control_204]": 151863,
144
+ "[control_205]": 151864,
145
+ "[control_206]": 151865,
146
+ "[control_207]": 151866,
147
+ "[control_208]": 151867,
148
+ "[control_209]": 151868,
149
+ "[control_20]": 151679,
150
+ "[control_210]": 151869,
151
+ "[control_211]": 151870,
152
+ "[control_212]": 151871,
153
+ "[control_213]": 151872,
154
+ "[control_214]": 151873,
155
+ "[control_215]": 151874,
156
+ "[control_216]": 151875,
157
+ "[control_217]": 151876,
158
+ "[control_218]": 151877,
159
+ "[control_219]": 151878,
160
+ "[control_21]": 151680,
161
+ "[control_220]": 151879,
162
+ "[control_221]": 151880,
163
+ "[control_222]": 151881,
164
+ "[control_223]": 151882,
165
+ "[control_224]": 151883,
166
+ "[control_225]": 151884,
167
+ "[control_226]": 151885,
168
+ "[control_227]": 151886,
169
+ "[control_228]": 151887,
170
+ "[control_229]": 151888,
171
+ "[control_22]": 151681,
172
+ "[control_230]": 151889,
173
+ "[control_231]": 151890,
174
+ "[control_232]": 151891,
175
+ "[control_233]": 151892,
176
+ "[control_234]": 151893,
177
+ "[control_235]": 151894,
178
+ "[control_236]": 151895,
179
+ "[control_237]": 151896,
180
+ "[control_238]": 151897,
181
+ "[control_239]": 151898,
182
+ "[control_23]": 151682,
183
+ "[control_240]": 151899,
184
+ "[control_241]": 151900,
185
+ "[control_242]": 151901,
186
+ "[control_243]": 151902,
187
+ "[control_244]": 151903,
188
+ "[control_245]": 151904,
189
+ "[control_246]": 151905,
190
+ "[control_247]": 151906,
191
+ "[control_248]": 151907,
192
+ "[control_249]": 151908,
193
+ "[control_24]": 151683,
194
+ "[control_250]": 151909,
195
+ "[control_251]": 151910,
196
+ "[control_252]": 151911,
197
+ "[control_253]": 151912,
198
+ "[control_254]": 151913,
199
+ "[control_255]": 151914,
200
+ "[control_256]": 151915,
201
+ "[control_257]": 151916,
202
+ "[control_258]": 151917,
203
+ "[control_259]": 151918,
204
+ "[control_25]": 151684,
205
+ "[control_260]": 151919,
206
+ "[control_261]": 151920,
207
+ "[control_262]": 151921,
208
+ "[control_263]": 151922,
209
+ "[control_264]": 151923,
210
+ "[control_265]": 151924,
211
+ "[control_266]": 151925,
212
+ "[control_267]": 151926,
213
+ "[control_268]": 151927,
214
+ "[control_269]": 151928,
215
+ "[control_26]": 151685,
216
+ "[control_270]": 151929,
217
+ "[control_271]": 151930,
218
+ "[control_272]": 151931,
219
+ "[control_273]": 151932,
220
+ "[control_274]": 151933,
221
+ "[control_275]": 151934,
222
+ "[control_276]": 151935,
223
+ "[control_277]": 151936,
224
+ "[control_278]": 151937,
225
+ "[control_279]": 151938,
226
+ "[control_27]": 151686,
227
+ "[control_280]": 151939,
228
+ "[control_281]": 151940,
229
+ "[control_282]": 151941,
230
+ "[control_283]": 151942,
231
+ "[control_284]": 151943,
232
+ "[control_285]": 151944,
233
+ "[control_286]": 151945,
234
+ "[control_287]": 151946,
235
+ "[control_288]": 151947,
236
+ "[control_289]": 151948,
237
+ "[control_28]": 151687,
238
+ "[control_290]": 151949,
239
+ "[control_291]": 151950,
240
+ "[control_292]": 151951,
241
+ "[control_293]": 151952,
242
+ "[control_294]": 151953,
243
+ "[control_295]": 151954,
244
+ "[control_296]": 151955,
245
+ "[control_297]": 151956,
246
+ "[control_298]": 151957,
247
+ "[control_299]": 151958,
248
+ "[control_29]": 151688,
249
+ "[control_300]": 151959,
250
+ "[control_301]": 151960,
251
+ "[control_302]": 151961,
252
+ "[control_303]": 151962,
253
+ "[control_304]": 151963,
254
+ "[control_305]": 151964,
255
+ "[control_306]": 151965,
256
+ "[control_307]": 151966,
257
+ "[control_308]": 151967,
258
+ "[control_309]": 151968,
259
+ "[control_30]": 151689,
260
+ "[control_310]": 151969,
261
+ "[control_311]": 151970,
262
+ "[control_312]": 151971,
263
+ "[control_313]": 151972,
264
+ "[control_314]": 151973,
265
+ "[control_315]": 151974,
266
+ "[control_316]": 151975,
267
+ "[control_317]": 151976,
268
+ "[control_318]": 151977,
269
+ "[control_319]": 151978,
270
+ "[control_31]": 151690,
271
+ "[control_320]": 151979,
272
+ "[control_321]": 151980,
273
+ "[control_322]": 151981,
274
+ "[control_323]": 151982,
275
+ "[control_324]": 151983,
276
+ "[control_325]": 151984,
277
+ "[control_326]": 151985,
278
+ "[control_327]": 151986,
279
+ "[control_328]": 151987,
280
+ "[control_329]": 151988,
281
+ "[control_32]": 151691,
282
+ "[control_330]": 151989,
283
+ "[control_331]": 151990,
284
+ "[control_332]": 151991,
285
+ "[control_333]": 151992,
286
+ "[control_334]": 151993,
287
+ "[control_335]": 151994,
288
+ "[control_336]": 151995,
289
+ "[control_337]": 151996,
290
+ "[control_338]": 151997,
291
+ "[control_339]": 151998,
292
+ "[control_33]": 151692,
293
+ "[control_340]": 151999,
294
+ "[control_341]": 152000,
295
+ "[control_342]": 152001,
296
+ "[control_343]": 152002,
297
+ "[control_344]": 152003,
298
+ "[control_345]": 152004,
299
+ "[control_346]": 152005,
300
+ "[control_347]": 152006,
301
+ "[control_348]": 152007,
302
+ "[control_349]": 152008,
303
+ "[control_34]": 151693,
304
+ "[control_350]": 152009,
305
+ "[control_351]": 152010,
306
+ "[control_352]": 152011,
307
+ "[control_353]": 152012,
308
+ "[control_354]": 152013,
309
+ "[control_355]": 152014,
310
+ "[control_356]": 152015,
311
+ "[control_357]": 152016,
312
+ "[control_358]": 152017,
313
+ "[control_359]": 152018,
314
+ "[control_35]": 151694,
315
+ "[control_360]": 152019,
316
+ "[control_361]": 152020,
317
+ "[control_362]": 152021,
318
+ "[control_363]": 152022,
319
+ "[control_364]": 152023,
320
+ "[control_365]": 152024,
321
+ "[control_366]": 152025,
322
+ "[control_367]": 152026,
323
+ "[control_368]": 152027,
324
+ "[control_369]": 152028,
325
+ "[control_36]": 151695,
326
+ "[control_370]": 152029,
327
+ "[control_371]": 152030,
328
+ "[control_372]": 152031,
329
+ "[control_373]": 152032,
330
+ "[control_374]": 152033,
331
+ "[control_375]": 152034,
332
+ "[control_376]": 152035,
333
+ "[control_377]": 152036,
334
+ "[control_378]": 152037,
335
+ "[control_379]": 152038,
336
+ "[control_37]": 151696,
337
+ "[control_380]": 152039,
338
+ "[control_381]": 152040,
339
+ "[control_382]": 152041,
340
+ "[control_383]": 152042,
341
+ "[control_384]": 152043,
342
+ "[control_385]": 152044,
343
+ "[control_386]": 152045,
344
+ "[control_387]": 152046,
345
+ "[control_388]": 152047,
346
+ "[control_389]": 152048,
347
+ "[control_38]": 151697,
348
+ "[control_390]": 152049,
349
+ "[control_391]": 152050,
350
+ "[control_392]": 152051,
351
+ "[control_393]": 152052,
352
+ "[control_394]": 152053,
353
+ "[control_395]": 152054,
354
+ "[control_396]": 152055,
355
+ "[control_397]": 152056,
356
+ "[control_398]": 152057,
357
+ "[control_399]": 152058,
358
+ "[control_39]": 151698,
359
+ "[control_400]": 152059,
360
+ "[control_401]": 152060,
361
+ "[control_402]": 152061,
362
+ "[control_403]": 152062,
363
+ "[control_404]": 152063,
364
+ "[control_405]": 152064,
365
+ "[control_406]": 152065,
366
+ "[control_407]": 152066,
367
+ "[control_408]": 152067,
368
+ "[control_409]": 152068,
369
+ "[control_40]": 151699,
370
+ "[control_410]": 152069,
371
+ "[control_411]": 152070,
372
+ "[control_412]": 152071,
373
+ "[control_413]": 152072,
374
+ "[control_414]": 152073,
375
+ "[control_415]": 152074,
376
+ "[control_416]": 152075,
377
+ "[control_417]": 152076,
378
+ "[control_418]": 152077,
379
+ "[control_419]": 152078,
380
+ "[control_41]": 151700,
381
+ "[control_420]": 152079,
382
+ "[control_421]": 152080,
383
+ "[control_422]": 152081,
384
+ "[control_423]": 152082,
385
+ "[control_424]": 152083,
386
+ "[control_425]": 152084,
387
+ "[control_426]": 152085,
388
+ "[control_427]": 152086,
389
+ "[control_428]": 152087,
390
+ "[control_429]": 152088,
391
+ "[control_42]": 151701,
392
+ "[control_430]": 152089,
393
+ "[control_431]": 152090,
394
+ "[control_432]": 152091,
395
+ "[control_433]": 152092,
396
+ "[control_434]": 152093,
397
+ "[control_435]": 152094,
398
+ "[control_436]": 152095,
399
+ "[control_437]": 152096,
400
+ "[control_438]": 152097,
401
+ "[control_439]": 152098,
402
+ "[control_43]": 151702,
403
+ "[control_440]": 152099,
404
+ "[control_441]": 152100,
405
+ "[control_442]": 152101,
406
+ "[control_443]": 152102,
407
+ "[control_444]": 152103,
408
+ "[control_445]": 152104,
409
+ "[control_446]": 152105,
410
+ "[control_447]": 152106,
411
+ "[control_448]": 152107,
412
+ "[control_449]": 152108,
413
+ "[control_44]": 151703,
414
+ "[control_450]": 152109,
415
+ "[control_451]": 152110,
416
+ "[control_452]": 152111,
417
+ "[control_453]": 152112,
418
+ "[control_454]": 152113,
419
+ "[control_455]": 152114,
420
+ "[control_456]": 152115,
421
+ "[control_457]": 152116,
422
+ "[control_458]": 152117,
423
+ "[control_459]": 152118,
424
+ "[control_45]": 151704,
425
+ "[control_460]": 152119,
426
+ "[control_461]": 152120,
427
+ "[control_462]": 152121,
428
+ "[control_463]": 152122,
429
+ "[control_464]": 152123,
430
+ "[control_465]": 152124,
431
+ "[control_466]": 152125,
432
+ "[control_467]": 152126,
433
+ "[control_468]": 152127,
434
+ "[control_469]": 152128,
435
+ "[control_46]": 151705,
436
+ "[control_470]": 152129,
437
+ "[control_471]": 152130,
438
+ "[control_472]": 152131,
439
+ "[control_473]": 152132,
440
+ "[control_474]": 152133,
441
+ "[control_475]": 152134,
442
+ "[control_476]": 152135,
443
+ "[control_477]": 152136,
444
+ "[control_478]": 152137,
445
+ "[control_479]": 152138,
446
+ "[control_47]": 151706,
447
+ "[control_480]": 152139,
448
+ "[control_481]": 152140,
449
+ "[control_482]": 152141,
450
+ "[control_483]": 152142,
451
+ "[control_484]": 152143,
452
+ "[control_485]": 152144,
453
+ "[control_486]": 152145,
454
+ "[control_487]": 152146,
455
+ "[control_488]": 152147,
456
+ "[control_489]": 152148,
457
+ "[control_48]": 151707,
458
+ "[control_490]": 152149,
459
+ "[control_491]": 152150,
460
+ "[control_492]": 152151,
461
+ "[control_493]": 152152,
462
+ "[control_494]": 152153,
463
+ "[control_495]": 152154,
464
+ "[control_496]": 152155,
465
+ "[control_497]": 152156,
466
+ "[control_498]": 152157,
467
+ "[control_499]": 152158,
468
+ "[control_49]": 151708,
469
+ "[control_500]": 152159,
470
+ "[control_501]": 152160,
471
+ "[control_502]": 152161,
472
+ "[control_503]": 152162,
473
+ "[control_504]": 152163,
474
+ "[control_505]": 152164,
475
+ "[control_506]": 152165,
476
+ "[control_507]": 152166,
477
+ "[control_508]": 152167,
478
+ "[control_509]": 152168,
479
+ "[control_50]": 151709,
480
+ "[control_510]": 152169,
481
+ "[control_511]": 152170,
482
+ "[control_512]": 152171,
483
+ "[control_513]": 152172,
484
+ "[control_514]": 152173,
485
+ "[control_515]": 152174,
486
+ "[control_516]": 152175,
487
+ "[control_517]": 152176,
488
+ "[control_518]": 152177,
489
+ "[control_519]": 152178,
490
+ "[control_51]": 151710,
491
+ "[control_520]": 152179,
492
+ "[control_521]": 152180,
493
+ "[control_522]": 152181,
494
+ "[control_523]": 152182,
495
+ "[control_524]": 152183,
496
+ "[control_525]": 152184,
497
+ "[control_526]": 152185,
498
+ "[control_527]": 152186,
499
+ "[control_528]": 152187,
500
+ "[control_529]": 152188,
501
+ "[control_52]": 151711,
502
+ "[control_530]": 152189,
503
+ "[control_531]": 152190,
504
+ "[control_532]": 152191,
505
+ "[control_533]": 152192,
506
+ "[control_534]": 152193,
507
+ "[control_535]": 152194,
508
+ "[control_536]": 152195,
509
+ "[control_537]": 152196,
510
+ "[control_538]": 152197,
511
+ "[control_539]": 152198,
512
+ "[control_53]": 151712,
513
+ "[control_540]": 152199,
514
+ "[control_541]": 152200,
515
+ "[control_542]": 152201,
516
+ "[control_543]": 152202,
517
+ "[control_544]": 152203,
518
+ "[control_545]": 152204,
519
+ "[control_546]": 152205,
520
+ "[control_547]": 152206,
521
+ "[control_548]": 152207,
522
+ "[control_549]": 152208,
523
+ "[control_54]": 151713,
524
+ "[control_550]": 152209,
525
+ "[control_551]": 152210,
526
+ "[control_552]": 152211,
527
+ "[control_553]": 152212,
528
+ "[control_554]": 152213,
529
+ "[control_555]": 152214,
530
+ "[control_556]": 152215,
531
+ "[control_557]": 152216,
532
+ "[control_558]": 152217,
533
+ "[control_559]": 152218,
534
+ "[control_55]": 151714,
535
+ "[control_560]": 152219,
536
+ "[control_561]": 152220,
537
+ "[control_562]": 152221,
538
+ "[control_563]": 152222,
539
+ "[control_564]": 152223,
540
+ "[control_565]": 152224,
541
+ "[control_566]": 152225,
542
+ "[control_567]": 152226,
543
+ "[control_568]": 152227,
544
+ "[control_569]": 152228,
545
+ "[control_56]": 151715,
546
+ "[control_570]": 152229,
547
+ "[control_571]": 152230,
548
+ "[control_572]": 152231,
549
+ "[control_573]": 152232,
550
+ "[control_574]": 152233,
551
+ "[control_575]": 152234,
552
+ "[control_576]": 152235,
553
+ "[control_577]": 152236,
554
+ "[control_578]": 152237,
555
+ "[control_579]": 152238,
556
+ "[control_57]": 151716,
557
+ "[control_580]": 152239,
558
+ "[control_581]": 152240,
559
+ "[control_582]": 152241,
560
+ "[control_583]": 152242,
561
+ "[control_584]": 152243,
562
+ "[control_585]": 152244,
563
+ "[control_586]": 152245,
564
+ "[control_587]": 152246,
565
+ "[control_588]": 152247,
566
+ "[control_589]": 152248,
567
+ "[control_58]": 151717,
568
+ "[control_590]": 152249,
569
+ "[control_591]": 152250,
570
+ "[control_592]": 152251,
571
+ "[control_593]": 152252,
572
+ "[control_594]": 152253,
573
+ "[control_595]": 152254,
574
+ "[control_596]": 152255,
575
+ "[control_597]": 152256,
576
+ "[control_598]": 152257,
577
+ "[control_599]": 152258,
578
+ "[control_59]": 151718,
579
+ "[control_600]": 152259,
580
+ "[control_601]": 152260,
581
+ "[control_602]": 152261,
582
+ "[control_603]": 152262,
583
+ "[control_604]": 152263,
584
+ "[control_605]": 152264,
585
+ "[control_606]": 152265,
586
+ "[control_607]": 152266,
587
+ "[control_608]": 152267,
588
+ "[control_609]": 152268,
589
+ "[control_60]": 151719,
590
+ "[control_610]": 152269,
591
+ "[control_611]": 152270,
592
+ "[control_612]": 152271,
593
+ "[control_613]": 152272,
594
+ "[control_614]": 152273,
595
+ "[control_615]": 152274,
596
+ "[control_616]": 152275,
597
+ "[control_617]": 152276,
598
+ "[control_618]": 152277,
599
+ "[control_619]": 152278,
600
+ "[control_61]": 151720,
601
+ "[control_620]": 152279,
602
+ "[control_621]": 152280,
603
+ "[control_622]": 152281,
604
+ "[control_623]": 152282,
605
+ "[control_624]": 152283,
606
+ "[control_625]": 152284,
607
+ "[control_626]": 152285,
608
+ "[control_627]": 152286,
609
+ "[control_628]": 152287,
610
+ "[control_629]": 152288,
611
+ "[control_62]": 151721,
612
+ "[control_630]": 152289,
613
+ "[control_631]": 152290,
614
+ "[control_632]": 152291,
615
+ "[control_633]": 152292,
616
+ "[control_634]": 152293,
617
+ "[control_635]": 152294,
618
+ "[control_636]": 152295,
619
+ "[control_637]": 152296,
620
+ "[control_638]": 152297,
621
+ "[control_639]": 152298,
622
+ "[control_63]": 151722,
623
+ "[control_640]": 152299,
624
+ "[control_641]": 152300,
625
+ "[control_642]": 152301,
626
+ "[control_643]": 152302,
627
+ "[control_644]": 152303,
628
+ "[control_645]": 152304,
629
+ "[control_646]": 152305,
630
+ "[control_647]": 152306,
631
+ "[control_648]": 152307,
632
+ "[control_649]": 152308,
633
+ "[control_64]": 151723,
634
+ "[control_650]": 152309,
635
+ "[control_651]": 152310,
636
+ "[control_652]": 152311,
637
+ "[control_653]": 152312,
638
+ "[control_654]": 152313,
639
+ "[control_655]": 152314,
640
+ "[control_656]": 152315,
641
+ "[control_657]": 152316,
642
+ "[control_658]": 152317,
643
+ "[control_659]": 152318,
644
+ "[control_65]": 151724,
645
+ "[control_660]": 152319,
646
+ "[control_661]": 152320,
647
+ "[control_662]": 152321,
648
+ "[control_663]": 152322,
649
+ "[control_664]": 152323,
650
+ "[control_665]": 152324,
651
+ "[control_666]": 152325,
652
+ "[control_667]": 152326,
653
+ "[control_668]": 152327,
654
+ "[control_669]": 152328,
655
+ "[control_66]": 151725,
656
+ "[control_670]": 152329,
657
+ "[control_671]": 152330,
658
+ "[control_672]": 152331,
659
+ "[control_673]": 152332,
660
+ "[control_674]": 152333,
661
+ "[control_675]": 152334,
662
+ "[control_676]": 152335,
663
+ "[control_677]": 152336,
664
+ "[control_678]": 152337,
665
+ "[control_679]": 152338,
666
+ "[control_67]": 151726,
667
+ "[control_680]": 152339,
668
+ "[control_681]": 152340,
669
+ "[control_682]": 152341,
670
+ "[control_683]": 152342,
671
+ "[control_684]": 152343,
672
+ "[control_685]": 152344,
673
+ "[control_686]": 152345,
674
+ "[control_687]": 152346,
675
+ "[control_688]": 152347,
676
+ "[control_689]": 152348,
677
+ "[control_68]": 151727,
678
+ "[control_690]": 152349,
679
+ "[control_691]": 152350,
680
+ "[control_692]": 152351,
681
+ "[control_693]": 152352,
682
+ "[control_694]": 152353,
683
+ "[control_695]": 152354,
684
+ "[control_696]": 152355,
685
+ "[control_697]": 152356,
686
+ "[control_698]": 152357,
687
+ "[control_699]": 152358,
688
+ "[control_69]": 151728,
689
+ "[control_700]": 152359,
690
+ "[control_701]": 152360,
691
+ "[control_702]": 152361,
692
+ "[control_703]": 152362,
693
+ "[control_704]": 152363,
694
+ "[control_705]": 152364,
695
+ "[control_706]": 152365,
696
+ "[control_707]": 152366,
697
+ "[control_708]": 152367,
698
+ "[control_709]": 152368,
699
+ "[control_70]": 151729,
700
+ "[control_710]": 152369,
701
+ "[control_711]": 152370,
702
+ "[control_712]": 152371,
703
+ "[control_713]": 152372,
704
+ "[control_714]": 152373,
705
+ "[control_715]": 152374,
706
+ "[control_716]": 152375,
707
+ "[control_717]": 152376,
708
+ "[control_718]": 152377,
709
+ "[control_719]": 152378,
710
+ "[control_71]": 151730,
711
+ "[control_720]": 152379,
712
+ "[control_721]": 152380,
713
+ "[control_722]": 152381,
714
+ "[control_723]": 152382,
715
+ "[control_724]": 152383,
716
+ "[control_725]": 152384,
717
+ "[control_726]": 152385,
718
+ "[control_727]": 152386,
719
+ "[control_728]": 152387,
720
+ "[control_729]": 152388,
721
+ "[control_72]": 151731,
722
+ "[control_730]": 152389,
723
+ "[control_731]": 152390,
724
+ "[control_732]": 152391,
725
+ "[control_733]": 152392,
726
+ "[control_734]": 152393,
727
+ "[control_735]": 152394,
728
+ "[control_736]": 152395,
729
+ "[control_737]": 152396,
730
+ "[control_738]": 152397,
731
+ "[control_739]": 152398,
732
+ "[control_73]": 151732,
733
+ "[control_740]": 152399,
734
+ "[control_741]": 152400,
735
+ "[control_742]": 152401,
736
+ "[control_743]": 152402,
737
+ "[control_744]": 152403,
738
+ "[control_745]": 152404,
739
+ "[control_746]": 152405,
740
+ "[control_747]": 152406,
741
+ "[control_748]": 152407,
742
+ "[control_749]": 152408,
743
+ "[control_74]": 151733,
744
+ "[control_750]": 152409,
745
+ "[control_751]": 152410,
746
+ "[control_752]": 152411,
747
+ "[control_753]": 152412,
748
+ "[control_754]": 152413,
749
+ "[control_755]": 152414,
750
+ "[control_756]": 152415,
751
+ "[control_757]": 152416,
752
+ "[control_758]": 152417,
753
+ "[control_759]": 152418,
754
+ "[control_75]": 151734,
755
+ "[control_760]": 152419,
756
+ "[control_761]": 152420,
757
+ "[control_762]": 152421,
758
+ "[control_763]": 152422,
759
+ "[control_764]": 152423,
760
+ "[control_765]": 152424,
761
+ "[control_766]": 152425,
762
+ "[control_767]": 152426,
763
+ "[control_768]": 152427,
764
+ "[control_769]": 152428,
765
+ "[control_76]": 151735,
766
+ "[control_770]": 152429,
767
+ "[control_771]": 152430,
768
+ "[control_772]": 152431,
769
+ "[control_773]": 152432,
770
+ "[control_774]": 152433,
771
+ "[control_775]": 152434,
772
+ "[control_776]": 152435,
773
+ "[control_777]": 152436,
774
+ "[control_778]": 152437,
775
+ "[control_779]": 152438,
776
+ "[control_77]": 151736,
777
+ "[control_780]": 152439,
778
+ "[control_781]": 152440,
779
+ "[control_782]": 152441,
780
+ "[control_783]": 152442,
781
+ "[control_784]": 152443,
782
+ "[control_785]": 152444,
783
+ "[control_786]": 152445,
784
+ "[control_787]": 152446,
785
+ "[control_788]": 152447,
786
+ "[control_789]": 152448,
787
+ "[control_78]": 151737,
788
+ "[control_790]": 152449,
789
+ "[control_791]": 152450,
790
+ "[control_792]": 152451,
791
+ "[control_793]": 152452,
792
+ "[control_794]": 152453,
793
+ "[control_795]": 152454,
794
+ "[control_796]": 152455,
795
+ "[control_797]": 152456,
796
+ "[control_798]": 152457,
797
+ "[control_799]": 152458,
798
+ "[control_79]": 151738,
799
+ "[control_800]": 152459,
800
+ "[control_801]": 152460,
801
+ "[control_802]": 152461,
802
+ "[control_803]": 152462,
803
+ "[control_804]": 152463,
804
+ "[control_805]": 152464,
805
+ "[control_806]": 152465,
806
+ "[control_807]": 152466,
807
+ "[control_808]": 152467,
808
+ "[control_809]": 152468,
809
+ "[control_80]": 151739,
810
+ "[control_810]": 152469,
811
+ "[control_811]": 152470,
812
+ "[control_812]": 152471,
813
+ "[control_813]": 152472,
814
+ "[control_814]": 152473,
815
+ "[control_815]": 152474,
816
+ "[control_816]": 152475,
817
+ "[control_817]": 152476,
818
+ "[control_818]": 152477,
819
+ "[control_819]": 152478,
820
+ "[control_81]": 151740,
821
+ "[control_820]": 152479,
822
+ "[control_821]": 152480,
823
+ "[control_822]": 152481,
824
+ "[control_823]": 152482,
825
+ "[control_824]": 152483,
826
+ "[control_825]": 152484,
827
+ "[control_826]": 152485,
828
+ "[control_827]": 152486,
829
+ "[control_828]": 152487,
830
+ "[control_829]": 152488,
831
+ "[control_82]": 151741,
832
+ "[control_830]": 152489,
833
+ "[control_831]": 152490,
834
+ "[control_832]": 152491,
835
+ "[control_833]": 152492,
836
+ "[control_834]": 152493,
837
+ "[control_835]": 152494,
838
+ "[control_836]": 152495,
839
+ "[control_837]": 152496,
840
+ "[control_838]": 152497,
841
+ "[control_839]": 152498,
842
+ "[control_83]": 151742,
843
+ "[control_840]": 152499,
844
+ "[control_841]": 152500,
845
+ "[control_842]": 152501,
846
+ "[control_843]": 152502,
847
+ "[control_844]": 152503,
848
+ "[control_845]": 152504,
849
+ "[control_846]": 152505,
850
+ "[control_847]": 152506,
851
+ "[control_848]": 152507,
852
+ "[control_849]": 152508,
853
+ "[control_84]": 151743,
854
+ "[control_850]": 152509,
855
+ "[control_851]": 152510,
856
+ "[control_852]": 152511,
857
+ "[control_853]": 152512,
858
+ "[control_854]": 152513,
859
+ "[control_855]": 152514,
860
+ "[control_856]": 152515,
861
+ "[control_857]": 152516,
862
+ "[control_858]": 152517,
863
+ "[control_859]": 152518,
864
+ "[control_85]": 151744,
865
+ "[control_860]": 152519,
866
+ "[control_861]": 152520,
867
+ "[control_862]": 152521,
868
+ "[control_863]": 152522,
869
+ "[control_864]": 152523,
870
+ "[control_865]": 152524,
871
+ "[control_866]": 152525,
872
+ "[control_867]": 152526,
873
+ "[control_868]": 152527,
874
+ "[control_869]": 152528,
875
+ "[control_86]": 151745,
876
+ "[control_870]": 152529,
877
+ "[control_871]": 152530,
878
+ "[control_872]": 152531,
879
+ "[control_873]": 152532,
880
+ "[control_874]": 152533,
881
+ "[control_875]": 152534,
882
+ "[control_876]": 152535,
883
+ "[control_877]": 152536,
884
+ "[control_878]": 152537,
885
+ "[control_879]": 152538,
886
+ "[control_87]": 151746,
887
+ "[control_880]": 152539,
888
+ "[control_881]": 152540,
889
+ "[control_882]": 152541,
890
+ "[control_883]": 152542,
891
+ "[control_884]": 152543,
892
+ "[control_885]": 152544,
893
+ "[control_886]": 152545,
894
+ "[control_887]": 152546,
895
+ "[control_888]": 152547,
896
+ "[control_889]": 152548,
897
+ "[control_88]": 151747,
898
+ "[control_890]": 152549,
899
+ "[control_891]": 152550,
900
+ "[control_892]": 152551,
901
+ "[control_893]": 152552,
902
+ "[control_894]": 152553,
903
+ "[control_895]": 152554,
904
+ "[control_896]": 152555,
905
+ "[control_897]": 152556,
906
+ "[control_898]": 152557,
907
+ "[control_899]": 152558,
908
+ "[control_89]": 151748,
909
+ "[control_900]": 152559,
910
+ "[control_901]": 152560,
911
+ "[control_902]": 152561,
912
+ "[control_903]": 152562,
913
+ "[control_904]": 152563,
914
+ "[control_905]": 152564,
915
+ "[control_906]": 152565,
916
+ "[control_907]": 152566,
917
+ "[control_908]": 152567,
918
+ "[control_909]": 152568,
919
+ "[control_90]": 151749,
920
+ "[control_910]": 152569,
921
+ "[control_911]": 152570,
922
+ "[control_912]": 152571,
923
+ "[control_913]": 152572,
924
+ "[control_914]": 152573,
925
+ "[control_915]": 152574,
926
+ "[control_916]": 152575,
927
+ "[control_917]": 152576,
928
+ "[control_918]": 152577,
929
+ "[control_919]": 152578,
930
+ "[control_91]": 151750,
931
+ "[control_920]": 152579,
932
+ "[control_921]": 152580,
933
+ "[control_922]": 152581,
934
+ "[control_923]": 152582,
935
+ "[control_924]": 152583,
936
+ "[control_925]": 152584,
937
+ "[control_926]": 152585,
938
+ "[control_927]": 152586,
939
+ "[control_928]": 152587,
940
+ "[control_929]": 152588,
941
+ "[control_92]": 151751,
942
+ "[control_930]": 152589,
943
+ "[control_931]": 152590,
944
+ "[control_932]": 152591,
945
+ "[control_933]": 152592,
946
+ "[control_934]": 152593,
947
+ "[control_935]": 152594,
948
+ "[control_936]": 152595,
949
+ "[control_937]": 152596,
950
+ "[control_938]": 152597,
951
+ "[control_939]": 152598,
952
+ "[control_93]": 151752,
953
+ "[control_940]": 152599,
954
+ "[control_941]": 152600,
955
+ "[control_942]": 152601,
956
+ "[control_943]": 152602,
957
+ "[control_944]": 152603,
958
+ "[control_945]": 152604,
959
+ "[control_946]": 152605,
960
+ "[control_947]": 152606,
961
+ "[control_948]": 152607,
962
+ "[control_949]": 152608,
963
+ "[control_94]": 151753,
964
+ "[control_950]": 152609,
965
+ "[control_951]": 152610,
966
+ "[control_952]": 152611,
967
+ "[control_953]": 152612,
968
+ "[control_954]": 152613,
969
+ "[control_955]": 152614,
970
+ "[control_956]": 152615,
971
+ "[control_957]": 152616,
972
+ "[control_958]": 152617,
973
+ "[control_959]": 152618,
974
+ "[control_95]": 151754,
975
+ "[control_960]": 152619,
976
+ "[control_961]": 152620,
977
+ "[control_962]": 152621,
978
+ "[control_963]": 152622,
979
+ "[control_964]": 152623,
980
+ "[control_965]": 152624,
981
+ "[control_966]": 152625,
982
+ "[control_967]": 152626,
983
+ "[control_968]": 152627,
984
+ "[control_969]": 152628,
985
+ "[control_96]": 151755,
986
+ "[control_970]": 152629,
987
+ "[control_971]": 152630,
988
+ "[control_972]": 152631,
989
+ "[control_973]": 152632,
990
+ "[control_974]": 152633,
991
+ "[control_975]": 152634,
992
+ "[control_976]": 152635,
993
+ "[control_977]": 152636,
994
+ "[control_978]": 152637,
995
+ "[control_979]": 152638,
996
+ "[control_97]": 151756,
997
+ "[control_980]": 152639,
998
+ "[control_981]": 152640,
999
+ "[control_982]": 152641,
1000
+ "[control_983]": 152642,
1001
+ "[control_984]": 152643,
1002
+ "[control_985]": 152644,
1003
+ "[control_986]": 152645,
1004
+ "[control_987]": 152646,
1005
+ "[control_988]": 152647,
1006
+ "[control_989]": 152648,
1007
+ "[control_98]": 151757,
1008
+ "[control_990]": 152649,
1009
+ "[control_991]": 152650,
1010
+ "[control_992]": 152651,
1011
+ "[control_993]": 152652,
1012
+ "[control_994]": 152653,
1013
+ "[control_995]": 152654,
1014
+ "[control_996]": 152655,
1015
+ "[control_997]": 152656,
1016
+ "[control_998]": 152657,
1017
+ "[control_999]": 152658,
1018
+ "[control_99]": 151758
1019
+ }
config.json ADDED
@@ -0,0 +1,72 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "architectures": [
3
+ "Qwen3ForCausalLM"
4
+ ],
5
+ "attention_bias": false,
6
+ "attention_dropout": 0.0,
7
+ "bos_token_id": 151643,
8
+ "eos_token_id": 151645,
9
+ "head_dim": 128,
10
+ "hidden_act": "silu",
11
+ "hidden_size": 5120,
12
+ "initializer_range": 0.02,
13
+ "intermediate_size": 17408,
14
+ "max_position_embeddings": 40960,
15
+ "max_window_layers": 40,
16
+ "model_type": "qwen3",
17
+ "num_attention_heads": 40,
18
+ "num_hidden_layers": 40,
19
+ "num_key_value_heads": 8,
20
+ "quantization_config": {
21
+ "config_groups": {
22
+ "group_0": {
23
+ "input_activations": {
24
+ "actorder": null,
25
+ "block_structure": null,
26
+ "dynamic": true,
27
+ "group_size": null,
28
+ "num_bits": 8,
29
+ "observer": null,
30
+ "observer_kwargs": {},
31
+ "strategy": "token",
32
+ "symmetric": true,
33
+ "type": "float"
34
+ },
35
+ "output_activations": null,
36
+ "targets": [
37
+ "Linear"
38
+ ],
39
+ "weights": {
40
+ "actorder": null,
41
+ "block_structure": null,
42
+ "dynamic": false,
43
+ "group_size": null,
44
+ "num_bits": 8,
45
+ "observer": "minmax",
46
+ "observer_kwargs": {},
47
+ "strategy": "channel",
48
+ "symmetric": true,
49
+ "type": "float"
50
+ }
51
+ }
52
+ },
53
+ "format": "float-quantized",
54
+ "global_compression_ratio": null,
55
+ "ignore": [
56
+ "lm_head"
57
+ ],
58
+ "kv_cache_scheme": null,
59
+ "quant_method": "compressed-tensors",
60
+ "quantization_status": "compressed"
61
+ },
62
+ "rms_norm_eps": 1e-06,
63
+ "rope_scaling": null,
64
+ "rope_theta": 1000000,
65
+ "sliding_window": null,
66
+ "tie_word_embeddings": false,
67
+ "torch_dtype": "bfloat16",
68
+ "transformers_version": "4.51.1",
69
+ "use_cache": true,
70
+ "use_sliding_window": false,
71
+ "vocab_size": 152660
72
+ }
generation_config.json ADDED
@@ -0,0 +1,13 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "bos_token_id": 151643,
3
+ "do_sample": true,
4
+ "eos_token_id": [
5
+ 151645,
6
+ 151643
7
+ ],
8
+ "pad_token_id": 151643,
9
+ "temperature": 0.6,
10
+ "top_k": 20,
11
+ "top_p": 0.95,
12
+ "transformers_version": "4.51.1"
13
+ }
merges.txt ADDED
The diff for this file is too large to render. See raw diff
 
model-00001-of-00004.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:dfc4b697f085fa8bd6d1d054bccf736b9f082e233f0344730e56ff3c822de4ff
3
+ size 4930469344
model-00002-of-00004.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4228280b53960b7d0eab62f2729dbcbd50e266ba97e2a47f5ede4a7b8aa4f131
3
+ size 4956434040
model-00003-of-00004.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8cc8d1187910fed69f6eac3734cf064cd194ef68345e1f3fca7e7ec8807e04be
3
+ size 4893503560
model-00004-of-00004.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:be72e24d1ea5dd8a7b8b1c03dfc877b2713b56e4d8697bd7541eaac55077d07d
3
+ size 1563238528
model.safetensors.index.json ADDED
@@ -0,0 +1,730 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "metadata": {
3
+ "total_size": 16343562240
4
+ },
5
+ "weight_map": {
6
+ "lm_head.weight": "model-00004-of-00004.safetensors",
7
+ "model.embed_tokens.weight": "model-00001-of-00004.safetensors",
8
+ "model.layers.0.input_layernorm.weight": "model-00001-of-00004.safetensors",
9
+ "model.layers.0.mlp.down_proj.weight": "model-00001-of-00004.safetensors",
10
+ "model.layers.0.mlp.down_proj.weight_scale": "model-00001-of-00004.safetensors",
11
+ "model.layers.0.mlp.gate_proj.weight": "model-00001-of-00004.safetensors",
12
+ "model.layers.0.mlp.gate_proj.weight_scale": "model-00001-of-00004.safetensors",
13
+ "model.layers.0.mlp.up_proj.weight": "model-00001-of-00004.safetensors",
14
+ "model.layers.0.mlp.up_proj.weight_scale": "model-00001-of-00004.safetensors",
15
+ "model.layers.0.post_attention_layernorm.weight": "model-00001-of-00004.safetensors",
16
+ "model.layers.0.self_attn.k_norm.weight": "model-00001-of-00004.safetensors",
17
+ "model.layers.0.self_attn.k_proj.weight": "model-00001-of-00004.safetensors",
18
+ "model.layers.0.self_attn.k_proj.weight_scale": "model-00001-of-00004.safetensors",
19
+ "model.layers.0.self_attn.o_proj.weight": "model-00001-of-00004.safetensors",
20
+ "model.layers.0.self_attn.o_proj.weight_scale": "model-00001-of-00004.safetensors",
21
+ "model.layers.0.self_attn.q_norm.weight": "model-00001-of-00004.safetensors",
22
+ "model.layers.0.self_attn.q_proj.weight": "model-00001-of-00004.safetensors",
23
+ "model.layers.0.self_attn.q_proj.weight_scale": "model-00001-of-00004.safetensors",
24
+ "model.layers.0.self_attn.v_proj.weight": "model-00001-of-00004.safetensors",
25
+ "model.layers.0.self_attn.v_proj.weight_scale": "model-00001-of-00004.safetensors",
26
+ "model.layers.1.input_layernorm.weight": "model-00001-of-00004.safetensors",
27
+ "model.layers.1.mlp.down_proj.weight": "model-00001-of-00004.safetensors",
28
+ "model.layers.1.mlp.down_proj.weight_scale": "model-00001-of-00004.safetensors",
29
+ "model.layers.1.mlp.gate_proj.weight": "model-00001-of-00004.safetensors",
30
+ "model.layers.1.mlp.gate_proj.weight_scale": "model-00001-of-00004.safetensors",
31
+ "model.layers.1.mlp.up_proj.weight": "model-00001-of-00004.safetensors",
32
+ "model.layers.1.mlp.up_proj.weight_scale": "model-00001-of-00004.safetensors",
33
+ "model.layers.1.post_attention_layernorm.weight": "model-00001-of-00004.safetensors",
34
+ "model.layers.1.self_attn.k_norm.weight": "model-00001-of-00004.safetensors",
35
+ "model.layers.1.self_attn.k_proj.weight": "model-00001-of-00004.safetensors",
36
+ "model.layers.1.self_attn.k_proj.weight_scale": "model-00001-of-00004.safetensors",
37
+ "model.layers.1.self_attn.o_proj.weight": "model-00001-of-00004.safetensors",
38
+ "model.layers.1.self_attn.o_proj.weight_scale": "model-00001-of-00004.safetensors",
39
+ "model.layers.1.self_attn.q_norm.weight": "model-00001-of-00004.safetensors",
40
+ "model.layers.1.self_attn.q_proj.weight": "model-00001-of-00004.safetensors",
41
+ "model.layers.1.self_attn.q_proj.weight_scale": "model-00001-of-00004.safetensors",
42
+ "model.layers.1.self_attn.v_proj.weight": "model-00001-of-00004.safetensors",
43
+ "model.layers.1.self_attn.v_proj.weight_scale": "model-00001-of-00004.safetensors",
44
+ "model.layers.10.input_layernorm.weight": "model-00002-of-00004.safetensors",
45
+ "model.layers.10.mlp.down_proj.weight": "model-00002-of-00004.safetensors",
46
+ "model.layers.10.mlp.down_proj.weight_scale": "model-00002-of-00004.safetensors",
47
+ "model.layers.10.mlp.gate_proj.weight": "model-00002-of-00004.safetensors",
48
+ "model.layers.10.mlp.gate_proj.weight_scale": "model-00002-of-00004.safetensors",
49
+ "model.layers.10.mlp.up_proj.weight": "model-00002-of-00004.safetensors",
50
+ "model.layers.10.mlp.up_proj.weight_scale": "model-00002-of-00004.safetensors",
51
+ "model.layers.10.post_attention_layernorm.weight": "model-00002-of-00004.safetensors",
52
+ "model.layers.10.self_attn.k_norm.weight": "model-00001-of-00004.safetensors",
53
+ "model.layers.10.self_attn.k_proj.weight": "model-00001-of-00004.safetensors",
54
+ "model.layers.10.self_attn.k_proj.weight_scale": "model-00001-of-00004.safetensors",
55
+ "model.layers.10.self_attn.o_proj.weight": "model-00001-of-00004.safetensors",
56
+ "model.layers.10.self_attn.o_proj.weight_scale": "model-00001-of-00004.safetensors",
57
+ "model.layers.10.self_attn.q_norm.weight": "model-00001-of-00004.safetensors",
58
+ "model.layers.10.self_attn.q_proj.weight": "model-00001-of-00004.safetensors",
59
+ "model.layers.10.self_attn.q_proj.weight_scale": "model-00001-of-00004.safetensors",
60
+ "model.layers.10.self_attn.v_proj.weight": "model-00001-of-00004.safetensors",
61
+ "model.layers.10.self_attn.v_proj.weight_scale": "model-00001-of-00004.safetensors",
62
+ "model.layers.11.input_layernorm.weight": "model-00002-of-00004.safetensors",
63
+ "model.layers.11.mlp.down_proj.weight": "model-00002-of-00004.safetensors",
64
+ "model.layers.11.mlp.down_proj.weight_scale": "model-00002-of-00004.safetensors",
65
+ "model.layers.11.mlp.gate_proj.weight": "model-00002-of-00004.safetensors",
66
+ "model.layers.11.mlp.gate_proj.weight_scale": "model-00002-of-00004.safetensors",
67
+ "model.layers.11.mlp.up_proj.weight": "model-00002-of-00004.safetensors",
68
+ "model.layers.11.mlp.up_proj.weight_scale": "model-00002-of-00004.safetensors",
69
+ "model.layers.11.post_attention_layernorm.weight": "model-00002-of-00004.safetensors",
70
+ "model.layers.11.self_attn.k_norm.weight": "model-00002-of-00004.safetensors",
71
+ "model.layers.11.self_attn.k_proj.weight": "model-00002-of-00004.safetensors",
72
+ "model.layers.11.self_attn.k_proj.weight_scale": "model-00002-of-00004.safetensors",
73
+ "model.layers.11.self_attn.o_proj.weight": "model-00002-of-00004.safetensors",
74
+ "model.layers.11.self_attn.o_proj.weight_scale": "model-00002-of-00004.safetensors",
75
+ "model.layers.11.self_attn.q_norm.weight": "model-00002-of-00004.safetensors",
76
+ "model.layers.11.self_attn.q_proj.weight": "model-00002-of-00004.safetensors",
77
+ "model.layers.11.self_attn.q_proj.weight_scale": "model-00002-of-00004.safetensors",
78
+ "model.layers.11.self_attn.v_proj.weight": "model-00002-of-00004.safetensors",
79
+ "model.layers.11.self_attn.v_proj.weight_scale": "model-00002-of-00004.safetensors",
80
+ "model.layers.12.input_layernorm.weight": "model-00002-of-00004.safetensors",
81
+ "model.layers.12.mlp.down_proj.weight": "model-00002-of-00004.safetensors",
82
+ "model.layers.12.mlp.down_proj.weight_scale": "model-00002-of-00004.safetensors",
83
+ "model.layers.12.mlp.gate_proj.weight": "model-00002-of-00004.safetensors",
84
+ "model.layers.12.mlp.gate_proj.weight_scale": "model-00002-of-00004.safetensors",
85
+ "model.layers.12.mlp.up_proj.weight": "model-00002-of-00004.safetensors",
86
+ "model.layers.12.mlp.up_proj.weight_scale": "model-00002-of-00004.safetensors",
87
+ "model.layers.12.post_attention_layernorm.weight": "model-00002-of-00004.safetensors",
88
+ "model.layers.12.self_attn.k_norm.weight": "model-00002-of-00004.safetensors",
89
+ "model.layers.12.self_attn.k_proj.weight": "model-00002-of-00004.safetensors",
90
+ "model.layers.12.self_attn.k_proj.weight_scale": "model-00002-of-00004.safetensors",
91
+ "model.layers.12.self_attn.o_proj.weight": "model-00002-of-00004.safetensors",
92
+ "model.layers.12.self_attn.o_proj.weight_scale": "model-00002-of-00004.safetensors",
93
+ "model.layers.12.self_attn.q_norm.weight": "model-00002-of-00004.safetensors",
94
+ "model.layers.12.self_attn.q_proj.weight": "model-00002-of-00004.safetensors",
95
+ "model.layers.12.self_attn.q_proj.weight_scale": "model-00002-of-00004.safetensors",
96
+ "model.layers.12.self_attn.v_proj.weight": "model-00002-of-00004.safetensors",
97
+ "model.layers.12.self_attn.v_proj.weight_scale": "model-00002-of-00004.safetensors",
98
+ "model.layers.13.input_layernorm.weight": "model-00002-of-00004.safetensors",
99
+ "model.layers.13.mlp.down_proj.weight": "model-00002-of-00004.safetensors",
100
+ "model.layers.13.mlp.down_proj.weight_scale": "model-00002-of-00004.safetensors",
101
+ "model.layers.13.mlp.gate_proj.weight": "model-00002-of-00004.safetensors",
102
+ "model.layers.13.mlp.gate_proj.weight_scale": "model-00002-of-00004.safetensors",
103
+ "model.layers.13.mlp.up_proj.weight": "model-00002-of-00004.safetensors",
104
+ "model.layers.13.mlp.up_proj.weight_scale": "model-00002-of-00004.safetensors",
105
+ "model.layers.13.post_attention_layernorm.weight": "model-00002-of-00004.safetensors",
106
+ "model.layers.13.self_attn.k_norm.weight": "model-00002-of-00004.safetensors",
107
+ "model.layers.13.self_attn.k_proj.weight": "model-00002-of-00004.safetensors",
108
+ "model.layers.13.self_attn.k_proj.weight_scale": "model-00002-of-00004.safetensors",
109
+ "model.layers.13.self_attn.o_proj.weight": "model-00002-of-00004.safetensors",
110
+ "model.layers.13.self_attn.o_proj.weight_scale": "model-00002-of-00004.safetensors",
111
+ "model.layers.13.self_attn.q_norm.weight": "model-00002-of-00004.safetensors",
112
+ "model.layers.13.self_attn.q_proj.weight": "model-00002-of-00004.safetensors",
113
+ "model.layers.13.self_attn.q_proj.weight_scale": "model-00002-of-00004.safetensors",
114
+ "model.layers.13.self_attn.v_proj.weight": "model-00002-of-00004.safetensors",
115
+ "model.layers.13.self_attn.v_proj.weight_scale": "model-00002-of-00004.safetensors",
116
+ "model.layers.14.input_layernorm.weight": "model-00002-of-00004.safetensors",
117
+ "model.layers.14.mlp.down_proj.weight": "model-00002-of-00004.safetensors",
118
+ "model.layers.14.mlp.down_proj.weight_scale": "model-00002-of-00004.safetensors",
119
+ "model.layers.14.mlp.gate_proj.weight": "model-00002-of-00004.safetensors",
120
+ "model.layers.14.mlp.gate_proj.weight_scale": "model-00002-of-00004.safetensors",
121
+ "model.layers.14.mlp.up_proj.weight": "model-00002-of-00004.safetensors",
122
+ "model.layers.14.mlp.up_proj.weight_scale": "model-00002-of-00004.safetensors",
123
+ "model.layers.14.post_attention_layernorm.weight": "model-00002-of-00004.safetensors",
124
+ "model.layers.14.self_attn.k_norm.weight": "model-00002-of-00004.safetensors",
125
+ "model.layers.14.self_attn.k_proj.weight": "model-00002-of-00004.safetensors",
126
+ "model.layers.14.self_attn.k_proj.weight_scale": "model-00002-of-00004.safetensors",
127
+ "model.layers.14.self_attn.o_proj.weight": "model-00002-of-00004.safetensors",
128
+ "model.layers.14.self_attn.o_proj.weight_scale": "model-00002-of-00004.safetensors",
129
+ "model.layers.14.self_attn.q_norm.weight": "model-00002-of-00004.safetensors",
130
+ "model.layers.14.self_attn.q_proj.weight": "model-00002-of-00004.safetensors",
131
+ "model.layers.14.self_attn.q_proj.weight_scale": "model-00002-of-00004.safetensors",
132
+ "model.layers.14.self_attn.v_proj.weight": "model-00002-of-00004.safetensors",
133
+ "model.layers.14.self_attn.v_proj.weight_scale": "model-00002-of-00004.safetensors",
134
+ "model.layers.15.input_layernorm.weight": "model-00002-of-00004.safetensors",
135
+ "model.layers.15.mlp.down_proj.weight": "model-00002-of-00004.safetensors",
136
+ "model.layers.15.mlp.down_proj.weight_scale": "model-00002-of-00004.safetensors",
137
+ "model.layers.15.mlp.gate_proj.weight": "model-00002-of-00004.safetensors",
138
+ "model.layers.15.mlp.gate_proj.weight_scale": "model-00002-of-00004.safetensors",
139
+ "model.layers.15.mlp.up_proj.weight": "model-00002-of-00004.safetensors",
140
+ "model.layers.15.mlp.up_proj.weight_scale": "model-00002-of-00004.safetensors",
141
+ "model.layers.15.post_attention_layernorm.weight": "model-00002-of-00004.safetensors",
142
+ "model.layers.15.self_attn.k_norm.weight": "model-00002-of-00004.safetensors",
143
+ "model.layers.15.self_attn.k_proj.weight": "model-00002-of-00004.safetensors",
144
+ "model.layers.15.self_attn.k_proj.weight_scale": "model-00002-of-00004.safetensors",
145
+ "model.layers.15.self_attn.o_proj.weight": "model-00002-of-00004.safetensors",
146
+ "model.layers.15.self_attn.o_proj.weight_scale": "model-00002-of-00004.safetensors",
147
+ "model.layers.15.self_attn.q_norm.weight": "model-00002-of-00004.safetensors",
148
+ "model.layers.15.self_attn.q_proj.weight": "model-00002-of-00004.safetensors",
149
+ "model.layers.15.self_attn.q_proj.weight_scale": "model-00002-of-00004.safetensors",
150
+ "model.layers.15.self_attn.v_proj.weight": "model-00002-of-00004.safetensors",
151
+ "model.layers.15.self_attn.v_proj.weight_scale": "model-00002-of-00004.safetensors",
152
+ "model.layers.16.input_layernorm.weight": "model-00002-of-00004.safetensors",
153
+ "model.layers.16.mlp.down_proj.weight": "model-00002-of-00004.safetensors",
154
+ "model.layers.16.mlp.down_proj.weight_scale": "model-00002-of-00004.safetensors",
155
+ "model.layers.16.mlp.gate_proj.weight": "model-00002-of-00004.safetensors",
156
+ "model.layers.16.mlp.gate_proj.weight_scale": "model-00002-of-00004.safetensors",
157
+ "model.layers.16.mlp.up_proj.weight": "model-00002-of-00004.safetensors",
158
+ "model.layers.16.mlp.up_proj.weight_scale": "model-00002-of-00004.safetensors",
159
+ "model.layers.16.post_attention_layernorm.weight": "model-00002-of-00004.safetensors",
160
+ "model.layers.16.self_attn.k_norm.weight": "model-00002-of-00004.safetensors",
161
+ "model.layers.16.self_attn.k_proj.weight": "model-00002-of-00004.safetensors",
162
+ "model.layers.16.self_attn.k_proj.weight_scale": "model-00002-of-00004.safetensors",
163
+ "model.layers.16.self_attn.o_proj.weight": "model-00002-of-00004.safetensors",
164
+ "model.layers.16.self_attn.o_proj.weight_scale": "model-00002-of-00004.safetensors",
165
+ "model.layers.16.self_attn.q_norm.weight": "model-00002-of-00004.safetensors",
166
+ "model.layers.16.self_attn.q_proj.weight": "model-00002-of-00004.safetensors",
167
+ "model.layers.16.self_attn.q_proj.weight_scale": "model-00002-of-00004.safetensors",
168
+ "model.layers.16.self_attn.v_proj.weight": "model-00002-of-00004.safetensors",
169
+ "model.layers.16.self_attn.v_proj.weight_scale": "model-00002-of-00004.safetensors",
170
+ "model.layers.17.input_layernorm.weight": "model-00002-of-00004.safetensors",
171
+ "model.layers.17.mlp.down_proj.weight": "model-00002-of-00004.safetensors",
172
+ "model.layers.17.mlp.down_proj.weight_scale": "model-00002-of-00004.safetensors",
173
+ "model.layers.17.mlp.gate_proj.weight": "model-00002-of-00004.safetensors",
174
+ "model.layers.17.mlp.gate_proj.weight_scale": "model-00002-of-00004.safetensors",
175
+ "model.layers.17.mlp.up_proj.weight": "model-00002-of-00004.safetensors",
176
+ "model.layers.17.mlp.up_proj.weight_scale": "model-00002-of-00004.safetensors",
177
+ "model.layers.17.post_attention_layernorm.weight": "model-00002-of-00004.safetensors",
178
+ "model.layers.17.self_attn.k_norm.weight": "model-00002-of-00004.safetensors",
179
+ "model.layers.17.self_attn.k_proj.weight": "model-00002-of-00004.safetensors",
180
+ "model.layers.17.self_attn.k_proj.weight_scale": "model-00002-of-00004.safetensors",
181
+ "model.layers.17.self_attn.o_proj.weight": "model-00002-of-00004.safetensors",
182
+ "model.layers.17.self_attn.o_proj.weight_scale": "model-00002-of-00004.safetensors",
183
+ "model.layers.17.self_attn.q_norm.weight": "model-00002-of-00004.safetensors",
184
+ "model.layers.17.self_attn.q_proj.weight": "model-00002-of-00004.safetensors",
185
+ "model.layers.17.self_attn.q_proj.weight_scale": "model-00002-of-00004.safetensors",
186
+ "model.layers.17.self_attn.v_proj.weight": "model-00002-of-00004.safetensors",
187
+ "model.layers.17.self_attn.v_proj.weight_scale": "model-00002-of-00004.safetensors",
188
+ "model.layers.18.input_layernorm.weight": "model-00002-of-00004.safetensors",
189
+ "model.layers.18.mlp.down_proj.weight": "model-00002-of-00004.safetensors",
190
+ "model.layers.18.mlp.down_proj.weight_scale": "model-00002-of-00004.safetensors",
191
+ "model.layers.18.mlp.gate_proj.weight": "model-00002-of-00004.safetensors",
192
+ "model.layers.18.mlp.gate_proj.weight_scale": "model-00002-of-00004.safetensors",
193
+ "model.layers.18.mlp.up_proj.weight": "model-00002-of-00004.safetensors",
194
+ "model.layers.18.mlp.up_proj.weight_scale": "model-00002-of-00004.safetensors",
195
+ "model.layers.18.post_attention_layernorm.weight": "model-00002-of-00004.safetensors",
196
+ "model.layers.18.self_attn.k_norm.weight": "model-00002-of-00004.safetensors",
197
+ "model.layers.18.self_attn.k_proj.weight": "model-00002-of-00004.safetensors",
198
+ "model.layers.18.self_attn.k_proj.weight_scale": "model-00002-of-00004.safetensors",
199
+ "model.layers.18.self_attn.o_proj.weight": "model-00002-of-00004.safetensors",
200
+ "model.layers.18.self_attn.o_proj.weight_scale": "model-00002-of-00004.safetensors",
201
+ "model.layers.18.self_attn.q_norm.weight": "model-00002-of-00004.safetensors",
202
+ "model.layers.18.self_attn.q_proj.weight": "model-00002-of-00004.safetensors",
203
+ "model.layers.18.self_attn.q_proj.weight_scale": "model-00002-of-00004.safetensors",
204
+ "model.layers.18.self_attn.v_proj.weight": "model-00002-of-00004.safetensors",
205
+ "model.layers.18.self_attn.v_proj.weight_scale": "model-00002-of-00004.safetensors",
206
+ "model.layers.19.input_layernorm.weight": "model-00002-of-00004.safetensors",
207
+ "model.layers.19.mlp.down_proj.weight": "model-00002-of-00004.safetensors",
208
+ "model.layers.19.mlp.down_proj.weight_scale": "model-00002-of-00004.safetensors",
209
+ "model.layers.19.mlp.gate_proj.weight": "model-00002-of-00004.safetensors",
210
+ "model.layers.19.mlp.gate_proj.weight_scale": "model-00002-of-00004.safetensors",
211
+ "model.layers.19.mlp.up_proj.weight": "model-00002-of-00004.safetensors",
212
+ "model.layers.19.mlp.up_proj.weight_scale": "model-00002-of-00004.safetensors",
213
+ "model.layers.19.post_attention_layernorm.weight": "model-00002-of-00004.safetensors",
214
+ "model.layers.19.self_attn.k_norm.weight": "model-00002-of-00004.safetensors",
215
+ "model.layers.19.self_attn.k_proj.weight": "model-00002-of-00004.safetensors",
216
+ "model.layers.19.self_attn.k_proj.weight_scale": "model-00002-of-00004.safetensors",
217
+ "model.layers.19.self_attn.o_proj.weight": "model-00002-of-00004.safetensors",
218
+ "model.layers.19.self_attn.o_proj.weight_scale": "model-00002-of-00004.safetensors",
219
+ "model.layers.19.self_attn.q_norm.weight": "model-00002-of-00004.safetensors",
220
+ "model.layers.19.self_attn.q_proj.weight": "model-00002-of-00004.safetensors",
221
+ "model.layers.19.self_attn.q_proj.weight_scale": "model-00002-of-00004.safetensors",
222
+ "model.layers.19.self_attn.v_proj.weight": "model-00002-of-00004.safetensors",
223
+ "model.layers.19.self_attn.v_proj.weight_scale": "model-00002-of-00004.safetensors",
224
+ "model.layers.2.input_layernorm.weight": "model-00001-of-00004.safetensors",
225
+ "model.layers.2.mlp.down_proj.weight": "model-00001-of-00004.safetensors",
226
+ "model.layers.2.mlp.down_proj.weight_scale": "model-00001-of-00004.safetensors",
227
+ "model.layers.2.mlp.gate_proj.weight": "model-00001-of-00004.safetensors",
228
+ "model.layers.2.mlp.gate_proj.weight_scale": "model-00001-of-00004.safetensors",
229
+ "model.layers.2.mlp.up_proj.weight": "model-00001-of-00004.safetensors",
230
+ "model.layers.2.mlp.up_proj.weight_scale": "model-00001-of-00004.safetensors",
231
+ "model.layers.2.post_attention_layernorm.weight": "model-00001-of-00004.safetensors",
232
+ "model.layers.2.self_attn.k_norm.weight": "model-00001-of-00004.safetensors",
233
+ "model.layers.2.self_attn.k_proj.weight": "model-00001-of-00004.safetensors",
234
+ "model.layers.2.self_attn.k_proj.weight_scale": "model-00001-of-00004.safetensors",
235
+ "model.layers.2.self_attn.o_proj.weight": "model-00001-of-00004.safetensors",
236
+ "model.layers.2.self_attn.o_proj.weight_scale": "model-00001-of-00004.safetensors",
237
+ "model.layers.2.self_attn.q_norm.weight": "model-00001-of-00004.safetensors",
238
+ "model.layers.2.self_attn.q_proj.weight": "model-00001-of-00004.safetensors",
239
+ "model.layers.2.self_attn.q_proj.weight_scale": "model-00001-of-00004.safetensors",
240
+ "model.layers.2.self_attn.v_proj.weight": "model-00001-of-00004.safetensors",
241
+ "model.layers.2.self_attn.v_proj.weight_scale": "model-00001-of-00004.safetensors",
242
+ "model.layers.20.input_layernorm.weight": "model-00002-of-00004.safetensors",
243
+ "model.layers.20.mlp.down_proj.weight": "model-00002-of-00004.safetensors",
244
+ "model.layers.20.mlp.down_proj.weight_scale": "model-00002-of-00004.safetensors",
245
+ "model.layers.20.mlp.gate_proj.weight": "model-00002-of-00004.safetensors",
246
+ "model.layers.20.mlp.gate_proj.weight_scale": "model-00002-of-00004.safetensors",
247
+ "model.layers.20.mlp.up_proj.weight": "model-00002-of-00004.safetensors",
248
+ "model.layers.20.mlp.up_proj.weight_scale": "model-00002-of-00004.safetensors",
249
+ "model.layers.20.post_attention_layernorm.weight": "model-00002-of-00004.safetensors",
250
+ "model.layers.20.self_attn.k_norm.weight": "model-00002-of-00004.safetensors",
251
+ "model.layers.20.self_attn.k_proj.weight": "model-00002-of-00004.safetensors",
252
+ "model.layers.20.self_attn.k_proj.weight_scale": "model-00002-of-00004.safetensors",
253
+ "model.layers.20.self_attn.o_proj.weight": "model-00002-of-00004.safetensors",
254
+ "model.layers.20.self_attn.o_proj.weight_scale": "model-00002-of-00004.safetensors",
255
+ "model.layers.20.self_attn.q_norm.weight": "model-00002-of-00004.safetensors",
256
+ "model.layers.20.self_attn.q_proj.weight": "model-00002-of-00004.safetensors",
257
+ "model.layers.20.self_attn.q_proj.weight_scale": "model-00002-of-00004.safetensors",
258
+ "model.layers.20.self_attn.v_proj.weight": "model-00002-of-00004.safetensors",
259
+ "model.layers.20.self_attn.v_proj.weight_scale": "model-00002-of-00004.safetensors",
260
+ "model.layers.21.input_layernorm.weight": "model-00002-of-00004.safetensors",
261
+ "model.layers.21.mlp.down_proj.weight": "model-00002-of-00004.safetensors",
262
+ "model.layers.21.mlp.down_proj.weight_scale": "model-00002-of-00004.safetensors",
263
+ "model.layers.21.mlp.gate_proj.weight": "model-00002-of-00004.safetensors",
264
+ "model.layers.21.mlp.gate_proj.weight_scale": "model-00002-of-00004.safetensors",
265
+ "model.layers.21.mlp.up_proj.weight": "model-00002-of-00004.safetensors",
266
+ "model.layers.21.mlp.up_proj.weight_scale": "model-00002-of-00004.safetensors",
267
+ "model.layers.21.post_attention_layernorm.weight": "model-00002-of-00004.safetensors",
268
+ "model.layers.21.self_attn.k_norm.weight": "model-00002-of-00004.safetensors",
269
+ "model.layers.21.self_attn.k_proj.weight": "model-00002-of-00004.safetensors",
270
+ "model.layers.21.self_attn.k_proj.weight_scale": "model-00002-of-00004.safetensors",
271
+ "model.layers.21.self_attn.o_proj.weight": "model-00002-of-00004.safetensors",
272
+ "model.layers.21.self_attn.o_proj.weight_scale": "model-00002-of-00004.safetensors",
273
+ "model.layers.21.self_attn.q_norm.weight": "model-00002-of-00004.safetensors",
274
+ "model.layers.21.self_attn.q_proj.weight": "model-00002-of-00004.safetensors",
275
+ "model.layers.21.self_attn.q_proj.weight_scale": "model-00002-of-00004.safetensors",
276
+ "model.layers.21.self_attn.v_proj.weight": "model-00002-of-00004.safetensors",
277
+ "model.layers.21.self_attn.v_proj.weight_scale": "model-00002-of-00004.safetensors",
278
+ "model.layers.22.input_layernorm.weight": "model-00002-of-00004.safetensors",
279
+ "model.layers.22.mlp.down_proj.weight": "model-00002-of-00004.safetensors",
280
+ "model.layers.22.mlp.down_proj.weight_scale": "model-00002-of-00004.safetensors",
281
+ "model.layers.22.mlp.gate_proj.weight": "model-00002-of-00004.safetensors",
282
+ "model.layers.22.mlp.gate_proj.weight_scale": "model-00002-of-00004.safetensors",
283
+ "model.layers.22.mlp.up_proj.weight": "model-00002-of-00004.safetensors",
284
+ "model.layers.22.mlp.up_proj.weight_scale": "model-00002-of-00004.safetensors",
285
+ "model.layers.22.post_attention_layernorm.weight": "model-00002-of-00004.safetensors",
286
+ "model.layers.22.self_attn.k_norm.weight": "model-00002-of-00004.safetensors",
287
+ "model.layers.22.self_attn.k_proj.weight": "model-00002-of-00004.safetensors",
288
+ "model.layers.22.self_attn.k_proj.weight_scale": "model-00002-of-00004.safetensors",
289
+ "model.layers.22.self_attn.o_proj.weight": "model-00002-of-00004.safetensors",
290
+ "model.layers.22.self_attn.o_proj.weight_scale": "model-00002-of-00004.safetensors",
291
+ "model.layers.22.self_attn.q_norm.weight": "model-00002-of-00004.safetensors",
292
+ "model.layers.22.self_attn.q_proj.weight": "model-00002-of-00004.safetensors",
293
+ "model.layers.22.self_attn.q_proj.weight_scale": "model-00002-of-00004.safetensors",
294
+ "model.layers.22.self_attn.v_proj.weight": "model-00002-of-00004.safetensors",
295
+ "model.layers.22.self_attn.v_proj.weight_scale": "model-00002-of-00004.safetensors",
296
+ "model.layers.23.input_layernorm.weight": "model-00002-of-00004.safetensors",
297
+ "model.layers.23.mlp.down_proj.weight": "model-00002-of-00004.safetensors",
298
+ "model.layers.23.mlp.down_proj.weight_scale": "model-00002-of-00004.safetensors",
299
+ "model.layers.23.mlp.gate_proj.weight": "model-00002-of-00004.safetensors",
300
+ "model.layers.23.mlp.gate_proj.weight_scale": "model-00002-of-00004.safetensors",
301
+ "model.layers.23.mlp.up_proj.weight": "model-00002-of-00004.safetensors",
302
+ "model.layers.23.mlp.up_proj.weight_scale": "model-00002-of-00004.safetensors",
303
+ "model.layers.23.post_attention_layernorm.weight": "model-00002-of-00004.safetensors",
304
+ "model.layers.23.self_attn.k_norm.weight": "model-00002-of-00004.safetensors",
305
+ "model.layers.23.self_attn.k_proj.weight": "model-00002-of-00004.safetensors",
306
+ "model.layers.23.self_attn.k_proj.weight_scale": "model-00002-of-00004.safetensors",
307
+ "model.layers.23.self_attn.o_proj.weight": "model-00002-of-00004.safetensors",
308
+ "model.layers.23.self_attn.o_proj.weight_scale": "model-00002-of-00004.safetensors",
309
+ "model.layers.23.self_attn.q_norm.weight": "model-00002-of-00004.safetensors",
310
+ "model.layers.23.self_attn.q_proj.weight": "model-00002-of-00004.safetensors",
311
+ "model.layers.23.self_attn.q_proj.weight_scale": "model-00002-of-00004.safetensors",
312
+ "model.layers.23.self_attn.v_proj.weight": "model-00002-of-00004.safetensors",
313
+ "model.layers.23.self_attn.v_proj.weight_scale": "model-00002-of-00004.safetensors",
314
+ "model.layers.24.input_layernorm.weight": "model-00002-of-00004.safetensors",
315
+ "model.layers.24.mlp.down_proj.weight": "model-00002-of-00004.safetensors",
316
+ "model.layers.24.mlp.down_proj.weight_scale": "model-00002-of-00004.safetensors",
317
+ "model.layers.24.mlp.gate_proj.weight": "model-00002-of-00004.safetensors",
318
+ "model.layers.24.mlp.gate_proj.weight_scale": "model-00002-of-00004.safetensors",
319
+ "model.layers.24.mlp.up_proj.weight": "model-00002-of-00004.safetensors",
320
+ "model.layers.24.mlp.up_proj.weight_scale": "model-00002-of-00004.safetensors",
321
+ "model.layers.24.post_attention_layernorm.weight": "model-00002-of-00004.safetensors",
322
+ "model.layers.24.self_attn.k_norm.weight": "model-00002-of-00004.safetensors",
323
+ "model.layers.24.self_attn.k_proj.weight": "model-00002-of-00004.safetensors",
324
+ "model.layers.24.self_attn.k_proj.weight_scale": "model-00002-of-00004.safetensors",
325
+ "model.layers.24.self_attn.o_proj.weight": "model-00002-of-00004.safetensors",
326
+ "model.layers.24.self_attn.o_proj.weight_scale": "model-00002-of-00004.safetensors",
327
+ "model.layers.24.self_attn.q_norm.weight": "model-00002-of-00004.safetensors",
328
+ "model.layers.24.self_attn.q_proj.weight": "model-00002-of-00004.safetensors",
329
+ "model.layers.24.self_attn.q_proj.weight_scale": "model-00002-of-00004.safetensors",
330
+ "model.layers.24.self_attn.v_proj.weight": "model-00002-of-00004.safetensors",
331
+ "model.layers.24.self_attn.v_proj.weight_scale": "model-00002-of-00004.safetensors",
332
+ "model.layers.25.input_layernorm.weight": "model-00003-of-00004.safetensors",
333
+ "model.layers.25.mlp.down_proj.weight": "model-00003-of-00004.safetensors",
334
+ "model.layers.25.mlp.down_proj.weight_scale": "model-00003-of-00004.safetensors",
335
+ "model.layers.25.mlp.gate_proj.weight": "model-00003-of-00004.safetensors",
336
+ "model.layers.25.mlp.gate_proj.weight_scale": "model-00003-of-00004.safetensors",
337
+ "model.layers.25.mlp.up_proj.weight": "model-00003-of-00004.safetensors",
338
+ "model.layers.25.mlp.up_proj.weight_scale": "model-00003-of-00004.safetensors",
339
+ "model.layers.25.post_attention_layernorm.weight": "model-00003-of-00004.safetensors",
340
+ "model.layers.25.self_attn.k_norm.weight": "model-00002-of-00004.safetensors",
341
+ "model.layers.25.self_attn.k_proj.weight": "model-00002-of-00004.safetensors",
342
+ "model.layers.25.self_attn.k_proj.weight_scale": "model-00002-of-00004.safetensors",
343
+ "model.layers.25.self_attn.o_proj.weight": "model-00002-of-00004.safetensors",
344
+ "model.layers.25.self_attn.o_proj.weight_scale": "model-00002-of-00004.safetensors",
345
+ "model.layers.25.self_attn.q_norm.weight": "model-00002-of-00004.safetensors",
346
+ "model.layers.25.self_attn.q_proj.weight": "model-00002-of-00004.safetensors",
347
+ "model.layers.25.self_attn.q_proj.weight_scale": "model-00002-of-00004.safetensors",
348
+ "model.layers.25.self_attn.v_proj.weight": "model-00002-of-00004.safetensors",
349
+ "model.layers.25.self_attn.v_proj.weight_scale": "model-00002-of-00004.safetensors",
350
+ "model.layers.26.input_layernorm.weight": "model-00003-of-00004.safetensors",
351
+ "model.layers.26.mlp.down_proj.weight": "model-00003-of-00004.safetensors",
352
+ "model.layers.26.mlp.down_proj.weight_scale": "model-00003-of-00004.safetensors",
353
+ "model.layers.26.mlp.gate_proj.weight": "model-00003-of-00004.safetensors",
354
+ "model.layers.26.mlp.gate_proj.weight_scale": "model-00003-of-00004.safetensors",
355
+ "model.layers.26.mlp.up_proj.weight": "model-00003-of-00004.safetensors",
356
+ "model.layers.26.mlp.up_proj.weight_scale": "model-00003-of-00004.safetensors",
357
+ "model.layers.26.post_attention_layernorm.weight": "model-00003-of-00004.safetensors",
358
+ "model.layers.26.self_attn.k_norm.weight": "model-00003-of-00004.safetensors",
359
+ "model.layers.26.self_attn.k_proj.weight": "model-00003-of-00004.safetensors",
360
+ "model.layers.26.self_attn.k_proj.weight_scale": "model-00003-of-00004.safetensors",
361
+ "model.layers.26.self_attn.o_proj.weight": "model-00003-of-00004.safetensors",
362
+ "model.layers.26.self_attn.o_proj.weight_scale": "model-00003-of-00004.safetensors",
363
+ "model.layers.26.self_attn.q_norm.weight": "model-00003-of-00004.safetensors",
364
+ "model.layers.26.self_attn.q_proj.weight": "model-00003-of-00004.safetensors",
365
+ "model.layers.26.self_attn.q_proj.weight_scale": "model-00003-of-00004.safetensors",
366
+ "model.layers.26.self_attn.v_proj.weight": "model-00003-of-00004.safetensors",
367
+ "model.layers.26.self_attn.v_proj.weight_scale": "model-00003-of-00004.safetensors",
368
+ "model.layers.27.input_layernorm.weight": "model-00003-of-00004.safetensors",
369
+ "model.layers.27.mlp.down_proj.weight": "model-00003-of-00004.safetensors",
370
+ "model.layers.27.mlp.down_proj.weight_scale": "model-00003-of-00004.safetensors",
371
+ "model.layers.27.mlp.gate_proj.weight": "model-00003-of-00004.safetensors",
372
+ "model.layers.27.mlp.gate_proj.weight_scale": "model-00003-of-00004.safetensors",
373
+ "model.layers.27.mlp.up_proj.weight": "model-00003-of-00004.safetensors",
374
+ "model.layers.27.mlp.up_proj.weight_scale": "model-00003-of-00004.safetensors",
375
+ "model.layers.27.post_attention_layernorm.weight": "model-00003-of-00004.safetensors",
376
+ "model.layers.27.self_attn.k_norm.weight": "model-00003-of-00004.safetensors",
377
+ "model.layers.27.self_attn.k_proj.weight": "model-00003-of-00004.safetensors",
378
+ "model.layers.27.self_attn.k_proj.weight_scale": "model-00003-of-00004.safetensors",
379
+ "model.layers.27.self_attn.o_proj.weight": "model-00003-of-00004.safetensors",
380
+ "model.layers.27.self_attn.o_proj.weight_scale": "model-00003-of-00004.safetensors",
381
+ "model.layers.27.self_attn.q_norm.weight": "model-00003-of-00004.safetensors",
382
+ "model.layers.27.self_attn.q_proj.weight": "model-00003-of-00004.safetensors",
383
+ "model.layers.27.self_attn.q_proj.weight_scale": "model-00003-of-00004.safetensors",
384
+ "model.layers.27.self_attn.v_proj.weight": "model-00003-of-00004.safetensors",
385
+ "model.layers.27.self_attn.v_proj.weight_scale": "model-00003-of-00004.safetensors",
386
+ "model.layers.28.input_layernorm.weight": "model-00003-of-00004.safetensors",
387
+ "model.layers.28.mlp.down_proj.weight": "model-00003-of-00004.safetensors",
388
+ "model.layers.28.mlp.down_proj.weight_scale": "model-00003-of-00004.safetensors",
389
+ "model.layers.28.mlp.gate_proj.weight": "model-00003-of-00004.safetensors",
390
+ "model.layers.28.mlp.gate_proj.weight_scale": "model-00003-of-00004.safetensors",
391
+ "model.layers.28.mlp.up_proj.weight": "model-00003-of-00004.safetensors",
392
+ "model.layers.28.mlp.up_proj.weight_scale": "model-00003-of-00004.safetensors",
393
+ "model.layers.28.post_attention_layernorm.weight": "model-00003-of-00004.safetensors",
394
+ "model.layers.28.self_attn.k_norm.weight": "model-00003-of-00004.safetensors",
395
+ "model.layers.28.self_attn.k_proj.weight": "model-00003-of-00004.safetensors",
396
+ "model.layers.28.self_attn.k_proj.weight_scale": "model-00003-of-00004.safetensors",
397
+ "model.layers.28.self_attn.o_proj.weight": "model-00003-of-00004.safetensors",
398
+ "model.layers.28.self_attn.o_proj.weight_scale": "model-00003-of-00004.safetensors",
399
+ "model.layers.28.self_attn.q_norm.weight": "model-00003-of-00004.safetensors",
400
+ "model.layers.28.self_attn.q_proj.weight": "model-00003-of-00004.safetensors",
401
+ "model.layers.28.self_attn.q_proj.weight_scale": "model-00003-of-00004.safetensors",
402
+ "model.layers.28.self_attn.v_proj.weight": "model-00003-of-00004.safetensors",
403
+ "model.layers.28.self_attn.v_proj.weight_scale": "model-00003-of-00004.safetensors",
404
+ "model.layers.29.input_layernorm.weight": "model-00003-of-00004.safetensors",
405
+ "model.layers.29.mlp.down_proj.weight": "model-00003-of-00004.safetensors",
406
+ "model.layers.29.mlp.down_proj.weight_scale": "model-00003-of-00004.safetensors",
407
+ "model.layers.29.mlp.gate_proj.weight": "model-00003-of-00004.safetensors",
408
+ "model.layers.29.mlp.gate_proj.weight_scale": "model-00003-of-00004.safetensors",
409
+ "model.layers.29.mlp.up_proj.weight": "model-00003-of-00004.safetensors",
410
+ "model.layers.29.mlp.up_proj.weight_scale": "model-00003-of-00004.safetensors",
411
+ "model.layers.29.post_attention_layernorm.weight": "model-00003-of-00004.safetensors",
412
+ "model.layers.29.self_attn.k_norm.weight": "model-00003-of-00004.safetensors",
413
+ "model.layers.29.self_attn.k_proj.weight": "model-00003-of-00004.safetensors",
414
+ "model.layers.29.self_attn.k_proj.weight_scale": "model-00003-of-00004.safetensors",
415
+ "model.layers.29.self_attn.o_proj.weight": "model-00003-of-00004.safetensors",
416
+ "model.layers.29.self_attn.o_proj.weight_scale": "model-00003-of-00004.safetensors",
417
+ "model.layers.29.self_attn.q_norm.weight": "model-00003-of-00004.safetensors",
418
+ "model.layers.29.self_attn.q_proj.weight": "model-00003-of-00004.safetensors",
419
+ "model.layers.29.self_attn.q_proj.weight_scale": "model-00003-of-00004.safetensors",
420
+ "model.layers.29.self_attn.v_proj.weight": "model-00003-of-00004.safetensors",
421
+ "model.layers.29.self_attn.v_proj.weight_scale": "model-00003-of-00004.safetensors",
422
+ "model.layers.3.input_layernorm.weight": "model-00001-of-00004.safetensors",
423
+ "model.layers.3.mlp.down_proj.weight": "model-00001-of-00004.safetensors",
424
+ "model.layers.3.mlp.down_proj.weight_scale": "model-00001-of-00004.safetensors",
425
+ "model.layers.3.mlp.gate_proj.weight": "model-00001-of-00004.safetensors",
426
+ "model.layers.3.mlp.gate_proj.weight_scale": "model-00001-of-00004.safetensors",
427
+ "model.layers.3.mlp.up_proj.weight": "model-00001-of-00004.safetensors",
428
+ "model.layers.3.mlp.up_proj.weight_scale": "model-00001-of-00004.safetensors",
429
+ "model.layers.3.post_attention_layernorm.weight": "model-00001-of-00004.safetensors",
430
+ "model.layers.3.self_attn.k_norm.weight": "model-00001-of-00004.safetensors",
431
+ "model.layers.3.self_attn.k_proj.weight": "model-00001-of-00004.safetensors",
432
+ "model.layers.3.self_attn.k_proj.weight_scale": "model-00001-of-00004.safetensors",
433
+ "model.layers.3.self_attn.o_proj.weight": "model-00001-of-00004.safetensors",
434
+ "model.layers.3.self_attn.o_proj.weight_scale": "model-00001-of-00004.safetensors",
435
+ "model.layers.3.self_attn.q_norm.weight": "model-00001-of-00004.safetensors",
436
+ "model.layers.3.self_attn.q_proj.weight": "model-00001-of-00004.safetensors",
437
+ "model.layers.3.self_attn.q_proj.weight_scale": "model-00001-of-00004.safetensors",
438
+ "model.layers.3.self_attn.v_proj.weight": "model-00001-of-00004.safetensors",
439
+ "model.layers.3.self_attn.v_proj.weight_scale": "model-00001-of-00004.safetensors",
440
+ "model.layers.30.input_layernorm.weight": "model-00003-of-00004.safetensors",
441
+ "model.layers.30.mlp.down_proj.weight": "model-00003-of-00004.safetensors",
442
+ "model.layers.30.mlp.down_proj.weight_scale": "model-00003-of-00004.safetensors",
443
+ "model.layers.30.mlp.gate_proj.weight": "model-00003-of-00004.safetensors",
444
+ "model.layers.30.mlp.gate_proj.weight_scale": "model-00003-of-00004.safetensors",
445
+ "model.layers.30.mlp.up_proj.weight": "model-00003-of-00004.safetensors",
446
+ "model.layers.30.mlp.up_proj.weight_scale": "model-00003-of-00004.safetensors",
447
+ "model.layers.30.post_attention_layernorm.weight": "model-00003-of-00004.safetensors",
448
+ "model.layers.30.self_attn.k_norm.weight": "model-00003-of-00004.safetensors",
449
+ "model.layers.30.self_attn.k_proj.weight": "model-00003-of-00004.safetensors",
450
+ "model.layers.30.self_attn.k_proj.weight_scale": "model-00003-of-00004.safetensors",
451
+ "model.layers.30.self_attn.o_proj.weight": "model-00003-of-00004.safetensors",
452
+ "model.layers.30.self_attn.o_proj.weight_scale": "model-00003-of-00004.safetensors",
453
+ "model.layers.30.self_attn.q_norm.weight": "model-00003-of-00004.safetensors",
454
+ "model.layers.30.self_attn.q_proj.weight": "model-00003-of-00004.safetensors",
455
+ "model.layers.30.self_attn.q_proj.weight_scale": "model-00003-of-00004.safetensors",
456
+ "model.layers.30.self_attn.v_proj.weight": "model-00003-of-00004.safetensors",
457
+ "model.layers.30.self_attn.v_proj.weight_scale": "model-00003-of-00004.safetensors",
458
+ "model.layers.31.input_layernorm.weight": "model-00003-of-00004.safetensors",
459
+ "model.layers.31.mlp.down_proj.weight": "model-00003-of-00004.safetensors",
460
+ "model.layers.31.mlp.down_proj.weight_scale": "model-00003-of-00004.safetensors",
461
+ "model.layers.31.mlp.gate_proj.weight": "model-00003-of-00004.safetensors",
462
+ "model.layers.31.mlp.gate_proj.weight_scale": "model-00003-of-00004.safetensors",
463
+ "model.layers.31.mlp.up_proj.weight": "model-00003-of-00004.safetensors",
464
+ "model.layers.31.mlp.up_proj.weight_scale": "model-00003-of-00004.safetensors",
465
+ "model.layers.31.post_attention_layernorm.weight": "model-00003-of-00004.safetensors",
466
+ "model.layers.31.self_attn.k_norm.weight": "model-00003-of-00004.safetensors",
467
+ "model.layers.31.self_attn.k_proj.weight": "model-00003-of-00004.safetensors",
468
+ "model.layers.31.self_attn.k_proj.weight_scale": "model-00003-of-00004.safetensors",
469
+ "model.layers.31.self_attn.o_proj.weight": "model-00003-of-00004.safetensors",
470
+ "model.layers.31.self_attn.o_proj.weight_scale": "model-00003-of-00004.safetensors",
471
+ "model.layers.31.self_attn.q_norm.weight": "model-00003-of-00004.safetensors",
472
+ "model.layers.31.self_attn.q_proj.weight": "model-00003-of-00004.safetensors",
473
+ "model.layers.31.self_attn.q_proj.weight_scale": "model-00003-of-00004.safetensors",
474
+ "model.layers.31.self_attn.v_proj.weight": "model-00003-of-00004.safetensors",
475
+ "model.layers.31.self_attn.v_proj.weight_scale": "model-00003-of-00004.safetensors",
476
+ "model.layers.32.input_layernorm.weight": "model-00003-of-00004.safetensors",
477
+ "model.layers.32.mlp.down_proj.weight": "model-00003-of-00004.safetensors",
478
+ "model.layers.32.mlp.down_proj.weight_scale": "model-00003-of-00004.safetensors",
479
+ "model.layers.32.mlp.gate_proj.weight": "model-00003-of-00004.safetensors",
480
+ "model.layers.32.mlp.gate_proj.weight_scale": "model-00003-of-00004.safetensors",
481
+ "model.layers.32.mlp.up_proj.weight": "model-00003-of-00004.safetensors",
482
+ "model.layers.32.mlp.up_proj.weight_scale": "model-00003-of-00004.safetensors",
483
+ "model.layers.32.post_attention_layernorm.weight": "model-00003-of-00004.safetensors",
484
+ "model.layers.32.self_attn.k_norm.weight": "model-00003-of-00004.safetensors",
485
+ "model.layers.32.self_attn.k_proj.weight": "model-00003-of-00004.safetensors",
486
+ "model.layers.32.self_attn.k_proj.weight_scale": "model-00003-of-00004.safetensors",
487
+ "model.layers.32.self_attn.o_proj.weight": "model-00003-of-00004.safetensors",
488
+ "model.layers.32.self_attn.o_proj.weight_scale": "model-00003-of-00004.safetensors",
489
+ "model.layers.32.self_attn.q_norm.weight": "model-00003-of-00004.safetensors",
490
+ "model.layers.32.self_attn.q_proj.weight": "model-00003-of-00004.safetensors",
491
+ "model.layers.32.self_attn.q_proj.weight_scale": "model-00003-of-00004.safetensors",
492
+ "model.layers.32.self_attn.v_proj.weight": "model-00003-of-00004.safetensors",
493
+ "model.layers.32.self_attn.v_proj.weight_scale": "model-00003-of-00004.safetensors",
494
+ "model.layers.33.input_layernorm.weight": "model-00003-of-00004.safetensors",
495
+ "model.layers.33.mlp.down_proj.weight": "model-00003-of-00004.safetensors",
496
+ "model.layers.33.mlp.down_proj.weight_scale": "model-00003-of-00004.safetensors",
497
+ "model.layers.33.mlp.gate_proj.weight": "model-00003-of-00004.safetensors",
498
+ "model.layers.33.mlp.gate_proj.weight_scale": "model-00003-of-00004.safetensors",
499
+ "model.layers.33.mlp.up_proj.weight": "model-00003-of-00004.safetensors",
500
+ "model.layers.33.mlp.up_proj.weight_scale": "model-00003-of-00004.safetensors",
501
+ "model.layers.33.post_attention_layernorm.weight": "model-00003-of-00004.safetensors",
502
+ "model.layers.33.self_attn.k_norm.weight": "model-00003-of-00004.safetensors",
503
+ "model.layers.33.self_attn.k_proj.weight": "model-00003-of-00004.safetensors",
504
+ "model.layers.33.self_attn.k_proj.weight_scale": "model-00003-of-00004.safetensors",
505
+ "model.layers.33.self_attn.o_proj.weight": "model-00003-of-00004.safetensors",
506
+ "model.layers.33.self_attn.o_proj.weight_scale": "model-00003-of-00004.safetensors",
507
+ "model.layers.33.self_attn.q_norm.weight": "model-00003-of-00004.safetensors",
508
+ "model.layers.33.self_attn.q_proj.weight": "model-00003-of-00004.safetensors",
509
+ "model.layers.33.self_attn.q_proj.weight_scale": "model-00003-of-00004.safetensors",
510
+ "model.layers.33.self_attn.v_proj.weight": "model-00003-of-00004.safetensors",
511
+ "model.layers.33.self_attn.v_proj.weight_scale": "model-00003-of-00004.safetensors",
512
+ "model.layers.34.input_layernorm.weight": "model-00003-of-00004.safetensors",
513
+ "model.layers.34.mlp.down_proj.weight": "model-00003-of-00004.safetensors",
514
+ "model.layers.34.mlp.down_proj.weight_scale": "model-00003-of-00004.safetensors",
515
+ "model.layers.34.mlp.gate_proj.weight": "model-00003-of-00004.safetensors",
516
+ "model.layers.34.mlp.gate_proj.weight_scale": "model-00003-of-00004.safetensors",
517
+ "model.layers.34.mlp.up_proj.weight": "model-00003-of-00004.safetensors",
518
+ "model.layers.34.mlp.up_proj.weight_scale": "model-00003-of-00004.safetensors",
519
+ "model.layers.34.post_attention_layernorm.weight": "model-00003-of-00004.safetensors",
520
+ "model.layers.34.self_attn.k_norm.weight": "model-00003-of-00004.safetensors",
521
+ "model.layers.34.self_attn.k_proj.weight": "model-00003-of-00004.safetensors",
522
+ "model.layers.34.self_attn.k_proj.weight_scale": "model-00003-of-00004.safetensors",
523
+ "model.layers.34.self_attn.o_proj.weight": "model-00003-of-00004.safetensors",
524
+ "model.layers.34.self_attn.o_proj.weight_scale": "model-00003-of-00004.safetensors",
525
+ "model.layers.34.self_attn.q_norm.weight": "model-00003-of-00004.safetensors",
526
+ "model.layers.34.self_attn.q_proj.weight": "model-00003-of-00004.safetensors",
527
+ "model.layers.34.self_attn.q_proj.weight_scale": "model-00003-of-00004.safetensors",
528
+ "model.layers.34.self_attn.v_proj.weight": "model-00003-of-00004.safetensors",
529
+ "model.layers.34.self_attn.v_proj.weight_scale": "model-00003-of-00004.safetensors",
530
+ "model.layers.35.input_layernorm.weight": "model-00003-of-00004.safetensors",
531
+ "model.layers.35.mlp.down_proj.weight": "model-00003-of-00004.safetensors",
532
+ "model.layers.35.mlp.down_proj.weight_scale": "model-00003-of-00004.safetensors",
533
+ "model.layers.35.mlp.gate_proj.weight": "model-00003-of-00004.safetensors",
534
+ "model.layers.35.mlp.gate_proj.weight_scale": "model-00003-of-00004.safetensors",
535
+ "model.layers.35.mlp.up_proj.weight": "model-00003-of-00004.safetensors",
536
+ "model.layers.35.mlp.up_proj.weight_scale": "model-00003-of-00004.safetensors",
537
+ "model.layers.35.post_attention_layernorm.weight": "model-00003-of-00004.safetensors",
538
+ "model.layers.35.self_attn.k_norm.weight": "model-00003-of-00004.safetensors",
539
+ "model.layers.35.self_attn.k_proj.weight": "model-00003-of-00004.safetensors",
540
+ "model.layers.35.self_attn.k_proj.weight_scale": "model-00003-of-00004.safetensors",
541
+ "model.layers.35.self_attn.o_proj.weight": "model-00003-of-00004.safetensors",
542
+ "model.layers.35.self_attn.o_proj.weight_scale": "model-00003-of-00004.safetensors",
543
+ "model.layers.35.self_attn.q_norm.weight": "model-00003-of-00004.safetensors",
544
+ "model.layers.35.self_attn.q_proj.weight": "model-00003-of-00004.safetensors",
545
+ "model.layers.35.self_attn.q_proj.weight_scale": "model-00003-of-00004.safetensors",
546
+ "model.layers.35.self_attn.v_proj.weight": "model-00003-of-00004.safetensors",
547
+ "model.layers.35.self_attn.v_proj.weight_scale": "model-00003-of-00004.safetensors",
548
+ "model.layers.36.input_layernorm.weight": "model-00003-of-00004.safetensors",
549
+ "model.layers.36.mlp.down_proj.weight": "model-00003-of-00004.safetensors",
550
+ "model.layers.36.mlp.down_proj.weight_scale": "model-00003-of-00004.safetensors",
551
+ "model.layers.36.mlp.gate_proj.weight": "model-00003-of-00004.safetensors",
552
+ "model.layers.36.mlp.gate_proj.weight_scale": "model-00003-of-00004.safetensors",
553
+ "model.layers.36.mlp.up_proj.weight": "model-00003-of-00004.safetensors",
554
+ "model.layers.36.mlp.up_proj.weight_scale": "model-00003-of-00004.safetensors",
555
+ "model.layers.36.post_attention_layernorm.weight": "model-00003-of-00004.safetensors",
556
+ "model.layers.36.self_attn.k_norm.weight": "model-00003-of-00004.safetensors",
557
+ "model.layers.36.self_attn.k_proj.weight": "model-00003-of-00004.safetensors",
558
+ "model.layers.36.self_attn.k_proj.weight_scale": "model-00003-of-00004.safetensors",
559
+ "model.layers.36.self_attn.o_proj.weight": "model-00003-of-00004.safetensors",
560
+ "model.layers.36.self_attn.o_proj.weight_scale": "model-00003-of-00004.safetensors",
561
+ "model.layers.36.self_attn.q_norm.weight": "model-00003-of-00004.safetensors",
562
+ "model.layers.36.self_attn.q_proj.weight": "model-00003-of-00004.safetensors",
563
+ "model.layers.36.self_attn.q_proj.weight_scale": "model-00003-of-00004.safetensors",
564
+ "model.layers.36.self_attn.v_proj.weight": "model-00003-of-00004.safetensors",
565
+ "model.layers.36.self_attn.v_proj.weight_scale": "model-00003-of-00004.safetensors",
566
+ "model.layers.37.input_layernorm.weight": "model-00003-of-00004.safetensors",
567
+ "model.layers.37.mlp.down_proj.weight": "model-00003-of-00004.safetensors",
568
+ "model.layers.37.mlp.down_proj.weight_scale": "model-00003-of-00004.safetensors",
569
+ "model.layers.37.mlp.gate_proj.weight": "model-00003-of-00004.safetensors",
570
+ "model.layers.37.mlp.gate_proj.weight_scale": "model-00003-of-00004.safetensors",
571
+ "model.layers.37.mlp.up_proj.weight": "model-00003-of-00004.safetensors",
572
+ "model.layers.37.mlp.up_proj.weight_scale": "model-00003-of-00004.safetensors",
573
+ "model.layers.37.post_attention_layernorm.weight": "model-00003-of-00004.safetensors",
574
+ "model.layers.37.self_attn.k_norm.weight": "model-00003-of-00004.safetensors",
575
+ "model.layers.37.self_attn.k_proj.weight": "model-00003-of-00004.safetensors",
576
+ "model.layers.37.self_attn.k_proj.weight_scale": "model-00003-of-00004.safetensors",
577
+ "model.layers.37.self_attn.o_proj.weight": "model-00003-of-00004.safetensors",
578
+ "model.layers.37.self_attn.o_proj.weight_scale": "model-00003-of-00004.safetensors",
579
+ "model.layers.37.self_attn.q_norm.weight": "model-00003-of-00004.safetensors",
580
+ "model.layers.37.self_attn.q_proj.weight": "model-00003-of-00004.safetensors",
581
+ "model.layers.37.self_attn.q_proj.weight_scale": "model-00003-of-00004.safetensors",
582
+ "model.layers.37.self_attn.v_proj.weight": "model-00003-of-00004.safetensors",
583
+ "model.layers.37.self_attn.v_proj.weight_scale": "model-00003-of-00004.safetensors",
584
+ "model.layers.38.input_layernorm.weight": "model-00003-of-00004.safetensors",
585
+ "model.layers.38.mlp.down_proj.weight": "model-00003-of-00004.safetensors",
586
+ "model.layers.38.mlp.down_proj.weight_scale": "model-00003-of-00004.safetensors",
587
+ "model.layers.38.mlp.gate_proj.weight": "model-00003-of-00004.safetensors",
588
+ "model.layers.38.mlp.gate_proj.weight_scale": "model-00003-of-00004.safetensors",
589
+ "model.layers.38.mlp.up_proj.weight": "model-00003-of-00004.safetensors",
590
+ "model.layers.38.mlp.up_proj.weight_scale": "model-00003-of-00004.safetensors",
591
+ "model.layers.38.post_attention_layernorm.weight": "model-00003-of-00004.safetensors",
592
+ "model.layers.38.self_attn.k_norm.weight": "model-00003-of-00004.safetensors",
593
+ "model.layers.38.self_attn.k_proj.weight": "model-00003-of-00004.safetensors",
594
+ "model.layers.38.self_attn.k_proj.weight_scale": "model-00003-of-00004.safetensors",
595
+ "model.layers.38.self_attn.o_proj.weight": "model-00003-of-00004.safetensors",
596
+ "model.layers.38.self_attn.o_proj.weight_scale": "model-00003-of-00004.safetensors",
597
+ "model.layers.38.self_attn.q_norm.weight": "model-00003-of-00004.safetensors",
598
+ "model.layers.38.self_attn.q_proj.weight": "model-00003-of-00004.safetensors",
599
+ "model.layers.38.self_attn.q_proj.weight_scale": "model-00003-of-00004.safetensors",
600
+ "model.layers.38.self_attn.v_proj.weight": "model-00003-of-00004.safetensors",
601
+ "model.layers.38.self_attn.v_proj.weight_scale": "model-00003-of-00004.safetensors",
602
+ "model.layers.39.input_layernorm.weight": "model-00003-of-00004.safetensors",
603
+ "model.layers.39.mlp.down_proj.weight": "model-00003-of-00004.safetensors",
604
+ "model.layers.39.mlp.down_proj.weight_scale": "model-00003-of-00004.safetensors",
605
+ "model.layers.39.mlp.gate_proj.weight": "model-00003-of-00004.safetensors",
606
+ "model.layers.39.mlp.gate_proj.weight_scale": "model-00003-of-00004.safetensors",
607
+ "model.layers.39.mlp.up_proj.weight": "model-00003-of-00004.safetensors",
608
+ "model.layers.39.mlp.up_proj.weight_scale": "model-00003-of-00004.safetensors",
609
+ "model.layers.39.post_attention_layernorm.weight": "model-00003-of-00004.safetensors",
610
+ "model.layers.39.self_attn.k_norm.weight": "model-00003-of-00004.safetensors",
611
+ "model.layers.39.self_attn.k_proj.weight": "model-00003-of-00004.safetensors",
612
+ "model.layers.39.self_attn.k_proj.weight_scale": "model-00003-of-00004.safetensors",
613
+ "model.layers.39.self_attn.o_proj.weight": "model-00003-of-00004.safetensors",
614
+ "model.layers.39.self_attn.o_proj.weight_scale": "model-00003-of-00004.safetensors",
615
+ "model.layers.39.self_attn.q_norm.weight": "model-00003-of-00004.safetensors",
616
+ "model.layers.39.self_attn.q_proj.weight": "model-00003-of-00004.safetensors",
617
+ "model.layers.39.self_attn.q_proj.weight_scale": "model-00003-of-00004.safetensors",
618
+ "model.layers.39.self_attn.v_proj.weight": "model-00003-of-00004.safetensors",
619
+ "model.layers.39.self_attn.v_proj.weight_scale": "model-00003-of-00004.safetensors",
620
+ "model.layers.4.input_layernorm.weight": "model-00001-of-00004.safetensors",
621
+ "model.layers.4.mlp.down_proj.weight": "model-00001-of-00004.safetensors",
622
+ "model.layers.4.mlp.down_proj.weight_scale": "model-00001-of-00004.safetensors",
623
+ "model.layers.4.mlp.gate_proj.weight": "model-00001-of-00004.safetensors",
624
+ "model.layers.4.mlp.gate_proj.weight_scale": "model-00001-of-00004.safetensors",
625
+ "model.layers.4.mlp.up_proj.weight": "model-00001-of-00004.safetensors",
626
+ "model.layers.4.mlp.up_proj.weight_scale": "model-00001-of-00004.safetensors",
627
+ "model.layers.4.post_attention_layernorm.weight": "model-00001-of-00004.safetensors",
628
+ "model.layers.4.self_attn.k_norm.weight": "model-00001-of-00004.safetensors",
629
+ "model.layers.4.self_attn.k_proj.weight": "model-00001-of-00004.safetensors",
630
+ "model.layers.4.self_attn.k_proj.weight_scale": "model-00001-of-00004.safetensors",
631
+ "model.layers.4.self_attn.o_proj.weight": "model-00001-of-00004.safetensors",
632
+ "model.layers.4.self_attn.o_proj.weight_scale": "model-00001-of-00004.safetensors",
633
+ "model.layers.4.self_attn.q_norm.weight": "model-00001-of-00004.safetensors",
634
+ "model.layers.4.self_attn.q_proj.weight": "model-00001-of-00004.safetensors",
635
+ "model.layers.4.self_attn.q_proj.weight_scale": "model-00001-of-00004.safetensors",
636
+ "model.layers.4.self_attn.v_proj.weight": "model-00001-of-00004.safetensors",
637
+ "model.layers.4.self_attn.v_proj.weight_scale": "model-00001-of-00004.safetensors",
638
+ "model.layers.5.input_layernorm.weight": "model-00001-of-00004.safetensors",
639
+ "model.layers.5.mlp.down_proj.weight": "model-00001-of-00004.safetensors",
640
+ "model.layers.5.mlp.down_proj.weight_scale": "model-00001-of-00004.safetensors",
641
+ "model.layers.5.mlp.gate_proj.weight": "model-00001-of-00004.safetensors",
642
+ "model.layers.5.mlp.gate_proj.weight_scale": "model-00001-of-00004.safetensors",
643
+ "model.layers.5.mlp.up_proj.weight": "model-00001-of-00004.safetensors",
644
+ "model.layers.5.mlp.up_proj.weight_scale": "model-00001-of-00004.safetensors",
645
+ "model.layers.5.post_attention_layernorm.weight": "model-00001-of-00004.safetensors",
646
+ "model.layers.5.self_attn.k_norm.weight": "model-00001-of-00004.safetensors",
647
+ "model.layers.5.self_attn.k_proj.weight": "model-00001-of-00004.safetensors",
648
+ "model.layers.5.self_attn.k_proj.weight_scale": "model-00001-of-00004.safetensors",
649
+ "model.layers.5.self_attn.o_proj.weight": "model-00001-of-00004.safetensors",
650
+ "model.layers.5.self_attn.o_proj.weight_scale": "model-00001-of-00004.safetensors",
651
+ "model.layers.5.self_attn.q_norm.weight": "model-00001-of-00004.safetensors",
652
+ "model.layers.5.self_attn.q_proj.weight": "model-00001-of-00004.safetensors",
653
+ "model.layers.5.self_attn.q_proj.weight_scale": "model-00001-of-00004.safetensors",
654
+ "model.layers.5.self_attn.v_proj.weight": "model-00001-of-00004.safetensors",
655
+ "model.layers.5.self_attn.v_proj.weight_scale": "model-00001-of-00004.safetensors",
656
+ "model.layers.6.input_layernorm.weight": "model-00001-of-00004.safetensors",
657
+ "model.layers.6.mlp.down_proj.weight": "model-00001-of-00004.safetensors",
658
+ "model.layers.6.mlp.down_proj.weight_scale": "model-00001-of-00004.safetensors",
659
+ "model.layers.6.mlp.gate_proj.weight": "model-00001-of-00004.safetensors",
660
+ "model.layers.6.mlp.gate_proj.weight_scale": "model-00001-of-00004.safetensors",
661
+ "model.layers.6.mlp.up_proj.weight": "model-00001-of-00004.safetensors",
662
+ "model.layers.6.mlp.up_proj.weight_scale": "model-00001-of-00004.safetensors",
663
+ "model.layers.6.post_attention_layernorm.weight": "model-00001-of-00004.safetensors",
664
+ "model.layers.6.self_attn.k_norm.weight": "model-00001-of-00004.safetensors",
665
+ "model.layers.6.self_attn.k_proj.weight": "model-00001-of-00004.safetensors",
666
+ "model.layers.6.self_attn.k_proj.weight_scale": "model-00001-of-00004.safetensors",
667
+ "model.layers.6.self_attn.o_proj.weight": "model-00001-of-00004.safetensors",
668
+ "model.layers.6.self_attn.o_proj.weight_scale": "model-00001-of-00004.safetensors",
669
+ "model.layers.6.self_attn.q_norm.weight": "model-00001-of-00004.safetensors",
670
+ "model.layers.6.self_attn.q_proj.weight": "model-00001-of-00004.safetensors",
671
+ "model.layers.6.self_attn.q_proj.weight_scale": "model-00001-of-00004.safetensors",
672
+ "model.layers.6.self_attn.v_proj.weight": "model-00001-of-00004.safetensors",
673
+ "model.layers.6.self_attn.v_proj.weight_scale": "model-00001-of-00004.safetensors",
674
+ "model.layers.7.input_layernorm.weight": "model-00001-of-00004.safetensors",
675
+ "model.layers.7.mlp.down_proj.weight": "model-00001-of-00004.safetensors",
676
+ "model.layers.7.mlp.down_proj.weight_scale": "model-00001-of-00004.safetensors",
677
+ "model.layers.7.mlp.gate_proj.weight": "model-00001-of-00004.safetensors",
678
+ "model.layers.7.mlp.gate_proj.weight_scale": "model-00001-of-00004.safetensors",
679
+ "model.layers.7.mlp.up_proj.weight": "model-00001-of-00004.safetensors",
680
+ "model.layers.7.mlp.up_proj.weight_scale": "model-00001-of-00004.safetensors",
681
+ "model.layers.7.post_attention_layernorm.weight": "model-00001-of-00004.safetensors",
682
+ "model.layers.7.self_attn.k_norm.weight": "model-00001-of-00004.safetensors",
683
+ "model.layers.7.self_attn.k_proj.weight": "model-00001-of-00004.safetensors",
684
+ "model.layers.7.self_attn.k_proj.weight_scale": "model-00001-of-00004.safetensors",
685
+ "model.layers.7.self_attn.o_proj.weight": "model-00001-of-00004.safetensors",
686
+ "model.layers.7.self_attn.o_proj.weight_scale": "model-00001-of-00004.safetensors",
687
+ "model.layers.7.self_attn.q_norm.weight": "model-00001-of-00004.safetensors",
688
+ "model.layers.7.self_attn.q_proj.weight": "model-00001-of-00004.safetensors",
689
+ "model.layers.7.self_attn.q_proj.weight_scale": "model-00001-of-00004.safetensors",
690
+ "model.layers.7.self_attn.v_proj.weight": "model-00001-of-00004.safetensors",
691
+ "model.layers.7.self_attn.v_proj.weight_scale": "model-00001-of-00004.safetensors",
692
+ "model.layers.8.input_layernorm.weight": "model-00001-of-00004.safetensors",
693
+ "model.layers.8.mlp.down_proj.weight": "model-00001-of-00004.safetensors",
694
+ "model.layers.8.mlp.down_proj.weight_scale": "model-00001-of-00004.safetensors",
695
+ "model.layers.8.mlp.gate_proj.weight": "model-00001-of-00004.safetensors",
696
+ "model.layers.8.mlp.gate_proj.weight_scale": "model-00001-of-00004.safetensors",
697
+ "model.layers.8.mlp.up_proj.weight": "model-00001-of-00004.safetensors",
698
+ "model.layers.8.mlp.up_proj.weight_scale": "model-00001-of-00004.safetensors",
699
+ "model.layers.8.post_attention_layernorm.weight": "model-00001-of-00004.safetensors",
700
+ "model.layers.8.self_attn.k_norm.weight": "model-00001-of-00004.safetensors",
701
+ "model.layers.8.self_attn.k_proj.weight": "model-00001-of-00004.safetensors",
702
+ "model.layers.8.self_attn.k_proj.weight_scale": "model-00001-of-00004.safetensors",
703
+ "model.layers.8.self_attn.o_proj.weight": "model-00001-of-00004.safetensors",
704
+ "model.layers.8.self_attn.o_proj.weight_scale": "model-00001-of-00004.safetensors",
705
+ "model.layers.8.self_attn.q_norm.weight": "model-00001-of-00004.safetensors",
706
+ "model.layers.8.self_attn.q_proj.weight": "model-00001-of-00004.safetensors",
707
+ "model.layers.8.self_attn.q_proj.weight_scale": "model-00001-of-00004.safetensors",
708
+ "model.layers.8.self_attn.v_proj.weight": "model-00001-of-00004.safetensors",
709
+ "model.layers.8.self_attn.v_proj.weight_scale": "model-00001-of-00004.safetensors",
710
+ "model.layers.9.input_layernorm.weight": "model-00001-of-00004.safetensors",
711
+ "model.layers.9.mlp.down_proj.weight": "model-00001-of-00004.safetensors",
712
+ "model.layers.9.mlp.down_proj.weight_scale": "model-00001-of-00004.safetensors",
713
+ "model.layers.9.mlp.gate_proj.weight": "model-00001-of-00004.safetensors",
714
+ "model.layers.9.mlp.gate_proj.weight_scale": "model-00001-of-00004.safetensors",
715
+ "model.layers.9.mlp.up_proj.weight": "model-00001-of-00004.safetensors",
716
+ "model.layers.9.mlp.up_proj.weight_scale": "model-00001-of-00004.safetensors",
717
+ "model.layers.9.post_attention_layernorm.weight": "model-00001-of-00004.safetensors",
718
+ "model.layers.9.self_attn.k_norm.weight": "model-00001-of-00004.safetensors",
719
+ "model.layers.9.self_attn.k_proj.weight": "model-00001-of-00004.safetensors",
720
+ "model.layers.9.self_attn.k_proj.weight_scale": "model-00001-of-00004.safetensors",
721
+ "model.layers.9.self_attn.o_proj.weight": "model-00001-of-00004.safetensors",
722
+ "model.layers.9.self_attn.o_proj.weight_scale": "model-00001-of-00004.safetensors",
723
+ "model.layers.9.self_attn.q_norm.weight": "model-00001-of-00004.safetensors",
724
+ "model.layers.9.self_attn.q_proj.weight": "model-00001-of-00004.safetensors",
725
+ "model.layers.9.self_attn.q_proj.weight_scale": "model-00001-of-00004.safetensors",
726
+ "model.layers.9.self_attn.v_proj.weight": "model-00001-of-00004.safetensors",
727
+ "model.layers.9.self_attn.v_proj.weight_scale": "model-00001-of-00004.safetensors",
728
+ "model.norm.weight": "model-00003-of-00004.safetensors"
729
+ }
730
+ }
recipe.yaml ADDED
@@ -0,0 +1,6 @@
 
 
 
 
 
 
 
1
+ default_stage:
2
+ default_modifiers:
3
+ QuantizationModifier:
4
+ ignore: [lm_head]
5
+ targets: [Linear]
6
+ scheme: FP8_DYNAMIC
special_tokens_map.json ADDED
@@ -0,0 +1,1009 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "additional_special_tokens": [
3
+ "[control_10]",
4
+ "[control_11]",
5
+ "[control_12]",
6
+ "[control_13]",
7
+ "[control_14]",
8
+ "[control_15]",
9
+ "[control_16]",
10
+ "[control_17]",
11
+ "[control_18]",
12
+ "[control_19]",
13
+ "[control_20]",
14
+ "[control_21]",
15
+ "[control_22]",
16
+ "[control_23]",
17
+ "[control_24]",
18
+ "[control_25]",
19
+ "[control_26]",
20
+ "[control_27]",
21
+ "[control_28]",
22
+ "[control_29]",
23
+ "[control_30]",
24
+ "[control_31]",
25
+ "[control_32]",
26
+ "[control_33]",
27
+ "[control_34]",
28
+ "[control_35]",
29
+ "[control_36]",
30
+ "[control_37]",
31
+ "[control_38]",
32
+ "[control_39]",
33
+ "[control_40]",
34
+ "[control_41]",
35
+ "[control_42]",
36
+ "[control_43]",
37
+ "[control_44]",
38
+ "[control_45]",
39
+ "[control_46]",
40
+ "[control_47]",
41
+ "[control_48]",
42
+ "[control_49]",
43
+ "[control_50]",
44
+ "[control_51]",
45
+ "[control_52]",
46
+ "[control_53]",
47
+ "[control_54]",
48
+ "[control_55]",
49
+ "[control_56]",
50
+ "[control_57]",
51
+ "[control_58]",
52
+ "[control_59]",
53
+ "[control_60]",
54
+ "[control_61]",
55
+ "[control_62]",
56
+ "[control_63]",
57
+ "[control_64]",
58
+ "[control_65]",
59
+ "[control_66]",
60
+ "[control_67]",
61
+ "[control_68]",
62
+ "[control_69]",
63
+ "[control_70]",
64
+ "[control_71]",
65
+ "[control_72]",
66
+ "[control_73]",
67
+ "[control_74]",
68
+ "[control_75]",
69
+ "[control_76]",
70
+ "[control_77]",
71
+ "[control_78]",
72
+ "[control_79]",
73
+ "[control_80]",
74
+ "[control_81]",
75
+ "[control_82]",
76
+ "[control_83]",
77
+ "[control_84]",
78
+ "[control_85]",
79
+ "[control_86]",
80
+ "[control_87]",
81
+ "[control_88]",
82
+ "[control_89]",
83
+ "[control_90]",
84
+ "[control_91]",
85
+ "[control_92]",
86
+ "[control_93]",
87
+ "[control_94]",
88
+ "[control_95]",
89
+ "[control_96]",
90
+ "[control_97]",
91
+ "[control_98]",
92
+ "[control_99]",
93
+ "[control_100]",
94
+ "[control_101]",
95
+ "[control_102]",
96
+ "[control_103]",
97
+ "[control_104]",
98
+ "[control_105]",
99
+ "[control_106]",
100
+ "[control_107]",
101
+ "[control_108]",
102
+ "[control_109]",
103
+ "[control_110]",
104
+ "[control_111]",
105
+ "[control_112]",
106
+ "[control_113]",
107
+ "[control_114]",
108
+ "[control_115]",
109
+ "[control_116]",
110
+ "[control_117]",
111
+ "[control_118]",
112
+ "[control_119]",
113
+ "[control_120]",
114
+ "[control_121]",
115
+ "[control_122]",
116
+ "[control_123]",
117
+ "[control_124]",
118
+ "[control_125]",
119
+ "[control_126]",
120
+ "[control_127]",
121
+ "[control_128]",
122
+ "[control_129]",
123
+ "[control_130]",
124
+ "[control_131]",
125
+ "[control_132]",
126
+ "[control_133]",
127
+ "[control_134]",
128
+ "[control_135]",
129
+ "[control_136]",
130
+ "[control_137]",
131
+ "[control_138]",
132
+ "[control_139]",
133
+ "[control_140]",
134
+ "[control_141]",
135
+ "[control_142]",
136
+ "[control_143]",
137
+ "[control_144]",
138
+ "[control_145]",
139
+ "[control_146]",
140
+ "[control_147]",
141
+ "[control_148]",
142
+ "[control_149]",
143
+ "[control_150]",
144
+ "[control_151]",
145
+ "[control_152]",
146
+ "[control_153]",
147
+ "[control_154]",
148
+ "[control_155]",
149
+ "[control_156]",
150
+ "[control_157]",
151
+ "[control_158]",
152
+ "[control_159]",
153
+ "[control_160]",
154
+ "[control_161]",
155
+ "[control_162]",
156
+ "[control_163]",
157
+ "[control_164]",
158
+ "[control_165]",
159
+ "[control_166]",
160
+ "[control_167]",
161
+ "[control_168]",
162
+ "[control_169]",
163
+ "[control_170]",
164
+ "[control_171]",
165
+ "[control_172]",
166
+ "[control_173]",
167
+ "[control_174]",
168
+ "[control_175]",
169
+ "[control_176]",
170
+ "[control_177]",
171
+ "[control_178]",
172
+ "[control_179]",
173
+ "[control_180]",
174
+ "[control_181]",
175
+ "[control_182]",
176
+ "[control_183]",
177
+ "[control_184]",
178
+ "[control_185]",
179
+ "[control_186]",
180
+ "[control_187]",
181
+ "[control_188]",
182
+ "[control_189]",
183
+ "[control_190]",
184
+ "[control_191]",
185
+ "[control_192]",
186
+ "[control_193]",
187
+ "[control_194]",
188
+ "[control_195]",
189
+ "[control_196]",
190
+ "[control_197]",
191
+ "[control_198]",
192
+ "[control_199]",
193
+ "[control_200]",
194
+ "[control_201]",
195
+ "[control_202]",
196
+ "[control_203]",
197
+ "[control_204]",
198
+ "[control_205]",
199
+ "[control_206]",
200
+ "[control_207]",
201
+ "[control_208]",
202
+ "[control_209]",
203
+ "[control_210]",
204
+ "[control_211]",
205
+ "[control_212]",
206
+ "[control_213]",
207
+ "[control_214]",
208
+ "[control_215]",
209
+ "[control_216]",
210
+ "[control_217]",
211
+ "[control_218]",
212
+ "[control_219]",
213
+ "[control_220]",
214
+ "[control_221]",
215
+ "[control_222]",
216
+ "[control_223]",
217
+ "[control_224]",
218
+ "[control_225]",
219
+ "[control_226]",
220
+ "[control_227]",
221
+ "[control_228]",
222
+ "[control_229]",
223
+ "[control_230]",
224
+ "[control_231]",
225
+ "[control_232]",
226
+ "[control_233]",
227
+ "[control_234]",
228
+ "[control_235]",
229
+ "[control_236]",
230
+ "[control_237]",
231
+ "[control_238]",
232
+ "[control_239]",
233
+ "[control_240]",
234
+ "[control_241]",
235
+ "[control_242]",
236
+ "[control_243]",
237
+ "[control_244]",
238
+ "[control_245]",
239
+ "[control_246]",
240
+ "[control_247]",
241
+ "[control_248]",
242
+ "[control_249]",
243
+ "[control_250]",
244
+ "[control_251]",
245
+ "[control_252]",
246
+ "[control_253]",
247
+ "[control_254]",
248
+ "[control_255]",
249
+ "[control_256]",
250
+ "[control_257]",
251
+ "[control_258]",
252
+ "[control_259]",
253
+ "[control_260]",
254
+ "[control_261]",
255
+ "[control_262]",
256
+ "[control_263]",
257
+ "[control_264]",
258
+ "[control_265]",
259
+ "[control_266]",
260
+ "[control_267]",
261
+ "[control_268]",
262
+ "[control_269]",
263
+ "[control_270]",
264
+ "[control_271]",
265
+ "[control_272]",
266
+ "[control_273]",
267
+ "[control_274]",
268
+ "[control_275]",
269
+ "[control_276]",
270
+ "[control_277]",
271
+ "[control_278]",
272
+ "[control_279]",
273
+ "[control_280]",
274
+ "[control_281]",
275
+ "[control_282]",
276
+ "[control_283]",
277
+ "[control_284]",
278
+ "[control_285]",
279
+ "[control_286]",
280
+ "[control_287]",
281
+ "[control_288]",
282
+ "[control_289]",
283
+ "[control_290]",
284
+ "[control_291]",
285
+ "[control_292]",
286
+ "[control_293]",
287
+ "[control_294]",
288
+ "[control_295]",
289
+ "[control_296]",
290
+ "[control_297]",
291
+ "[control_298]",
292
+ "[control_299]",
293
+ "[control_300]",
294
+ "[control_301]",
295
+ "[control_302]",
296
+ "[control_303]",
297
+ "[control_304]",
298
+ "[control_305]",
299
+ "[control_306]",
300
+ "[control_307]",
301
+ "[control_308]",
302
+ "[control_309]",
303
+ "[control_310]",
304
+ "[control_311]",
305
+ "[control_312]",
306
+ "[control_313]",
307
+ "[control_314]",
308
+ "[control_315]",
309
+ "[control_316]",
310
+ "[control_317]",
311
+ "[control_318]",
312
+ "[control_319]",
313
+ "[control_320]",
314
+ "[control_321]",
315
+ "[control_322]",
316
+ "[control_323]",
317
+ "[control_324]",
318
+ "[control_325]",
319
+ "[control_326]",
320
+ "[control_327]",
321
+ "[control_328]",
322
+ "[control_329]",
323
+ "[control_330]",
324
+ "[control_331]",
325
+ "[control_332]",
326
+ "[control_333]",
327
+ "[control_334]",
328
+ "[control_335]",
329
+ "[control_336]",
330
+ "[control_337]",
331
+ "[control_338]",
332
+ "[control_339]",
333
+ "[control_340]",
334
+ "[control_341]",
335
+ "[control_342]",
336
+ "[control_343]",
337
+ "[control_344]",
338
+ "[control_345]",
339
+ "[control_346]",
340
+ "[control_347]",
341
+ "[control_348]",
342
+ "[control_349]",
343
+ "[control_350]",
344
+ "[control_351]",
345
+ "[control_352]",
346
+ "[control_353]",
347
+ "[control_354]",
348
+ "[control_355]",
349
+ "[control_356]",
350
+ "[control_357]",
351
+ "[control_358]",
352
+ "[control_359]",
353
+ "[control_360]",
354
+ "[control_361]",
355
+ "[control_362]",
356
+ "[control_363]",
357
+ "[control_364]",
358
+ "[control_365]",
359
+ "[control_366]",
360
+ "[control_367]",
361
+ "[control_368]",
362
+ "[control_369]",
363
+ "[control_370]",
364
+ "[control_371]",
365
+ "[control_372]",
366
+ "[control_373]",
367
+ "[control_374]",
368
+ "[control_375]",
369
+ "[control_376]",
370
+ "[control_377]",
371
+ "[control_378]",
372
+ "[control_379]",
373
+ "[control_380]",
374
+ "[control_381]",
375
+ "[control_382]",
376
+ "[control_383]",
377
+ "[control_384]",
378
+ "[control_385]",
379
+ "[control_386]",
380
+ "[control_387]",
381
+ "[control_388]",
382
+ "[control_389]",
383
+ "[control_390]",
384
+ "[control_391]",
385
+ "[control_392]",
386
+ "[control_393]",
387
+ "[control_394]",
388
+ "[control_395]",
389
+ "[control_396]",
390
+ "[control_397]",
391
+ "[control_398]",
392
+ "[control_399]",
393
+ "[control_400]",
394
+ "[control_401]",
395
+ "[control_402]",
396
+ "[control_403]",
397
+ "[control_404]",
398
+ "[control_405]",
399
+ "[control_406]",
400
+ "[control_407]",
401
+ "[control_408]",
402
+ "[control_409]",
403
+ "[control_410]",
404
+ "[control_411]",
405
+ "[control_412]",
406
+ "[control_413]",
407
+ "[control_414]",
408
+ "[control_415]",
409
+ "[control_416]",
410
+ "[control_417]",
411
+ "[control_418]",
412
+ "[control_419]",
413
+ "[control_420]",
414
+ "[control_421]",
415
+ "[control_422]",
416
+ "[control_423]",
417
+ "[control_424]",
418
+ "[control_425]",
419
+ "[control_426]",
420
+ "[control_427]",
421
+ "[control_428]",
422
+ "[control_429]",
423
+ "[control_430]",
424
+ "[control_431]",
425
+ "[control_432]",
426
+ "[control_433]",
427
+ "[control_434]",
428
+ "[control_435]",
429
+ "[control_436]",
430
+ "[control_437]",
431
+ "[control_438]",
432
+ "[control_439]",
433
+ "[control_440]",
434
+ "[control_441]",
435
+ "[control_442]",
436
+ "[control_443]",
437
+ "[control_444]",
438
+ "[control_445]",
439
+ "[control_446]",
440
+ "[control_447]",
441
+ "[control_448]",
442
+ "[control_449]",
443
+ "[control_450]",
444
+ "[control_451]",
445
+ "[control_452]",
446
+ "[control_453]",
447
+ "[control_454]",
448
+ "[control_455]",
449
+ "[control_456]",
450
+ "[control_457]",
451
+ "[control_458]",
452
+ "[control_459]",
453
+ "[control_460]",
454
+ "[control_461]",
455
+ "[control_462]",
456
+ "[control_463]",
457
+ "[control_464]",
458
+ "[control_465]",
459
+ "[control_466]",
460
+ "[control_467]",
461
+ "[control_468]",
462
+ "[control_469]",
463
+ "[control_470]",
464
+ "[control_471]",
465
+ "[control_472]",
466
+ "[control_473]",
467
+ "[control_474]",
468
+ "[control_475]",
469
+ "[control_476]",
470
+ "[control_477]",
471
+ "[control_478]",
472
+ "[control_479]",
473
+ "[control_480]",
474
+ "[control_481]",
475
+ "[control_482]",
476
+ "[control_483]",
477
+ "[control_484]",
478
+ "[control_485]",
479
+ "[control_486]",
480
+ "[control_487]",
481
+ "[control_488]",
482
+ "[control_489]",
483
+ "[control_490]",
484
+ "[control_491]",
485
+ "[control_492]",
486
+ "[control_493]",
487
+ "[control_494]",
488
+ "[control_495]",
489
+ "[control_496]",
490
+ "[control_497]",
491
+ "[control_498]",
492
+ "[control_499]",
493
+ "[control_500]",
494
+ "[control_501]",
495
+ "[control_502]",
496
+ "[control_503]",
497
+ "[control_504]",
498
+ "[control_505]",
499
+ "[control_506]",
500
+ "[control_507]",
501
+ "[control_508]",
502
+ "[control_509]",
503
+ "[control_510]",
504
+ "[control_511]",
505
+ "[control_512]",
506
+ "[control_513]",
507
+ "[control_514]",
508
+ "[control_515]",
509
+ "[control_516]",
510
+ "[control_517]",
511
+ "[control_518]",
512
+ "[control_519]",
513
+ "[control_520]",
514
+ "[control_521]",
515
+ "[control_522]",
516
+ "[control_523]",
517
+ "[control_524]",
518
+ "[control_525]",
519
+ "[control_526]",
520
+ "[control_527]",
521
+ "[control_528]",
522
+ "[control_529]",
523
+ "[control_530]",
524
+ "[control_531]",
525
+ "[control_532]",
526
+ "[control_533]",
527
+ "[control_534]",
528
+ "[control_535]",
529
+ "[control_536]",
530
+ "[control_537]",
531
+ "[control_538]",
532
+ "[control_539]",
533
+ "[control_540]",
534
+ "[control_541]",
535
+ "[control_542]",
536
+ "[control_543]",
537
+ "[control_544]",
538
+ "[control_545]",
539
+ "[control_546]",
540
+ "[control_547]",
541
+ "[control_548]",
542
+ "[control_549]",
543
+ "[control_550]",
544
+ "[control_551]",
545
+ "[control_552]",
546
+ "[control_553]",
547
+ "[control_554]",
548
+ "[control_555]",
549
+ "[control_556]",
550
+ "[control_557]",
551
+ "[control_558]",
552
+ "[control_559]",
553
+ "[control_560]",
554
+ "[control_561]",
555
+ "[control_562]",
556
+ "[control_563]",
557
+ "[control_564]",
558
+ "[control_565]",
559
+ "[control_566]",
560
+ "[control_567]",
561
+ "[control_568]",
562
+ "[control_569]",
563
+ "[control_570]",
564
+ "[control_571]",
565
+ "[control_572]",
566
+ "[control_573]",
567
+ "[control_574]",
568
+ "[control_575]",
569
+ "[control_576]",
570
+ "[control_577]",
571
+ "[control_578]",
572
+ "[control_579]",
573
+ "[control_580]",
574
+ "[control_581]",
575
+ "[control_582]",
576
+ "[control_583]",
577
+ "[control_584]",
578
+ "[control_585]",
579
+ "[control_586]",
580
+ "[control_587]",
581
+ "[control_588]",
582
+ "[control_589]",
583
+ "[control_590]",
584
+ "[control_591]",
585
+ "[control_592]",
586
+ "[control_593]",
587
+ "[control_594]",
588
+ "[control_595]",
589
+ "[control_596]",
590
+ "[control_597]",
591
+ "[control_598]",
592
+ "[control_599]",
593
+ "[control_600]",
594
+ "[control_601]",
595
+ "[control_602]",
596
+ "[control_603]",
597
+ "[control_604]",
598
+ "[control_605]",
599
+ "[control_606]",
600
+ "[control_607]",
601
+ "[control_608]",
602
+ "[control_609]",
603
+ "[control_610]",
604
+ "[control_611]",
605
+ "[control_612]",
606
+ "[control_613]",
607
+ "[control_614]",
608
+ "[control_615]",
609
+ "[control_616]",
610
+ "[control_617]",
611
+ "[control_618]",
612
+ "[control_619]",
613
+ "[control_620]",
614
+ "[control_621]",
615
+ "[control_622]",
616
+ "[control_623]",
617
+ "[control_624]",
618
+ "[control_625]",
619
+ "[control_626]",
620
+ "[control_627]",
621
+ "[control_628]",
622
+ "[control_629]",
623
+ "[control_630]",
624
+ "[control_631]",
625
+ "[control_632]",
626
+ "[control_633]",
627
+ "[control_634]",
628
+ "[control_635]",
629
+ "[control_636]",
630
+ "[control_637]",
631
+ "[control_638]",
632
+ "[control_639]",
633
+ "[control_640]",
634
+ "[control_641]",
635
+ "[control_642]",
636
+ "[control_643]",
637
+ "[control_644]",
638
+ "[control_645]",
639
+ "[control_646]",
640
+ "[control_647]",
641
+ "[control_648]",
642
+ "[control_649]",
643
+ "[control_650]",
644
+ "[control_651]",
645
+ "[control_652]",
646
+ "[control_653]",
647
+ "[control_654]",
648
+ "[control_655]",
649
+ "[control_656]",
650
+ "[control_657]",
651
+ "[control_658]",
652
+ "[control_659]",
653
+ "[control_660]",
654
+ "[control_661]",
655
+ "[control_662]",
656
+ "[control_663]",
657
+ "[control_664]",
658
+ "[control_665]",
659
+ "[control_666]",
660
+ "[control_667]",
661
+ "[control_668]",
662
+ "[control_669]",
663
+ "[control_670]",
664
+ "[control_671]",
665
+ "[control_672]",
666
+ "[control_673]",
667
+ "[control_674]",
668
+ "[control_675]",
669
+ "[control_676]",
670
+ "[control_677]",
671
+ "[control_678]",
672
+ "[control_679]",
673
+ "[control_680]",
674
+ "[control_681]",
675
+ "[control_682]",
676
+ "[control_683]",
677
+ "[control_684]",
678
+ "[control_685]",
679
+ "[control_686]",
680
+ "[control_687]",
681
+ "[control_688]",
682
+ "[control_689]",
683
+ "[control_690]",
684
+ "[control_691]",
685
+ "[control_692]",
686
+ "[control_693]",
687
+ "[control_694]",
688
+ "[control_695]",
689
+ "[control_696]",
690
+ "[control_697]",
691
+ "[control_698]",
692
+ "[control_699]",
693
+ "[control_700]",
694
+ "[control_701]",
695
+ "[control_702]",
696
+ "[control_703]",
697
+ "[control_704]",
698
+ "[control_705]",
699
+ "[control_706]",
700
+ "[control_707]",
701
+ "[control_708]",
702
+ "[control_709]",
703
+ "[control_710]",
704
+ "[control_711]",
705
+ "[control_712]",
706
+ "[control_713]",
707
+ "[control_714]",
708
+ "[control_715]",
709
+ "[control_716]",
710
+ "[control_717]",
711
+ "[control_718]",
712
+ "[control_719]",
713
+ "[control_720]",
714
+ "[control_721]",
715
+ "[control_722]",
716
+ "[control_723]",
717
+ "[control_724]",
718
+ "[control_725]",
719
+ "[control_726]",
720
+ "[control_727]",
721
+ "[control_728]",
722
+ "[control_729]",
723
+ "[control_730]",
724
+ "[control_731]",
725
+ "[control_732]",
726
+ "[control_733]",
727
+ "[control_734]",
728
+ "[control_735]",
729
+ "[control_736]",
730
+ "[control_737]",
731
+ "[control_738]",
732
+ "[control_739]",
733
+ "[control_740]",
734
+ "[control_741]",
735
+ "[control_742]",
736
+ "[control_743]",
737
+ "[control_744]",
738
+ "[control_745]",
739
+ "[control_746]",
740
+ "[control_747]",
741
+ "[control_748]",
742
+ "[control_749]",
743
+ "[control_750]",
744
+ "[control_751]",
745
+ "[control_752]",
746
+ "[control_753]",
747
+ "[control_754]",
748
+ "[control_755]",
749
+ "[control_756]",
750
+ "[control_757]",
751
+ "[control_758]",
752
+ "[control_759]",
753
+ "[control_760]",
754
+ "[control_761]",
755
+ "[control_762]",
756
+ "[control_763]",
757
+ "[control_764]",
758
+ "[control_765]",
759
+ "[control_766]",
760
+ "[control_767]",
761
+ "[control_768]",
762
+ "[control_769]",
763
+ "[control_770]",
764
+ "[control_771]",
765
+ "[control_772]",
766
+ "[control_773]",
767
+ "[control_774]",
768
+ "[control_775]",
769
+ "[control_776]",
770
+ "[control_777]",
771
+ "[control_778]",
772
+ "[control_779]",
773
+ "[control_780]",
774
+ "[control_781]",
775
+ "[control_782]",
776
+ "[control_783]",
777
+ "[control_784]",
778
+ "[control_785]",
779
+ "[control_786]",
780
+ "[control_787]",
781
+ "[control_788]",
782
+ "[control_789]",
783
+ "[control_790]",
784
+ "[control_791]",
785
+ "[control_792]",
786
+ "[control_793]",
787
+ "[control_794]",
788
+ "[control_795]",
789
+ "[control_796]",
790
+ "[control_797]",
791
+ "[control_798]",
792
+ "[control_799]",
793
+ "[control_800]",
794
+ "[control_801]",
795
+ "[control_802]",
796
+ "[control_803]",
797
+ "[control_804]",
798
+ "[control_805]",
799
+ "[control_806]",
800
+ "[control_807]",
801
+ "[control_808]",
802
+ "[control_809]",
803
+ "[control_810]",
804
+ "[control_811]",
805
+ "[control_812]",
806
+ "[control_813]",
807
+ "[control_814]",
808
+ "[control_815]",
809
+ "[control_816]",
810
+ "[control_817]",
811
+ "[control_818]",
812
+ "[control_819]",
813
+ "[control_820]",
814
+ "[control_821]",
815
+ "[control_822]",
816
+ "[control_823]",
817
+ "[control_824]",
818
+ "[control_825]",
819
+ "[control_826]",
820
+ "[control_827]",
821
+ "[control_828]",
822
+ "[control_829]",
823
+ "[control_830]",
824
+ "[control_831]",
825
+ "[control_832]",
826
+ "[control_833]",
827
+ "[control_834]",
828
+ "[control_835]",
829
+ "[control_836]",
830
+ "[control_837]",
831
+ "[control_838]",
832
+ "[control_839]",
833
+ "[control_840]",
834
+ "[control_841]",
835
+ "[control_842]",
836
+ "[control_843]",
837
+ "[control_844]",
838
+ "[control_845]",
839
+ "[control_846]",
840
+ "[control_847]",
841
+ "[control_848]",
842
+ "[control_849]",
843
+ "[control_850]",
844
+ "[control_851]",
845
+ "[control_852]",
846
+ "[control_853]",
847
+ "[control_854]",
848
+ "[control_855]",
849
+ "[control_856]",
850
+ "[control_857]",
851
+ "[control_858]",
852
+ "[control_859]",
853
+ "[control_860]",
854
+ "[control_861]",
855
+ "[control_862]",
856
+ "[control_863]",
857
+ "[control_864]",
858
+ "[control_865]",
859
+ "[control_866]",
860
+ "[control_867]",
861
+ "[control_868]",
862
+ "[control_869]",
863
+ "[control_870]",
864
+ "[control_871]",
865
+ "[control_872]",
866
+ "[control_873]",
867
+ "[control_874]",
868
+ "[control_875]",
869
+ "[control_876]",
870
+ "[control_877]",
871
+ "[control_878]",
872
+ "[control_879]",
873
+ "[control_880]",
874
+ "[control_881]",
875
+ "[control_882]",
876
+ "[control_883]",
877
+ "[control_884]",
878
+ "[control_885]",
879
+ "[control_886]",
880
+ "[control_887]",
881
+ "[control_888]",
882
+ "[control_889]",
883
+ "[control_890]",
884
+ "[control_891]",
885
+ "[control_892]",
886
+ "[control_893]",
887
+ "[control_894]",
888
+ "[control_895]",
889
+ "[control_896]",
890
+ "[control_897]",
891
+ "[control_898]",
892
+ "[control_899]",
893
+ "[control_900]",
894
+ "[control_901]",
895
+ "[control_902]",
896
+ "[control_903]",
897
+ "[control_904]",
898
+ "[control_905]",
899
+ "[control_906]",
900
+ "[control_907]",
901
+ "[control_908]",
902
+ "[control_909]",
903
+ "[control_910]",
904
+ "[control_911]",
905
+ "[control_912]",
906
+ "[control_913]",
907
+ "[control_914]",
908
+ "[control_915]",
909
+ "[control_916]",
910
+ "[control_917]",
911
+ "[control_918]",
912
+ "[control_919]",
913
+ "[control_920]",
914
+ "[control_921]",
915
+ "[control_922]",
916
+ "[control_923]",
917
+ "[control_924]",
918
+ "[control_925]",
919
+ "[control_926]",
920
+ "[control_927]",
921
+ "[control_928]",
922
+ "[control_929]",
923
+ "[control_930]",
924
+ "[control_931]",
925
+ "[control_932]",
926
+ "[control_933]",
927
+ "[control_934]",
928
+ "[control_935]",
929
+ "[control_936]",
930
+ "[control_937]",
931
+ "[control_938]",
932
+ "[control_939]",
933
+ "[control_940]",
934
+ "[control_941]",
935
+ "[control_942]",
936
+ "[control_943]",
937
+ "[control_944]",
938
+ "[control_945]",
939
+ "[control_946]",
940
+ "[control_947]",
941
+ "[control_948]",
942
+ "[control_949]",
943
+ "[control_950]",
944
+ "[control_951]",
945
+ "[control_952]",
946
+ "[control_953]",
947
+ "[control_954]",
948
+ "[control_955]",
949
+ "[control_956]",
950
+ "[control_957]",
951
+ "[control_958]",
952
+ "[control_959]",
953
+ "[control_960]",
954
+ "[control_961]",
955
+ "[control_962]",
956
+ "[control_963]",
957
+ "[control_964]",
958
+ "[control_965]",
959
+ "[control_966]",
960
+ "[control_967]",
961
+ "[control_968]",
962
+ "[control_969]",
963
+ "[control_970]",
964
+ "[control_971]",
965
+ "[control_972]",
966
+ "[control_973]",
967
+ "[control_974]",
968
+ "[control_975]",
969
+ "[control_976]",
970
+ "[control_977]",
971
+ "[control_978]",
972
+ "[control_979]",
973
+ "[control_980]",
974
+ "[control_981]",
975
+ "[control_982]",
976
+ "[control_983]",
977
+ "[control_984]",
978
+ "[control_985]",
979
+ "[control_986]",
980
+ "[control_987]",
981
+ "[control_988]",
982
+ "[control_989]",
983
+ "[control_990]",
984
+ "[control_991]",
985
+ "[control_992]",
986
+ "[control_993]",
987
+ "[control_994]",
988
+ "[control_995]",
989
+ "[control_996]",
990
+ "[control_997]",
991
+ "[control_998]",
992
+ "[control_999]",
993
+ "[control_1000]"
994
+ ],
995
+ "eos_token": {
996
+ "content": "<|im_end|>",
997
+ "lstrip": false,
998
+ "normalized": false,
999
+ "rstrip": false,
1000
+ "single_word": false
1001
+ },
1002
+ "pad_token": {
1003
+ "content": "<|endoftext|>",
1004
+ "lstrip": false,
1005
+ "normalized": false,
1006
+ "rstrip": false,
1007
+ "single_word": false
1008
+ }
1009
+ }
tokenizer.json ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:328a87eddd7f2f23d04dd5f861aceadf75f9af4563253644a1b6c11bc09a19ef
3
+ size 11610855
tokenizer_config.json ADDED
The diff for this file is too large to render. See raw diff
 
vocab.json ADDED
The diff for this file is too large to render. See raw diff