Upload folder using huggingface_hub
Browse files- .gitattributes +1 -0
- added_tokens.json +1019 -0
- config.json +72 -0
- generation_config.json +13 -0
- merges.txt +0 -0
- model-00001-of-00004.safetensors +3 -0
- model-00002-of-00004.safetensors +3 -0
- model-00003-of-00004.safetensors +3 -0
- model-00004-of-00004.safetensors +3 -0
- model.safetensors.index.json +730 -0
- recipe.yaml +6 -0
- special_tokens_map.json +1009 -0
- tokenizer.json +3 -0
- tokenizer_config.json +0 -0
- vocab.json +0 -0
.gitattributes
CHANGED
@@ -33,3 +33,4 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
|
|
33 |
*.zip filter=lfs diff=lfs merge=lfs -text
|
34 |
*.zst filter=lfs diff=lfs merge=lfs -text
|
35 |
*tfevents* filter=lfs diff=lfs merge=lfs -text
|
|
|
|
33 |
*.zip filter=lfs diff=lfs merge=lfs -text
|
34 |
*.zst filter=lfs diff=lfs merge=lfs -text
|
35 |
*tfevents* filter=lfs diff=lfs merge=lfs -text
|
36 |
+
tokenizer.json filter=lfs diff=lfs merge=lfs -text
|
added_tokens.json
ADDED
@@ -0,0 +1,1019 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"</think>": 151668,
|
3 |
+
"</tool_call>": 151658,
|
4 |
+
"</tool_response>": 151666,
|
5 |
+
"<think>": 151667,
|
6 |
+
"<tool_call>": 151657,
|
7 |
+
"<tool_response>": 151665,
|
8 |
+
"<|box_end|>": 151649,
|
9 |
+
"<|box_start|>": 151648,
|
10 |
+
"<|endoftext|>": 151643,
|
11 |
+
"<|file_sep|>": 151664,
|
12 |
+
"<|fim_middle|>": 151660,
|
13 |
+
"<|fim_pad|>": 151662,
|
14 |
+
"<|fim_prefix|>": 151659,
|
15 |
+
"<|fim_suffix|>": 151661,
|
16 |
+
"<|im_end|>": 151645,
|
17 |
+
"<|im_start|>": 151644,
|
18 |
+
"<|image_pad|>": 151655,
|
19 |
+
"<|object_ref_end|>": 151647,
|
20 |
+
"<|object_ref_start|>": 151646,
|
21 |
+
"<|quad_end|>": 151651,
|
22 |
+
"<|quad_start|>": 151650,
|
23 |
+
"<|repo_name|>": 151663,
|
24 |
+
"<|video_pad|>": 151656,
|
25 |
+
"<|vision_end|>": 151653,
|
26 |
+
"<|vision_pad|>": 151654,
|
27 |
+
"<|vision_start|>": 151652,
|
28 |
+
"[control_1000]": 152659,
|
29 |
+
"[control_100]": 151759,
|
30 |
+
"[control_101]": 151760,
|
31 |
+
"[control_102]": 151761,
|
32 |
+
"[control_103]": 151762,
|
33 |
+
"[control_104]": 151763,
|
34 |
+
"[control_105]": 151764,
|
35 |
+
"[control_106]": 151765,
|
36 |
+
"[control_107]": 151766,
|
37 |
+
"[control_108]": 151767,
|
38 |
+
"[control_109]": 151768,
|
39 |
+
"[control_10]": 151669,
|
40 |
+
"[control_110]": 151769,
|
41 |
+
"[control_111]": 151770,
|
42 |
+
"[control_112]": 151771,
|
43 |
+
"[control_113]": 151772,
|
44 |
+
"[control_114]": 151773,
|
45 |
+
"[control_115]": 151774,
|
46 |
+
"[control_116]": 151775,
|
47 |
+
"[control_117]": 151776,
|
48 |
+
"[control_118]": 151777,
|
49 |
+
"[control_119]": 151778,
|
50 |
+
"[control_11]": 151670,
|
51 |
+
"[control_120]": 151779,
|
52 |
+
"[control_121]": 151780,
|
53 |
+
"[control_122]": 151781,
|
54 |
+
"[control_123]": 151782,
|
55 |
+
"[control_124]": 151783,
|
56 |
+
"[control_125]": 151784,
|
57 |
+
"[control_126]": 151785,
|
58 |
+
"[control_127]": 151786,
|
59 |
+
"[control_128]": 151787,
|
60 |
+
"[control_129]": 151788,
|
61 |
+
"[control_12]": 151671,
|
62 |
+
"[control_130]": 151789,
|
63 |
+
"[control_131]": 151790,
|
64 |
+
"[control_132]": 151791,
|
65 |
+
"[control_133]": 151792,
|
66 |
+
"[control_134]": 151793,
|
67 |
+
"[control_135]": 151794,
|
68 |
+
"[control_136]": 151795,
|
69 |
+
"[control_137]": 151796,
|
70 |
+
"[control_138]": 151797,
|
71 |
+
"[control_139]": 151798,
|
72 |
+
"[control_13]": 151672,
|
73 |
+
"[control_140]": 151799,
|
74 |
+
"[control_141]": 151800,
|
75 |
+
"[control_142]": 151801,
|
76 |
+
"[control_143]": 151802,
|
77 |
+
"[control_144]": 151803,
|
78 |
+
"[control_145]": 151804,
|
79 |
+
"[control_146]": 151805,
|
80 |
+
"[control_147]": 151806,
|
81 |
+
"[control_148]": 151807,
|
82 |
+
"[control_149]": 151808,
|
83 |
+
"[control_14]": 151673,
|
84 |
+
"[control_150]": 151809,
|
85 |
+
"[control_151]": 151810,
|
86 |
+
"[control_152]": 151811,
|
87 |
+
"[control_153]": 151812,
|
88 |
+
"[control_154]": 151813,
|
89 |
+
"[control_155]": 151814,
|
90 |
+
"[control_156]": 151815,
|
91 |
+
"[control_157]": 151816,
|
92 |
+
"[control_158]": 151817,
|
93 |
+
"[control_159]": 151818,
|
94 |
+
"[control_15]": 151674,
|
95 |
+
"[control_160]": 151819,
|
96 |
+
"[control_161]": 151820,
|
97 |
+
"[control_162]": 151821,
|
98 |
+
"[control_163]": 151822,
|
99 |
+
"[control_164]": 151823,
|
100 |
+
"[control_165]": 151824,
|
101 |
+
"[control_166]": 151825,
|
102 |
+
"[control_167]": 151826,
|
103 |
+
"[control_168]": 151827,
|
104 |
+
"[control_169]": 151828,
|
105 |
+
"[control_16]": 151675,
|
106 |
+
"[control_170]": 151829,
|
107 |
+
"[control_171]": 151830,
|
108 |
+
"[control_172]": 151831,
|
109 |
+
"[control_173]": 151832,
|
110 |
+
"[control_174]": 151833,
|
111 |
+
"[control_175]": 151834,
|
112 |
+
"[control_176]": 151835,
|
113 |
+
"[control_177]": 151836,
|
114 |
+
"[control_178]": 151837,
|
115 |
+
"[control_179]": 151838,
|
116 |
+
"[control_17]": 151676,
|
117 |
+
"[control_180]": 151839,
|
118 |
+
"[control_181]": 151840,
|
119 |
+
"[control_182]": 151841,
|
120 |
+
"[control_183]": 151842,
|
121 |
+
"[control_184]": 151843,
|
122 |
+
"[control_185]": 151844,
|
123 |
+
"[control_186]": 151845,
|
124 |
+
"[control_187]": 151846,
|
125 |
+
"[control_188]": 151847,
|
126 |
+
"[control_189]": 151848,
|
127 |
+
"[control_18]": 151677,
|
128 |
+
"[control_190]": 151849,
|
129 |
+
"[control_191]": 151850,
|
130 |
+
"[control_192]": 151851,
|
131 |
+
"[control_193]": 151852,
|
132 |
+
"[control_194]": 151853,
|
133 |
+
"[control_195]": 151854,
|
134 |
+
"[control_196]": 151855,
|
135 |
+
"[control_197]": 151856,
|
136 |
+
"[control_198]": 151857,
|
137 |
+
"[control_199]": 151858,
|
138 |
+
"[control_19]": 151678,
|
139 |
+
"[control_200]": 151859,
|
140 |
+
"[control_201]": 151860,
|
141 |
+
"[control_202]": 151861,
|
142 |
+
"[control_203]": 151862,
|
143 |
+
"[control_204]": 151863,
|
144 |
+
"[control_205]": 151864,
|
145 |
+
"[control_206]": 151865,
|
146 |
+
"[control_207]": 151866,
|
147 |
+
"[control_208]": 151867,
|
148 |
+
"[control_209]": 151868,
|
149 |
+
"[control_20]": 151679,
|
150 |
+
"[control_210]": 151869,
|
151 |
+
"[control_211]": 151870,
|
152 |
+
"[control_212]": 151871,
|
153 |
+
"[control_213]": 151872,
|
154 |
+
"[control_214]": 151873,
|
155 |
+
"[control_215]": 151874,
|
156 |
+
"[control_216]": 151875,
|
157 |
+
"[control_217]": 151876,
|
158 |
+
"[control_218]": 151877,
|
159 |
+
"[control_219]": 151878,
|
160 |
+
"[control_21]": 151680,
|
161 |
+
"[control_220]": 151879,
|
162 |
+
"[control_221]": 151880,
|
163 |
+
"[control_222]": 151881,
|
164 |
+
"[control_223]": 151882,
|
165 |
+
"[control_224]": 151883,
|
166 |
+
"[control_225]": 151884,
|
167 |
+
"[control_226]": 151885,
|
168 |
+
"[control_227]": 151886,
|
169 |
+
"[control_228]": 151887,
|
170 |
+
"[control_229]": 151888,
|
171 |
+
"[control_22]": 151681,
|
172 |
+
"[control_230]": 151889,
|
173 |
+
"[control_231]": 151890,
|
174 |
+
"[control_232]": 151891,
|
175 |
+
"[control_233]": 151892,
|
176 |
+
"[control_234]": 151893,
|
177 |
+
"[control_235]": 151894,
|
178 |
+
"[control_236]": 151895,
|
179 |
+
"[control_237]": 151896,
|
180 |
+
"[control_238]": 151897,
|
181 |
+
"[control_239]": 151898,
|
182 |
+
"[control_23]": 151682,
|
183 |
+
"[control_240]": 151899,
|
184 |
+
"[control_241]": 151900,
|
185 |
+
"[control_242]": 151901,
|
186 |
+
"[control_243]": 151902,
|
187 |
+
"[control_244]": 151903,
|
188 |
+
"[control_245]": 151904,
|
189 |
+
"[control_246]": 151905,
|
190 |
+
"[control_247]": 151906,
|
191 |
+
"[control_248]": 151907,
|
192 |
+
"[control_249]": 151908,
|
193 |
+
"[control_24]": 151683,
|
194 |
+
"[control_250]": 151909,
|
195 |
+
"[control_251]": 151910,
|
196 |
+
"[control_252]": 151911,
|
197 |
+
"[control_253]": 151912,
|
198 |
+
"[control_254]": 151913,
|
199 |
+
"[control_255]": 151914,
|
200 |
+
"[control_256]": 151915,
|
201 |
+
"[control_257]": 151916,
|
202 |
+
"[control_258]": 151917,
|
203 |
+
"[control_259]": 151918,
|
204 |
+
"[control_25]": 151684,
|
205 |
+
"[control_260]": 151919,
|
206 |
+
"[control_261]": 151920,
|
207 |
+
"[control_262]": 151921,
|
208 |
+
"[control_263]": 151922,
|
209 |
+
"[control_264]": 151923,
|
210 |
+
"[control_265]": 151924,
|
211 |
+
"[control_266]": 151925,
|
212 |
+
"[control_267]": 151926,
|
213 |
+
"[control_268]": 151927,
|
214 |
+
"[control_269]": 151928,
|
215 |
+
"[control_26]": 151685,
|
216 |
+
"[control_270]": 151929,
|
217 |
+
"[control_271]": 151930,
|
218 |
+
"[control_272]": 151931,
|
219 |
+
"[control_273]": 151932,
|
220 |
+
"[control_274]": 151933,
|
221 |
+
"[control_275]": 151934,
|
222 |
+
"[control_276]": 151935,
|
223 |
+
"[control_277]": 151936,
|
224 |
+
"[control_278]": 151937,
|
225 |
+
"[control_279]": 151938,
|
226 |
+
"[control_27]": 151686,
|
227 |
+
"[control_280]": 151939,
|
228 |
+
"[control_281]": 151940,
|
229 |
+
"[control_282]": 151941,
|
230 |
+
"[control_283]": 151942,
|
231 |
+
"[control_284]": 151943,
|
232 |
+
"[control_285]": 151944,
|
233 |
+
"[control_286]": 151945,
|
234 |
+
"[control_287]": 151946,
|
235 |
+
"[control_288]": 151947,
|
236 |
+
"[control_289]": 151948,
|
237 |
+
"[control_28]": 151687,
|
238 |
+
"[control_290]": 151949,
|
239 |
+
"[control_291]": 151950,
|
240 |
+
"[control_292]": 151951,
|
241 |
+
"[control_293]": 151952,
|
242 |
+
"[control_294]": 151953,
|
243 |
+
"[control_295]": 151954,
|
244 |
+
"[control_296]": 151955,
|
245 |
+
"[control_297]": 151956,
|
246 |
+
"[control_298]": 151957,
|
247 |
+
"[control_299]": 151958,
|
248 |
+
"[control_29]": 151688,
|
249 |
+
"[control_300]": 151959,
|
250 |
+
"[control_301]": 151960,
|
251 |
+
"[control_302]": 151961,
|
252 |
+
"[control_303]": 151962,
|
253 |
+
"[control_304]": 151963,
|
254 |
+
"[control_305]": 151964,
|
255 |
+
"[control_306]": 151965,
|
256 |
+
"[control_307]": 151966,
|
257 |
+
"[control_308]": 151967,
|
258 |
+
"[control_309]": 151968,
|
259 |
+
"[control_30]": 151689,
|
260 |
+
"[control_310]": 151969,
|
261 |
+
"[control_311]": 151970,
|
262 |
+
"[control_312]": 151971,
|
263 |
+
"[control_313]": 151972,
|
264 |
+
"[control_314]": 151973,
|
265 |
+
"[control_315]": 151974,
|
266 |
+
"[control_316]": 151975,
|
267 |
+
"[control_317]": 151976,
|
268 |
+
"[control_318]": 151977,
|
269 |
+
"[control_319]": 151978,
|
270 |
+
"[control_31]": 151690,
|
271 |
+
"[control_320]": 151979,
|
272 |
+
"[control_321]": 151980,
|
273 |
+
"[control_322]": 151981,
|
274 |
+
"[control_323]": 151982,
|
275 |
+
"[control_324]": 151983,
|
276 |
+
"[control_325]": 151984,
|
277 |
+
"[control_326]": 151985,
|
278 |
+
"[control_327]": 151986,
|
279 |
+
"[control_328]": 151987,
|
280 |
+
"[control_329]": 151988,
|
281 |
+
"[control_32]": 151691,
|
282 |
+
"[control_330]": 151989,
|
283 |
+
"[control_331]": 151990,
|
284 |
+
"[control_332]": 151991,
|
285 |
+
"[control_333]": 151992,
|
286 |
+
"[control_334]": 151993,
|
287 |
+
"[control_335]": 151994,
|
288 |
+
"[control_336]": 151995,
|
289 |
+
"[control_337]": 151996,
|
290 |
+
"[control_338]": 151997,
|
291 |
+
"[control_339]": 151998,
|
292 |
+
"[control_33]": 151692,
|
293 |
+
"[control_340]": 151999,
|
294 |
+
"[control_341]": 152000,
|
295 |
+
"[control_342]": 152001,
|
296 |
+
"[control_343]": 152002,
|
297 |
+
"[control_344]": 152003,
|
298 |
+
"[control_345]": 152004,
|
299 |
+
"[control_346]": 152005,
|
300 |
+
"[control_347]": 152006,
|
301 |
+
"[control_348]": 152007,
|
302 |
+
"[control_349]": 152008,
|
303 |
+
"[control_34]": 151693,
|
304 |
+
"[control_350]": 152009,
|
305 |
+
"[control_351]": 152010,
|
306 |
+
"[control_352]": 152011,
|
307 |
+
"[control_353]": 152012,
|
308 |
+
"[control_354]": 152013,
|
309 |
+
"[control_355]": 152014,
|
310 |
+
"[control_356]": 152015,
|
311 |
+
"[control_357]": 152016,
|
312 |
+
"[control_358]": 152017,
|
313 |
+
"[control_359]": 152018,
|
314 |
+
"[control_35]": 151694,
|
315 |
+
"[control_360]": 152019,
|
316 |
+
"[control_361]": 152020,
|
317 |
+
"[control_362]": 152021,
|
318 |
+
"[control_363]": 152022,
|
319 |
+
"[control_364]": 152023,
|
320 |
+
"[control_365]": 152024,
|
321 |
+
"[control_366]": 152025,
|
322 |
+
"[control_367]": 152026,
|
323 |
+
"[control_368]": 152027,
|
324 |
+
"[control_369]": 152028,
|
325 |
+
"[control_36]": 151695,
|
326 |
+
"[control_370]": 152029,
|
327 |
+
"[control_371]": 152030,
|
328 |
+
"[control_372]": 152031,
|
329 |
+
"[control_373]": 152032,
|
330 |
+
"[control_374]": 152033,
|
331 |
+
"[control_375]": 152034,
|
332 |
+
"[control_376]": 152035,
|
333 |
+
"[control_377]": 152036,
|
334 |
+
"[control_378]": 152037,
|
335 |
+
"[control_379]": 152038,
|
336 |
+
"[control_37]": 151696,
|
337 |
+
"[control_380]": 152039,
|
338 |
+
"[control_381]": 152040,
|
339 |
+
"[control_382]": 152041,
|
340 |
+
"[control_383]": 152042,
|
341 |
+
"[control_384]": 152043,
|
342 |
+
"[control_385]": 152044,
|
343 |
+
"[control_386]": 152045,
|
344 |
+
"[control_387]": 152046,
|
345 |
+
"[control_388]": 152047,
|
346 |
+
"[control_389]": 152048,
|
347 |
+
"[control_38]": 151697,
|
348 |
+
"[control_390]": 152049,
|
349 |
+
"[control_391]": 152050,
|
350 |
+
"[control_392]": 152051,
|
351 |
+
"[control_393]": 152052,
|
352 |
+
"[control_394]": 152053,
|
353 |
+
"[control_395]": 152054,
|
354 |
+
"[control_396]": 152055,
|
355 |
+
"[control_397]": 152056,
|
356 |
+
"[control_398]": 152057,
|
357 |
+
"[control_399]": 152058,
|
358 |
+
"[control_39]": 151698,
|
359 |
+
"[control_400]": 152059,
|
360 |
+
"[control_401]": 152060,
|
361 |
+
"[control_402]": 152061,
|
362 |
+
"[control_403]": 152062,
|
363 |
+
"[control_404]": 152063,
|
364 |
+
"[control_405]": 152064,
|
365 |
+
"[control_406]": 152065,
|
366 |
+
"[control_407]": 152066,
|
367 |
+
"[control_408]": 152067,
|
368 |
+
"[control_409]": 152068,
|
369 |
+
"[control_40]": 151699,
|
370 |
+
"[control_410]": 152069,
|
371 |
+
"[control_411]": 152070,
|
372 |
+
"[control_412]": 152071,
|
373 |
+
"[control_413]": 152072,
|
374 |
+
"[control_414]": 152073,
|
375 |
+
"[control_415]": 152074,
|
376 |
+
"[control_416]": 152075,
|
377 |
+
"[control_417]": 152076,
|
378 |
+
"[control_418]": 152077,
|
379 |
+
"[control_419]": 152078,
|
380 |
+
"[control_41]": 151700,
|
381 |
+
"[control_420]": 152079,
|
382 |
+
"[control_421]": 152080,
|
383 |
+
"[control_422]": 152081,
|
384 |
+
"[control_423]": 152082,
|
385 |
+
"[control_424]": 152083,
|
386 |
+
"[control_425]": 152084,
|
387 |
+
"[control_426]": 152085,
|
388 |
+
"[control_427]": 152086,
|
389 |
+
"[control_428]": 152087,
|
390 |
+
"[control_429]": 152088,
|
391 |
+
"[control_42]": 151701,
|
392 |
+
"[control_430]": 152089,
|
393 |
+
"[control_431]": 152090,
|
394 |
+
"[control_432]": 152091,
|
395 |
+
"[control_433]": 152092,
|
396 |
+
"[control_434]": 152093,
|
397 |
+
"[control_435]": 152094,
|
398 |
+
"[control_436]": 152095,
|
399 |
+
"[control_437]": 152096,
|
400 |
+
"[control_438]": 152097,
|
401 |
+
"[control_439]": 152098,
|
402 |
+
"[control_43]": 151702,
|
403 |
+
"[control_440]": 152099,
|
404 |
+
"[control_441]": 152100,
|
405 |
+
"[control_442]": 152101,
|
406 |
+
"[control_443]": 152102,
|
407 |
+
"[control_444]": 152103,
|
408 |
+
"[control_445]": 152104,
|
409 |
+
"[control_446]": 152105,
|
410 |
+
"[control_447]": 152106,
|
411 |
+
"[control_448]": 152107,
|
412 |
+
"[control_449]": 152108,
|
413 |
+
"[control_44]": 151703,
|
414 |
+
"[control_450]": 152109,
|
415 |
+
"[control_451]": 152110,
|
416 |
+
"[control_452]": 152111,
|
417 |
+
"[control_453]": 152112,
|
418 |
+
"[control_454]": 152113,
|
419 |
+
"[control_455]": 152114,
|
420 |
+
"[control_456]": 152115,
|
421 |
+
"[control_457]": 152116,
|
422 |
+
"[control_458]": 152117,
|
423 |
+
"[control_459]": 152118,
|
424 |
+
"[control_45]": 151704,
|
425 |
+
"[control_460]": 152119,
|
426 |
+
"[control_461]": 152120,
|
427 |
+
"[control_462]": 152121,
|
428 |
+
"[control_463]": 152122,
|
429 |
+
"[control_464]": 152123,
|
430 |
+
"[control_465]": 152124,
|
431 |
+
"[control_466]": 152125,
|
432 |
+
"[control_467]": 152126,
|
433 |
+
"[control_468]": 152127,
|
434 |
+
"[control_469]": 152128,
|
435 |
+
"[control_46]": 151705,
|
436 |
+
"[control_470]": 152129,
|
437 |
+
"[control_471]": 152130,
|
438 |
+
"[control_472]": 152131,
|
439 |
+
"[control_473]": 152132,
|
440 |
+
"[control_474]": 152133,
|
441 |
+
"[control_475]": 152134,
|
442 |
+
"[control_476]": 152135,
|
443 |
+
"[control_477]": 152136,
|
444 |
+
"[control_478]": 152137,
|
445 |
+
"[control_479]": 152138,
|
446 |
+
"[control_47]": 151706,
|
447 |
+
"[control_480]": 152139,
|
448 |
+
"[control_481]": 152140,
|
449 |
+
"[control_482]": 152141,
|
450 |
+
"[control_483]": 152142,
|
451 |
+
"[control_484]": 152143,
|
452 |
+
"[control_485]": 152144,
|
453 |
+
"[control_486]": 152145,
|
454 |
+
"[control_487]": 152146,
|
455 |
+
"[control_488]": 152147,
|
456 |
+
"[control_489]": 152148,
|
457 |
+
"[control_48]": 151707,
|
458 |
+
"[control_490]": 152149,
|
459 |
+
"[control_491]": 152150,
|
460 |
+
"[control_492]": 152151,
|
461 |
+
"[control_493]": 152152,
|
462 |
+
"[control_494]": 152153,
|
463 |
+
"[control_495]": 152154,
|
464 |
+
"[control_496]": 152155,
|
465 |
+
"[control_497]": 152156,
|
466 |
+
"[control_498]": 152157,
|
467 |
+
"[control_499]": 152158,
|
468 |
+
"[control_49]": 151708,
|
469 |
+
"[control_500]": 152159,
|
470 |
+
"[control_501]": 152160,
|
471 |
+
"[control_502]": 152161,
|
472 |
+
"[control_503]": 152162,
|
473 |
+
"[control_504]": 152163,
|
474 |
+
"[control_505]": 152164,
|
475 |
+
"[control_506]": 152165,
|
476 |
+
"[control_507]": 152166,
|
477 |
+
"[control_508]": 152167,
|
478 |
+
"[control_509]": 152168,
|
479 |
+
"[control_50]": 151709,
|
480 |
+
"[control_510]": 152169,
|
481 |
+
"[control_511]": 152170,
|
482 |
+
"[control_512]": 152171,
|
483 |
+
"[control_513]": 152172,
|
484 |
+
"[control_514]": 152173,
|
485 |
+
"[control_515]": 152174,
|
486 |
+
"[control_516]": 152175,
|
487 |
+
"[control_517]": 152176,
|
488 |
+
"[control_518]": 152177,
|
489 |
+
"[control_519]": 152178,
|
490 |
+
"[control_51]": 151710,
|
491 |
+
"[control_520]": 152179,
|
492 |
+
"[control_521]": 152180,
|
493 |
+
"[control_522]": 152181,
|
494 |
+
"[control_523]": 152182,
|
495 |
+
"[control_524]": 152183,
|
496 |
+
"[control_525]": 152184,
|
497 |
+
"[control_526]": 152185,
|
498 |
+
"[control_527]": 152186,
|
499 |
+
"[control_528]": 152187,
|
500 |
+
"[control_529]": 152188,
|
501 |
+
"[control_52]": 151711,
|
502 |
+
"[control_530]": 152189,
|
503 |
+
"[control_531]": 152190,
|
504 |
+
"[control_532]": 152191,
|
505 |
+
"[control_533]": 152192,
|
506 |
+
"[control_534]": 152193,
|
507 |
+
"[control_535]": 152194,
|
508 |
+
"[control_536]": 152195,
|
509 |
+
"[control_537]": 152196,
|
510 |
+
"[control_538]": 152197,
|
511 |
+
"[control_539]": 152198,
|
512 |
+
"[control_53]": 151712,
|
513 |
+
"[control_540]": 152199,
|
514 |
+
"[control_541]": 152200,
|
515 |
+
"[control_542]": 152201,
|
516 |
+
"[control_543]": 152202,
|
517 |
+
"[control_544]": 152203,
|
518 |
+
"[control_545]": 152204,
|
519 |
+
"[control_546]": 152205,
|
520 |
+
"[control_547]": 152206,
|
521 |
+
"[control_548]": 152207,
|
522 |
+
"[control_549]": 152208,
|
523 |
+
"[control_54]": 151713,
|
524 |
+
"[control_550]": 152209,
|
525 |
+
"[control_551]": 152210,
|
526 |
+
"[control_552]": 152211,
|
527 |
+
"[control_553]": 152212,
|
528 |
+
"[control_554]": 152213,
|
529 |
+
"[control_555]": 152214,
|
530 |
+
"[control_556]": 152215,
|
531 |
+
"[control_557]": 152216,
|
532 |
+
"[control_558]": 152217,
|
533 |
+
"[control_559]": 152218,
|
534 |
+
"[control_55]": 151714,
|
535 |
+
"[control_560]": 152219,
|
536 |
+
"[control_561]": 152220,
|
537 |
+
"[control_562]": 152221,
|
538 |
+
"[control_563]": 152222,
|
539 |
+
"[control_564]": 152223,
|
540 |
+
"[control_565]": 152224,
|
541 |
+
"[control_566]": 152225,
|
542 |
+
"[control_567]": 152226,
|
543 |
+
"[control_568]": 152227,
|
544 |
+
"[control_569]": 152228,
|
545 |
+
"[control_56]": 151715,
|
546 |
+
"[control_570]": 152229,
|
547 |
+
"[control_571]": 152230,
|
548 |
+
"[control_572]": 152231,
|
549 |
+
"[control_573]": 152232,
|
550 |
+
"[control_574]": 152233,
|
551 |
+
"[control_575]": 152234,
|
552 |
+
"[control_576]": 152235,
|
553 |
+
"[control_577]": 152236,
|
554 |
+
"[control_578]": 152237,
|
555 |
+
"[control_579]": 152238,
|
556 |
+
"[control_57]": 151716,
|
557 |
+
"[control_580]": 152239,
|
558 |
+
"[control_581]": 152240,
|
559 |
+
"[control_582]": 152241,
|
560 |
+
"[control_583]": 152242,
|
561 |
+
"[control_584]": 152243,
|
562 |
+
"[control_585]": 152244,
|
563 |
+
"[control_586]": 152245,
|
564 |
+
"[control_587]": 152246,
|
565 |
+
"[control_588]": 152247,
|
566 |
+
"[control_589]": 152248,
|
567 |
+
"[control_58]": 151717,
|
568 |
+
"[control_590]": 152249,
|
569 |
+
"[control_591]": 152250,
|
570 |
+
"[control_592]": 152251,
|
571 |
+
"[control_593]": 152252,
|
572 |
+
"[control_594]": 152253,
|
573 |
+
"[control_595]": 152254,
|
574 |
+
"[control_596]": 152255,
|
575 |
+
"[control_597]": 152256,
|
576 |
+
"[control_598]": 152257,
|
577 |
+
"[control_599]": 152258,
|
578 |
+
"[control_59]": 151718,
|
579 |
+
"[control_600]": 152259,
|
580 |
+
"[control_601]": 152260,
|
581 |
+
"[control_602]": 152261,
|
582 |
+
"[control_603]": 152262,
|
583 |
+
"[control_604]": 152263,
|
584 |
+
"[control_605]": 152264,
|
585 |
+
"[control_606]": 152265,
|
586 |
+
"[control_607]": 152266,
|
587 |
+
"[control_608]": 152267,
|
588 |
+
"[control_609]": 152268,
|
589 |
+
"[control_60]": 151719,
|
590 |
+
"[control_610]": 152269,
|
591 |
+
"[control_611]": 152270,
|
592 |
+
"[control_612]": 152271,
|
593 |
+
"[control_613]": 152272,
|
594 |
+
"[control_614]": 152273,
|
595 |
+
"[control_615]": 152274,
|
596 |
+
"[control_616]": 152275,
|
597 |
+
"[control_617]": 152276,
|
598 |
+
"[control_618]": 152277,
|
599 |
+
"[control_619]": 152278,
|
600 |
+
"[control_61]": 151720,
|
601 |
+
"[control_620]": 152279,
|
602 |
+
"[control_621]": 152280,
|
603 |
+
"[control_622]": 152281,
|
604 |
+
"[control_623]": 152282,
|
605 |
+
"[control_624]": 152283,
|
606 |
+
"[control_625]": 152284,
|
607 |
+
"[control_626]": 152285,
|
608 |
+
"[control_627]": 152286,
|
609 |
+
"[control_628]": 152287,
|
610 |
+
"[control_629]": 152288,
|
611 |
+
"[control_62]": 151721,
|
612 |
+
"[control_630]": 152289,
|
613 |
+
"[control_631]": 152290,
|
614 |
+
"[control_632]": 152291,
|
615 |
+
"[control_633]": 152292,
|
616 |
+
"[control_634]": 152293,
|
617 |
+
"[control_635]": 152294,
|
618 |
+
"[control_636]": 152295,
|
619 |
+
"[control_637]": 152296,
|
620 |
+
"[control_638]": 152297,
|
621 |
+
"[control_639]": 152298,
|
622 |
+
"[control_63]": 151722,
|
623 |
+
"[control_640]": 152299,
|
624 |
+
"[control_641]": 152300,
|
625 |
+
"[control_642]": 152301,
|
626 |
+
"[control_643]": 152302,
|
627 |
+
"[control_644]": 152303,
|
628 |
+
"[control_645]": 152304,
|
629 |
+
"[control_646]": 152305,
|
630 |
+
"[control_647]": 152306,
|
631 |
+
"[control_648]": 152307,
|
632 |
+
"[control_649]": 152308,
|
633 |
+
"[control_64]": 151723,
|
634 |
+
"[control_650]": 152309,
|
635 |
+
"[control_651]": 152310,
|
636 |
+
"[control_652]": 152311,
|
637 |
+
"[control_653]": 152312,
|
638 |
+
"[control_654]": 152313,
|
639 |
+
"[control_655]": 152314,
|
640 |
+
"[control_656]": 152315,
|
641 |
+
"[control_657]": 152316,
|
642 |
+
"[control_658]": 152317,
|
643 |
+
"[control_659]": 152318,
|
644 |
+
"[control_65]": 151724,
|
645 |
+
"[control_660]": 152319,
|
646 |
+
"[control_661]": 152320,
|
647 |
+
"[control_662]": 152321,
|
648 |
+
"[control_663]": 152322,
|
649 |
+
"[control_664]": 152323,
|
650 |
+
"[control_665]": 152324,
|
651 |
+
"[control_666]": 152325,
|
652 |
+
"[control_667]": 152326,
|
653 |
+
"[control_668]": 152327,
|
654 |
+
"[control_669]": 152328,
|
655 |
+
"[control_66]": 151725,
|
656 |
+
"[control_670]": 152329,
|
657 |
+
"[control_671]": 152330,
|
658 |
+
"[control_672]": 152331,
|
659 |
+
"[control_673]": 152332,
|
660 |
+
"[control_674]": 152333,
|
661 |
+
"[control_675]": 152334,
|
662 |
+
"[control_676]": 152335,
|
663 |
+
"[control_677]": 152336,
|
664 |
+
"[control_678]": 152337,
|
665 |
+
"[control_679]": 152338,
|
666 |
+
"[control_67]": 151726,
|
667 |
+
"[control_680]": 152339,
|
668 |
+
"[control_681]": 152340,
|
669 |
+
"[control_682]": 152341,
|
670 |
+
"[control_683]": 152342,
|
671 |
+
"[control_684]": 152343,
|
672 |
+
"[control_685]": 152344,
|
673 |
+
"[control_686]": 152345,
|
674 |
+
"[control_687]": 152346,
|
675 |
+
"[control_688]": 152347,
|
676 |
+
"[control_689]": 152348,
|
677 |
+
"[control_68]": 151727,
|
678 |
+
"[control_690]": 152349,
|
679 |
+
"[control_691]": 152350,
|
680 |
+
"[control_692]": 152351,
|
681 |
+
"[control_693]": 152352,
|
682 |
+
"[control_694]": 152353,
|
683 |
+
"[control_695]": 152354,
|
684 |
+
"[control_696]": 152355,
|
685 |
+
"[control_697]": 152356,
|
686 |
+
"[control_698]": 152357,
|
687 |
+
"[control_699]": 152358,
|
688 |
+
"[control_69]": 151728,
|
689 |
+
"[control_700]": 152359,
|
690 |
+
"[control_701]": 152360,
|
691 |
+
"[control_702]": 152361,
|
692 |
+
"[control_703]": 152362,
|
693 |
+
"[control_704]": 152363,
|
694 |
+
"[control_705]": 152364,
|
695 |
+
"[control_706]": 152365,
|
696 |
+
"[control_707]": 152366,
|
697 |
+
"[control_708]": 152367,
|
698 |
+
"[control_709]": 152368,
|
699 |
+
"[control_70]": 151729,
|
700 |
+
"[control_710]": 152369,
|
701 |
+
"[control_711]": 152370,
|
702 |
+
"[control_712]": 152371,
|
703 |
+
"[control_713]": 152372,
|
704 |
+
"[control_714]": 152373,
|
705 |
+
"[control_715]": 152374,
|
706 |
+
"[control_716]": 152375,
|
707 |
+
"[control_717]": 152376,
|
708 |
+
"[control_718]": 152377,
|
709 |
+
"[control_719]": 152378,
|
710 |
+
"[control_71]": 151730,
|
711 |
+
"[control_720]": 152379,
|
712 |
+
"[control_721]": 152380,
|
713 |
+
"[control_722]": 152381,
|
714 |
+
"[control_723]": 152382,
|
715 |
+
"[control_724]": 152383,
|
716 |
+
"[control_725]": 152384,
|
717 |
+
"[control_726]": 152385,
|
718 |
+
"[control_727]": 152386,
|
719 |
+
"[control_728]": 152387,
|
720 |
+
"[control_729]": 152388,
|
721 |
+
"[control_72]": 151731,
|
722 |
+
"[control_730]": 152389,
|
723 |
+
"[control_731]": 152390,
|
724 |
+
"[control_732]": 152391,
|
725 |
+
"[control_733]": 152392,
|
726 |
+
"[control_734]": 152393,
|
727 |
+
"[control_735]": 152394,
|
728 |
+
"[control_736]": 152395,
|
729 |
+
"[control_737]": 152396,
|
730 |
+
"[control_738]": 152397,
|
731 |
+
"[control_739]": 152398,
|
732 |
+
"[control_73]": 151732,
|
733 |
+
"[control_740]": 152399,
|
734 |
+
"[control_741]": 152400,
|
735 |
+
"[control_742]": 152401,
|
736 |
+
"[control_743]": 152402,
|
737 |
+
"[control_744]": 152403,
|
738 |
+
"[control_745]": 152404,
|
739 |
+
"[control_746]": 152405,
|
740 |
+
"[control_747]": 152406,
|
741 |
+
"[control_748]": 152407,
|
742 |
+
"[control_749]": 152408,
|
743 |
+
"[control_74]": 151733,
|
744 |
+
"[control_750]": 152409,
|
745 |
+
"[control_751]": 152410,
|
746 |
+
"[control_752]": 152411,
|
747 |
+
"[control_753]": 152412,
|
748 |
+
"[control_754]": 152413,
|
749 |
+
"[control_755]": 152414,
|
750 |
+
"[control_756]": 152415,
|
751 |
+
"[control_757]": 152416,
|
752 |
+
"[control_758]": 152417,
|
753 |
+
"[control_759]": 152418,
|
754 |
+
"[control_75]": 151734,
|
755 |
+
"[control_760]": 152419,
|
756 |
+
"[control_761]": 152420,
|
757 |
+
"[control_762]": 152421,
|
758 |
+
"[control_763]": 152422,
|
759 |
+
"[control_764]": 152423,
|
760 |
+
"[control_765]": 152424,
|
761 |
+
"[control_766]": 152425,
|
762 |
+
"[control_767]": 152426,
|
763 |
+
"[control_768]": 152427,
|
764 |
+
"[control_769]": 152428,
|
765 |
+
"[control_76]": 151735,
|
766 |
+
"[control_770]": 152429,
|
767 |
+
"[control_771]": 152430,
|
768 |
+
"[control_772]": 152431,
|
769 |
+
"[control_773]": 152432,
|
770 |
+
"[control_774]": 152433,
|
771 |
+
"[control_775]": 152434,
|
772 |
+
"[control_776]": 152435,
|
773 |
+
"[control_777]": 152436,
|
774 |
+
"[control_778]": 152437,
|
775 |
+
"[control_779]": 152438,
|
776 |
+
"[control_77]": 151736,
|
777 |
+
"[control_780]": 152439,
|
778 |
+
"[control_781]": 152440,
|
779 |
+
"[control_782]": 152441,
|
780 |
+
"[control_783]": 152442,
|
781 |
+
"[control_784]": 152443,
|
782 |
+
"[control_785]": 152444,
|
783 |
+
"[control_786]": 152445,
|
784 |
+
"[control_787]": 152446,
|
785 |
+
"[control_788]": 152447,
|
786 |
+
"[control_789]": 152448,
|
787 |
+
"[control_78]": 151737,
|
788 |
+
"[control_790]": 152449,
|
789 |
+
"[control_791]": 152450,
|
790 |
+
"[control_792]": 152451,
|
791 |
+
"[control_793]": 152452,
|
792 |
+
"[control_794]": 152453,
|
793 |
+
"[control_795]": 152454,
|
794 |
+
"[control_796]": 152455,
|
795 |
+
"[control_797]": 152456,
|
796 |
+
"[control_798]": 152457,
|
797 |
+
"[control_799]": 152458,
|
798 |
+
"[control_79]": 151738,
|
799 |
+
"[control_800]": 152459,
|
800 |
+
"[control_801]": 152460,
|
801 |
+
"[control_802]": 152461,
|
802 |
+
"[control_803]": 152462,
|
803 |
+
"[control_804]": 152463,
|
804 |
+
"[control_805]": 152464,
|
805 |
+
"[control_806]": 152465,
|
806 |
+
"[control_807]": 152466,
|
807 |
+
"[control_808]": 152467,
|
808 |
+
"[control_809]": 152468,
|
809 |
+
"[control_80]": 151739,
|
810 |
+
"[control_810]": 152469,
|
811 |
+
"[control_811]": 152470,
|
812 |
+
"[control_812]": 152471,
|
813 |
+
"[control_813]": 152472,
|
814 |
+
"[control_814]": 152473,
|
815 |
+
"[control_815]": 152474,
|
816 |
+
"[control_816]": 152475,
|
817 |
+
"[control_817]": 152476,
|
818 |
+
"[control_818]": 152477,
|
819 |
+
"[control_819]": 152478,
|
820 |
+
"[control_81]": 151740,
|
821 |
+
"[control_820]": 152479,
|
822 |
+
"[control_821]": 152480,
|
823 |
+
"[control_822]": 152481,
|
824 |
+
"[control_823]": 152482,
|
825 |
+
"[control_824]": 152483,
|
826 |
+
"[control_825]": 152484,
|
827 |
+
"[control_826]": 152485,
|
828 |
+
"[control_827]": 152486,
|
829 |
+
"[control_828]": 152487,
|
830 |
+
"[control_829]": 152488,
|
831 |
+
"[control_82]": 151741,
|
832 |
+
"[control_830]": 152489,
|
833 |
+
"[control_831]": 152490,
|
834 |
+
"[control_832]": 152491,
|
835 |
+
"[control_833]": 152492,
|
836 |
+
"[control_834]": 152493,
|
837 |
+
"[control_835]": 152494,
|
838 |
+
"[control_836]": 152495,
|
839 |
+
"[control_837]": 152496,
|
840 |
+
"[control_838]": 152497,
|
841 |
+
"[control_839]": 152498,
|
842 |
+
"[control_83]": 151742,
|
843 |
+
"[control_840]": 152499,
|
844 |
+
"[control_841]": 152500,
|
845 |
+
"[control_842]": 152501,
|
846 |
+
"[control_843]": 152502,
|
847 |
+
"[control_844]": 152503,
|
848 |
+
"[control_845]": 152504,
|
849 |
+
"[control_846]": 152505,
|
850 |
+
"[control_847]": 152506,
|
851 |
+
"[control_848]": 152507,
|
852 |
+
"[control_849]": 152508,
|
853 |
+
"[control_84]": 151743,
|
854 |
+
"[control_850]": 152509,
|
855 |
+
"[control_851]": 152510,
|
856 |
+
"[control_852]": 152511,
|
857 |
+
"[control_853]": 152512,
|
858 |
+
"[control_854]": 152513,
|
859 |
+
"[control_855]": 152514,
|
860 |
+
"[control_856]": 152515,
|
861 |
+
"[control_857]": 152516,
|
862 |
+
"[control_858]": 152517,
|
863 |
+
"[control_859]": 152518,
|
864 |
+
"[control_85]": 151744,
|
865 |
+
"[control_860]": 152519,
|
866 |
+
"[control_861]": 152520,
|
867 |
+
"[control_862]": 152521,
|
868 |
+
"[control_863]": 152522,
|
869 |
+
"[control_864]": 152523,
|
870 |
+
"[control_865]": 152524,
|
871 |
+
"[control_866]": 152525,
|
872 |
+
"[control_867]": 152526,
|
873 |
+
"[control_868]": 152527,
|
874 |
+
"[control_869]": 152528,
|
875 |
+
"[control_86]": 151745,
|
876 |
+
"[control_870]": 152529,
|
877 |
+
"[control_871]": 152530,
|
878 |
+
"[control_872]": 152531,
|
879 |
+
"[control_873]": 152532,
|
880 |
+
"[control_874]": 152533,
|
881 |
+
"[control_875]": 152534,
|
882 |
+
"[control_876]": 152535,
|
883 |
+
"[control_877]": 152536,
|
884 |
+
"[control_878]": 152537,
|
885 |
+
"[control_879]": 152538,
|
886 |
+
"[control_87]": 151746,
|
887 |
+
"[control_880]": 152539,
|
888 |
+
"[control_881]": 152540,
|
889 |
+
"[control_882]": 152541,
|
890 |
+
"[control_883]": 152542,
|
891 |
+
"[control_884]": 152543,
|
892 |
+
"[control_885]": 152544,
|
893 |
+
"[control_886]": 152545,
|
894 |
+
"[control_887]": 152546,
|
895 |
+
"[control_888]": 152547,
|
896 |
+
"[control_889]": 152548,
|
897 |
+
"[control_88]": 151747,
|
898 |
+
"[control_890]": 152549,
|
899 |
+
"[control_891]": 152550,
|
900 |
+
"[control_892]": 152551,
|
901 |
+
"[control_893]": 152552,
|
902 |
+
"[control_894]": 152553,
|
903 |
+
"[control_895]": 152554,
|
904 |
+
"[control_896]": 152555,
|
905 |
+
"[control_897]": 152556,
|
906 |
+
"[control_898]": 152557,
|
907 |
+
"[control_899]": 152558,
|
908 |
+
"[control_89]": 151748,
|
909 |
+
"[control_900]": 152559,
|
910 |
+
"[control_901]": 152560,
|
911 |
+
"[control_902]": 152561,
|
912 |
+
"[control_903]": 152562,
|
913 |
+
"[control_904]": 152563,
|
914 |
+
"[control_905]": 152564,
|
915 |
+
"[control_906]": 152565,
|
916 |
+
"[control_907]": 152566,
|
917 |
+
"[control_908]": 152567,
|
918 |
+
"[control_909]": 152568,
|
919 |
+
"[control_90]": 151749,
|
920 |
+
"[control_910]": 152569,
|
921 |
+
"[control_911]": 152570,
|
922 |
+
"[control_912]": 152571,
|
923 |
+
"[control_913]": 152572,
|
924 |
+
"[control_914]": 152573,
|
925 |
+
"[control_915]": 152574,
|
926 |
+
"[control_916]": 152575,
|
927 |
+
"[control_917]": 152576,
|
928 |
+
"[control_918]": 152577,
|
929 |
+
"[control_919]": 152578,
|
930 |
+
"[control_91]": 151750,
|
931 |
+
"[control_920]": 152579,
|
932 |
+
"[control_921]": 152580,
|
933 |
+
"[control_922]": 152581,
|
934 |
+
"[control_923]": 152582,
|
935 |
+
"[control_924]": 152583,
|
936 |
+
"[control_925]": 152584,
|
937 |
+
"[control_926]": 152585,
|
938 |
+
"[control_927]": 152586,
|
939 |
+
"[control_928]": 152587,
|
940 |
+
"[control_929]": 152588,
|
941 |
+
"[control_92]": 151751,
|
942 |
+
"[control_930]": 152589,
|
943 |
+
"[control_931]": 152590,
|
944 |
+
"[control_932]": 152591,
|
945 |
+
"[control_933]": 152592,
|
946 |
+
"[control_934]": 152593,
|
947 |
+
"[control_935]": 152594,
|
948 |
+
"[control_936]": 152595,
|
949 |
+
"[control_937]": 152596,
|
950 |
+
"[control_938]": 152597,
|
951 |
+
"[control_939]": 152598,
|
952 |
+
"[control_93]": 151752,
|
953 |
+
"[control_940]": 152599,
|
954 |
+
"[control_941]": 152600,
|
955 |
+
"[control_942]": 152601,
|
956 |
+
"[control_943]": 152602,
|
957 |
+
"[control_944]": 152603,
|
958 |
+
"[control_945]": 152604,
|
959 |
+
"[control_946]": 152605,
|
960 |
+
"[control_947]": 152606,
|
961 |
+
"[control_948]": 152607,
|
962 |
+
"[control_949]": 152608,
|
963 |
+
"[control_94]": 151753,
|
964 |
+
"[control_950]": 152609,
|
965 |
+
"[control_951]": 152610,
|
966 |
+
"[control_952]": 152611,
|
967 |
+
"[control_953]": 152612,
|
968 |
+
"[control_954]": 152613,
|
969 |
+
"[control_955]": 152614,
|
970 |
+
"[control_956]": 152615,
|
971 |
+
"[control_957]": 152616,
|
972 |
+
"[control_958]": 152617,
|
973 |
+
"[control_959]": 152618,
|
974 |
+
"[control_95]": 151754,
|
975 |
+
"[control_960]": 152619,
|
976 |
+
"[control_961]": 152620,
|
977 |
+
"[control_962]": 152621,
|
978 |
+
"[control_963]": 152622,
|
979 |
+
"[control_964]": 152623,
|
980 |
+
"[control_965]": 152624,
|
981 |
+
"[control_966]": 152625,
|
982 |
+
"[control_967]": 152626,
|
983 |
+
"[control_968]": 152627,
|
984 |
+
"[control_969]": 152628,
|
985 |
+
"[control_96]": 151755,
|
986 |
+
"[control_970]": 152629,
|
987 |
+
"[control_971]": 152630,
|
988 |
+
"[control_972]": 152631,
|
989 |
+
"[control_973]": 152632,
|
990 |
+
"[control_974]": 152633,
|
991 |
+
"[control_975]": 152634,
|
992 |
+
"[control_976]": 152635,
|
993 |
+
"[control_977]": 152636,
|
994 |
+
"[control_978]": 152637,
|
995 |
+
"[control_979]": 152638,
|
996 |
+
"[control_97]": 151756,
|
997 |
+
"[control_980]": 152639,
|
998 |
+
"[control_981]": 152640,
|
999 |
+
"[control_982]": 152641,
|
1000 |
+
"[control_983]": 152642,
|
1001 |
+
"[control_984]": 152643,
|
1002 |
+
"[control_985]": 152644,
|
1003 |
+
"[control_986]": 152645,
|
1004 |
+
"[control_987]": 152646,
|
1005 |
+
"[control_988]": 152647,
|
1006 |
+
"[control_989]": 152648,
|
1007 |
+
"[control_98]": 151757,
|
1008 |
+
"[control_990]": 152649,
|
1009 |
+
"[control_991]": 152650,
|
1010 |
+
"[control_992]": 152651,
|
1011 |
+
"[control_993]": 152652,
|
1012 |
+
"[control_994]": 152653,
|
1013 |
+
"[control_995]": 152654,
|
1014 |
+
"[control_996]": 152655,
|
1015 |
+
"[control_997]": 152656,
|
1016 |
+
"[control_998]": 152657,
|
1017 |
+
"[control_999]": 152658,
|
1018 |
+
"[control_99]": 151758
|
1019 |
+
}
|
config.json
ADDED
@@ -0,0 +1,72 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"architectures": [
|
3 |
+
"Qwen3ForCausalLM"
|
4 |
+
],
|
5 |
+
"attention_bias": false,
|
6 |
+
"attention_dropout": 0.0,
|
7 |
+
"bos_token_id": 151643,
|
8 |
+
"eos_token_id": 151645,
|
9 |
+
"head_dim": 128,
|
10 |
+
"hidden_act": "silu",
|
11 |
+
"hidden_size": 5120,
|
12 |
+
"initializer_range": 0.02,
|
13 |
+
"intermediate_size": 17408,
|
14 |
+
"max_position_embeddings": 40960,
|
15 |
+
"max_window_layers": 40,
|
16 |
+
"model_type": "qwen3",
|
17 |
+
"num_attention_heads": 40,
|
18 |
+
"num_hidden_layers": 40,
|
19 |
+
"num_key_value_heads": 8,
|
20 |
+
"quantization_config": {
|
21 |
+
"config_groups": {
|
22 |
+
"group_0": {
|
23 |
+
"input_activations": {
|
24 |
+
"actorder": null,
|
25 |
+
"block_structure": null,
|
26 |
+
"dynamic": true,
|
27 |
+
"group_size": null,
|
28 |
+
"num_bits": 8,
|
29 |
+
"observer": null,
|
30 |
+
"observer_kwargs": {},
|
31 |
+
"strategy": "token",
|
32 |
+
"symmetric": true,
|
33 |
+
"type": "float"
|
34 |
+
},
|
35 |
+
"output_activations": null,
|
36 |
+
"targets": [
|
37 |
+
"Linear"
|
38 |
+
],
|
39 |
+
"weights": {
|
40 |
+
"actorder": null,
|
41 |
+
"block_structure": null,
|
42 |
+
"dynamic": false,
|
43 |
+
"group_size": null,
|
44 |
+
"num_bits": 8,
|
45 |
+
"observer": "minmax",
|
46 |
+
"observer_kwargs": {},
|
47 |
+
"strategy": "channel",
|
48 |
+
"symmetric": true,
|
49 |
+
"type": "float"
|
50 |
+
}
|
51 |
+
}
|
52 |
+
},
|
53 |
+
"format": "float-quantized",
|
54 |
+
"global_compression_ratio": null,
|
55 |
+
"ignore": [
|
56 |
+
"lm_head"
|
57 |
+
],
|
58 |
+
"kv_cache_scheme": null,
|
59 |
+
"quant_method": "compressed-tensors",
|
60 |
+
"quantization_status": "compressed"
|
61 |
+
},
|
62 |
+
"rms_norm_eps": 1e-06,
|
63 |
+
"rope_scaling": null,
|
64 |
+
"rope_theta": 1000000,
|
65 |
+
"sliding_window": null,
|
66 |
+
"tie_word_embeddings": false,
|
67 |
+
"torch_dtype": "bfloat16",
|
68 |
+
"transformers_version": "4.51.1",
|
69 |
+
"use_cache": true,
|
70 |
+
"use_sliding_window": false,
|
71 |
+
"vocab_size": 152660
|
72 |
+
}
|
generation_config.json
ADDED
@@ -0,0 +1,13 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"bos_token_id": 151643,
|
3 |
+
"do_sample": true,
|
4 |
+
"eos_token_id": [
|
5 |
+
151645,
|
6 |
+
151643
|
7 |
+
],
|
8 |
+
"pad_token_id": 151643,
|
9 |
+
"temperature": 0.6,
|
10 |
+
"top_k": 20,
|
11 |
+
"top_p": 0.95,
|
12 |
+
"transformers_version": "4.51.1"
|
13 |
+
}
|
merges.txt
ADDED
The diff for this file is too large to render.
See raw diff
|
|
model-00001-of-00004.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:dfc4b697f085fa8bd6d1d054bccf736b9f082e233f0344730e56ff3c822de4ff
|
3 |
+
size 4930469344
|
model-00002-of-00004.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:4228280b53960b7d0eab62f2729dbcbd50e266ba97e2a47f5ede4a7b8aa4f131
|
3 |
+
size 4956434040
|
model-00003-of-00004.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:8cc8d1187910fed69f6eac3734cf064cd194ef68345e1f3fca7e7ec8807e04be
|
3 |
+
size 4893503560
|
model-00004-of-00004.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:be72e24d1ea5dd8a7b8b1c03dfc877b2713b56e4d8697bd7541eaac55077d07d
|
3 |
+
size 1563238528
|
model.safetensors.index.json
ADDED
@@ -0,0 +1,730 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"metadata": {
|
3 |
+
"total_size": 16343562240
|
4 |
+
},
|
5 |
+
"weight_map": {
|
6 |
+
"lm_head.weight": "model-00004-of-00004.safetensors",
|
7 |
+
"model.embed_tokens.weight": "model-00001-of-00004.safetensors",
|
8 |
+
"model.layers.0.input_layernorm.weight": "model-00001-of-00004.safetensors",
|
9 |
+
"model.layers.0.mlp.down_proj.weight": "model-00001-of-00004.safetensors",
|
10 |
+
"model.layers.0.mlp.down_proj.weight_scale": "model-00001-of-00004.safetensors",
|
11 |
+
"model.layers.0.mlp.gate_proj.weight": "model-00001-of-00004.safetensors",
|
12 |
+
"model.layers.0.mlp.gate_proj.weight_scale": "model-00001-of-00004.safetensors",
|
13 |
+
"model.layers.0.mlp.up_proj.weight": "model-00001-of-00004.safetensors",
|
14 |
+
"model.layers.0.mlp.up_proj.weight_scale": "model-00001-of-00004.safetensors",
|
15 |
+
"model.layers.0.post_attention_layernorm.weight": "model-00001-of-00004.safetensors",
|
16 |
+
"model.layers.0.self_attn.k_norm.weight": "model-00001-of-00004.safetensors",
|
17 |
+
"model.layers.0.self_attn.k_proj.weight": "model-00001-of-00004.safetensors",
|
18 |
+
"model.layers.0.self_attn.k_proj.weight_scale": "model-00001-of-00004.safetensors",
|
19 |
+
"model.layers.0.self_attn.o_proj.weight": "model-00001-of-00004.safetensors",
|
20 |
+
"model.layers.0.self_attn.o_proj.weight_scale": "model-00001-of-00004.safetensors",
|
21 |
+
"model.layers.0.self_attn.q_norm.weight": "model-00001-of-00004.safetensors",
|
22 |
+
"model.layers.0.self_attn.q_proj.weight": "model-00001-of-00004.safetensors",
|
23 |
+
"model.layers.0.self_attn.q_proj.weight_scale": "model-00001-of-00004.safetensors",
|
24 |
+
"model.layers.0.self_attn.v_proj.weight": "model-00001-of-00004.safetensors",
|
25 |
+
"model.layers.0.self_attn.v_proj.weight_scale": "model-00001-of-00004.safetensors",
|
26 |
+
"model.layers.1.input_layernorm.weight": "model-00001-of-00004.safetensors",
|
27 |
+
"model.layers.1.mlp.down_proj.weight": "model-00001-of-00004.safetensors",
|
28 |
+
"model.layers.1.mlp.down_proj.weight_scale": "model-00001-of-00004.safetensors",
|
29 |
+
"model.layers.1.mlp.gate_proj.weight": "model-00001-of-00004.safetensors",
|
30 |
+
"model.layers.1.mlp.gate_proj.weight_scale": "model-00001-of-00004.safetensors",
|
31 |
+
"model.layers.1.mlp.up_proj.weight": "model-00001-of-00004.safetensors",
|
32 |
+
"model.layers.1.mlp.up_proj.weight_scale": "model-00001-of-00004.safetensors",
|
33 |
+
"model.layers.1.post_attention_layernorm.weight": "model-00001-of-00004.safetensors",
|
34 |
+
"model.layers.1.self_attn.k_norm.weight": "model-00001-of-00004.safetensors",
|
35 |
+
"model.layers.1.self_attn.k_proj.weight": "model-00001-of-00004.safetensors",
|
36 |
+
"model.layers.1.self_attn.k_proj.weight_scale": "model-00001-of-00004.safetensors",
|
37 |
+
"model.layers.1.self_attn.o_proj.weight": "model-00001-of-00004.safetensors",
|
38 |
+
"model.layers.1.self_attn.o_proj.weight_scale": "model-00001-of-00004.safetensors",
|
39 |
+
"model.layers.1.self_attn.q_norm.weight": "model-00001-of-00004.safetensors",
|
40 |
+
"model.layers.1.self_attn.q_proj.weight": "model-00001-of-00004.safetensors",
|
41 |
+
"model.layers.1.self_attn.q_proj.weight_scale": "model-00001-of-00004.safetensors",
|
42 |
+
"model.layers.1.self_attn.v_proj.weight": "model-00001-of-00004.safetensors",
|
43 |
+
"model.layers.1.self_attn.v_proj.weight_scale": "model-00001-of-00004.safetensors",
|
44 |
+
"model.layers.10.input_layernorm.weight": "model-00002-of-00004.safetensors",
|
45 |
+
"model.layers.10.mlp.down_proj.weight": "model-00002-of-00004.safetensors",
|
46 |
+
"model.layers.10.mlp.down_proj.weight_scale": "model-00002-of-00004.safetensors",
|
47 |
+
"model.layers.10.mlp.gate_proj.weight": "model-00002-of-00004.safetensors",
|
48 |
+
"model.layers.10.mlp.gate_proj.weight_scale": "model-00002-of-00004.safetensors",
|
49 |
+
"model.layers.10.mlp.up_proj.weight": "model-00002-of-00004.safetensors",
|
50 |
+
"model.layers.10.mlp.up_proj.weight_scale": "model-00002-of-00004.safetensors",
|
51 |
+
"model.layers.10.post_attention_layernorm.weight": "model-00002-of-00004.safetensors",
|
52 |
+
"model.layers.10.self_attn.k_norm.weight": "model-00001-of-00004.safetensors",
|
53 |
+
"model.layers.10.self_attn.k_proj.weight": "model-00001-of-00004.safetensors",
|
54 |
+
"model.layers.10.self_attn.k_proj.weight_scale": "model-00001-of-00004.safetensors",
|
55 |
+
"model.layers.10.self_attn.o_proj.weight": "model-00001-of-00004.safetensors",
|
56 |
+
"model.layers.10.self_attn.o_proj.weight_scale": "model-00001-of-00004.safetensors",
|
57 |
+
"model.layers.10.self_attn.q_norm.weight": "model-00001-of-00004.safetensors",
|
58 |
+
"model.layers.10.self_attn.q_proj.weight": "model-00001-of-00004.safetensors",
|
59 |
+
"model.layers.10.self_attn.q_proj.weight_scale": "model-00001-of-00004.safetensors",
|
60 |
+
"model.layers.10.self_attn.v_proj.weight": "model-00001-of-00004.safetensors",
|
61 |
+
"model.layers.10.self_attn.v_proj.weight_scale": "model-00001-of-00004.safetensors",
|
62 |
+
"model.layers.11.input_layernorm.weight": "model-00002-of-00004.safetensors",
|
63 |
+
"model.layers.11.mlp.down_proj.weight": "model-00002-of-00004.safetensors",
|
64 |
+
"model.layers.11.mlp.down_proj.weight_scale": "model-00002-of-00004.safetensors",
|
65 |
+
"model.layers.11.mlp.gate_proj.weight": "model-00002-of-00004.safetensors",
|
66 |
+
"model.layers.11.mlp.gate_proj.weight_scale": "model-00002-of-00004.safetensors",
|
67 |
+
"model.layers.11.mlp.up_proj.weight": "model-00002-of-00004.safetensors",
|
68 |
+
"model.layers.11.mlp.up_proj.weight_scale": "model-00002-of-00004.safetensors",
|
69 |
+
"model.layers.11.post_attention_layernorm.weight": "model-00002-of-00004.safetensors",
|
70 |
+
"model.layers.11.self_attn.k_norm.weight": "model-00002-of-00004.safetensors",
|
71 |
+
"model.layers.11.self_attn.k_proj.weight": "model-00002-of-00004.safetensors",
|
72 |
+
"model.layers.11.self_attn.k_proj.weight_scale": "model-00002-of-00004.safetensors",
|
73 |
+
"model.layers.11.self_attn.o_proj.weight": "model-00002-of-00004.safetensors",
|
74 |
+
"model.layers.11.self_attn.o_proj.weight_scale": "model-00002-of-00004.safetensors",
|
75 |
+
"model.layers.11.self_attn.q_norm.weight": "model-00002-of-00004.safetensors",
|
76 |
+
"model.layers.11.self_attn.q_proj.weight": "model-00002-of-00004.safetensors",
|
77 |
+
"model.layers.11.self_attn.q_proj.weight_scale": "model-00002-of-00004.safetensors",
|
78 |
+
"model.layers.11.self_attn.v_proj.weight": "model-00002-of-00004.safetensors",
|
79 |
+
"model.layers.11.self_attn.v_proj.weight_scale": "model-00002-of-00004.safetensors",
|
80 |
+
"model.layers.12.input_layernorm.weight": "model-00002-of-00004.safetensors",
|
81 |
+
"model.layers.12.mlp.down_proj.weight": "model-00002-of-00004.safetensors",
|
82 |
+
"model.layers.12.mlp.down_proj.weight_scale": "model-00002-of-00004.safetensors",
|
83 |
+
"model.layers.12.mlp.gate_proj.weight": "model-00002-of-00004.safetensors",
|
84 |
+
"model.layers.12.mlp.gate_proj.weight_scale": "model-00002-of-00004.safetensors",
|
85 |
+
"model.layers.12.mlp.up_proj.weight": "model-00002-of-00004.safetensors",
|
86 |
+
"model.layers.12.mlp.up_proj.weight_scale": "model-00002-of-00004.safetensors",
|
87 |
+
"model.layers.12.post_attention_layernorm.weight": "model-00002-of-00004.safetensors",
|
88 |
+
"model.layers.12.self_attn.k_norm.weight": "model-00002-of-00004.safetensors",
|
89 |
+
"model.layers.12.self_attn.k_proj.weight": "model-00002-of-00004.safetensors",
|
90 |
+
"model.layers.12.self_attn.k_proj.weight_scale": "model-00002-of-00004.safetensors",
|
91 |
+
"model.layers.12.self_attn.o_proj.weight": "model-00002-of-00004.safetensors",
|
92 |
+
"model.layers.12.self_attn.o_proj.weight_scale": "model-00002-of-00004.safetensors",
|
93 |
+
"model.layers.12.self_attn.q_norm.weight": "model-00002-of-00004.safetensors",
|
94 |
+
"model.layers.12.self_attn.q_proj.weight": "model-00002-of-00004.safetensors",
|
95 |
+
"model.layers.12.self_attn.q_proj.weight_scale": "model-00002-of-00004.safetensors",
|
96 |
+
"model.layers.12.self_attn.v_proj.weight": "model-00002-of-00004.safetensors",
|
97 |
+
"model.layers.12.self_attn.v_proj.weight_scale": "model-00002-of-00004.safetensors",
|
98 |
+
"model.layers.13.input_layernorm.weight": "model-00002-of-00004.safetensors",
|
99 |
+
"model.layers.13.mlp.down_proj.weight": "model-00002-of-00004.safetensors",
|
100 |
+
"model.layers.13.mlp.down_proj.weight_scale": "model-00002-of-00004.safetensors",
|
101 |
+
"model.layers.13.mlp.gate_proj.weight": "model-00002-of-00004.safetensors",
|
102 |
+
"model.layers.13.mlp.gate_proj.weight_scale": "model-00002-of-00004.safetensors",
|
103 |
+
"model.layers.13.mlp.up_proj.weight": "model-00002-of-00004.safetensors",
|
104 |
+
"model.layers.13.mlp.up_proj.weight_scale": "model-00002-of-00004.safetensors",
|
105 |
+
"model.layers.13.post_attention_layernorm.weight": "model-00002-of-00004.safetensors",
|
106 |
+
"model.layers.13.self_attn.k_norm.weight": "model-00002-of-00004.safetensors",
|
107 |
+
"model.layers.13.self_attn.k_proj.weight": "model-00002-of-00004.safetensors",
|
108 |
+
"model.layers.13.self_attn.k_proj.weight_scale": "model-00002-of-00004.safetensors",
|
109 |
+
"model.layers.13.self_attn.o_proj.weight": "model-00002-of-00004.safetensors",
|
110 |
+
"model.layers.13.self_attn.o_proj.weight_scale": "model-00002-of-00004.safetensors",
|
111 |
+
"model.layers.13.self_attn.q_norm.weight": "model-00002-of-00004.safetensors",
|
112 |
+
"model.layers.13.self_attn.q_proj.weight": "model-00002-of-00004.safetensors",
|
113 |
+
"model.layers.13.self_attn.q_proj.weight_scale": "model-00002-of-00004.safetensors",
|
114 |
+
"model.layers.13.self_attn.v_proj.weight": "model-00002-of-00004.safetensors",
|
115 |
+
"model.layers.13.self_attn.v_proj.weight_scale": "model-00002-of-00004.safetensors",
|
116 |
+
"model.layers.14.input_layernorm.weight": "model-00002-of-00004.safetensors",
|
117 |
+
"model.layers.14.mlp.down_proj.weight": "model-00002-of-00004.safetensors",
|
118 |
+
"model.layers.14.mlp.down_proj.weight_scale": "model-00002-of-00004.safetensors",
|
119 |
+
"model.layers.14.mlp.gate_proj.weight": "model-00002-of-00004.safetensors",
|
120 |
+
"model.layers.14.mlp.gate_proj.weight_scale": "model-00002-of-00004.safetensors",
|
121 |
+
"model.layers.14.mlp.up_proj.weight": "model-00002-of-00004.safetensors",
|
122 |
+
"model.layers.14.mlp.up_proj.weight_scale": "model-00002-of-00004.safetensors",
|
123 |
+
"model.layers.14.post_attention_layernorm.weight": "model-00002-of-00004.safetensors",
|
124 |
+
"model.layers.14.self_attn.k_norm.weight": "model-00002-of-00004.safetensors",
|
125 |
+
"model.layers.14.self_attn.k_proj.weight": "model-00002-of-00004.safetensors",
|
126 |
+
"model.layers.14.self_attn.k_proj.weight_scale": "model-00002-of-00004.safetensors",
|
127 |
+
"model.layers.14.self_attn.o_proj.weight": "model-00002-of-00004.safetensors",
|
128 |
+
"model.layers.14.self_attn.o_proj.weight_scale": "model-00002-of-00004.safetensors",
|
129 |
+
"model.layers.14.self_attn.q_norm.weight": "model-00002-of-00004.safetensors",
|
130 |
+
"model.layers.14.self_attn.q_proj.weight": "model-00002-of-00004.safetensors",
|
131 |
+
"model.layers.14.self_attn.q_proj.weight_scale": "model-00002-of-00004.safetensors",
|
132 |
+
"model.layers.14.self_attn.v_proj.weight": "model-00002-of-00004.safetensors",
|
133 |
+
"model.layers.14.self_attn.v_proj.weight_scale": "model-00002-of-00004.safetensors",
|
134 |
+
"model.layers.15.input_layernorm.weight": "model-00002-of-00004.safetensors",
|
135 |
+
"model.layers.15.mlp.down_proj.weight": "model-00002-of-00004.safetensors",
|
136 |
+
"model.layers.15.mlp.down_proj.weight_scale": "model-00002-of-00004.safetensors",
|
137 |
+
"model.layers.15.mlp.gate_proj.weight": "model-00002-of-00004.safetensors",
|
138 |
+
"model.layers.15.mlp.gate_proj.weight_scale": "model-00002-of-00004.safetensors",
|
139 |
+
"model.layers.15.mlp.up_proj.weight": "model-00002-of-00004.safetensors",
|
140 |
+
"model.layers.15.mlp.up_proj.weight_scale": "model-00002-of-00004.safetensors",
|
141 |
+
"model.layers.15.post_attention_layernorm.weight": "model-00002-of-00004.safetensors",
|
142 |
+
"model.layers.15.self_attn.k_norm.weight": "model-00002-of-00004.safetensors",
|
143 |
+
"model.layers.15.self_attn.k_proj.weight": "model-00002-of-00004.safetensors",
|
144 |
+
"model.layers.15.self_attn.k_proj.weight_scale": "model-00002-of-00004.safetensors",
|
145 |
+
"model.layers.15.self_attn.o_proj.weight": "model-00002-of-00004.safetensors",
|
146 |
+
"model.layers.15.self_attn.o_proj.weight_scale": "model-00002-of-00004.safetensors",
|
147 |
+
"model.layers.15.self_attn.q_norm.weight": "model-00002-of-00004.safetensors",
|
148 |
+
"model.layers.15.self_attn.q_proj.weight": "model-00002-of-00004.safetensors",
|
149 |
+
"model.layers.15.self_attn.q_proj.weight_scale": "model-00002-of-00004.safetensors",
|
150 |
+
"model.layers.15.self_attn.v_proj.weight": "model-00002-of-00004.safetensors",
|
151 |
+
"model.layers.15.self_attn.v_proj.weight_scale": "model-00002-of-00004.safetensors",
|
152 |
+
"model.layers.16.input_layernorm.weight": "model-00002-of-00004.safetensors",
|
153 |
+
"model.layers.16.mlp.down_proj.weight": "model-00002-of-00004.safetensors",
|
154 |
+
"model.layers.16.mlp.down_proj.weight_scale": "model-00002-of-00004.safetensors",
|
155 |
+
"model.layers.16.mlp.gate_proj.weight": "model-00002-of-00004.safetensors",
|
156 |
+
"model.layers.16.mlp.gate_proj.weight_scale": "model-00002-of-00004.safetensors",
|
157 |
+
"model.layers.16.mlp.up_proj.weight": "model-00002-of-00004.safetensors",
|
158 |
+
"model.layers.16.mlp.up_proj.weight_scale": "model-00002-of-00004.safetensors",
|
159 |
+
"model.layers.16.post_attention_layernorm.weight": "model-00002-of-00004.safetensors",
|
160 |
+
"model.layers.16.self_attn.k_norm.weight": "model-00002-of-00004.safetensors",
|
161 |
+
"model.layers.16.self_attn.k_proj.weight": "model-00002-of-00004.safetensors",
|
162 |
+
"model.layers.16.self_attn.k_proj.weight_scale": "model-00002-of-00004.safetensors",
|
163 |
+
"model.layers.16.self_attn.o_proj.weight": "model-00002-of-00004.safetensors",
|
164 |
+
"model.layers.16.self_attn.o_proj.weight_scale": "model-00002-of-00004.safetensors",
|
165 |
+
"model.layers.16.self_attn.q_norm.weight": "model-00002-of-00004.safetensors",
|
166 |
+
"model.layers.16.self_attn.q_proj.weight": "model-00002-of-00004.safetensors",
|
167 |
+
"model.layers.16.self_attn.q_proj.weight_scale": "model-00002-of-00004.safetensors",
|
168 |
+
"model.layers.16.self_attn.v_proj.weight": "model-00002-of-00004.safetensors",
|
169 |
+
"model.layers.16.self_attn.v_proj.weight_scale": "model-00002-of-00004.safetensors",
|
170 |
+
"model.layers.17.input_layernorm.weight": "model-00002-of-00004.safetensors",
|
171 |
+
"model.layers.17.mlp.down_proj.weight": "model-00002-of-00004.safetensors",
|
172 |
+
"model.layers.17.mlp.down_proj.weight_scale": "model-00002-of-00004.safetensors",
|
173 |
+
"model.layers.17.mlp.gate_proj.weight": "model-00002-of-00004.safetensors",
|
174 |
+
"model.layers.17.mlp.gate_proj.weight_scale": "model-00002-of-00004.safetensors",
|
175 |
+
"model.layers.17.mlp.up_proj.weight": "model-00002-of-00004.safetensors",
|
176 |
+
"model.layers.17.mlp.up_proj.weight_scale": "model-00002-of-00004.safetensors",
|
177 |
+
"model.layers.17.post_attention_layernorm.weight": "model-00002-of-00004.safetensors",
|
178 |
+
"model.layers.17.self_attn.k_norm.weight": "model-00002-of-00004.safetensors",
|
179 |
+
"model.layers.17.self_attn.k_proj.weight": "model-00002-of-00004.safetensors",
|
180 |
+
"model.layers.17.self_attn.k_proj.weight_scale": "model-00002-of-00004.safetensors",
|
181 |
+
"model.layers.17.self_attn.o_proj.weight": "model-00002-of-00004.safetensors",
|
182 |
+
"model.layers.17.self_attn.o_proj.weight_scale": "model-00002-of-00004.safetensors",
|
183 |
+
"model.layers.17.self_attn.q_norm.weight": "model-00002-of-00004.safetensors",
|
184 |
+
"model.layers.17.self_attn.q_proj.weight": "model-00002-of-00004.safetensors",
|
185 |
+
"model.layers.17.self_attn.q_proj.weight_scale": "model-00002-of-00004.safetensors",
|
186 |
+
"model.layers.17.self_attn.v_proj.weight": "model-00002-of-00004.safetensors",
|
187 |
+
"model.layers.17.self_attn.v_proj.weight_scale": "model-00002-of-00004.safetensors",
|
188 |
+
"model.layers.18.input_layernorm.weight": "model-00002-of-00004.safetensors",
|
189 |
+
"model.layers.18.mlp.down_proj.weight": "model-00002-of-00004.safetensors",
|
190 |
+
"model.layers.18.mlp.down_proj.weight_scale": "model-00002-of-00004.safetensors",
|
191 |
+
"model.layers.18.mlp.gate_proj.weight": "model-00002-of-00004.safetensors",
|
192 |
+
"model.layers.18.mlp.gate_proj.weight_scale": "model-00002-of-00004.safetensors",
|
193 |
+
"model.layers.18.mlp.up_proj.weight": "model-00002-of-00004.safetensors",
|
194 |
+
"model.layers.18.mlp.up_proj.weight_scale": "model-00002-of-00004.safetensors",
|
195 |
+
"model.layers.18.post_attention_layernorm.weight": "model-00002-of-00004.safetensors",
|
196 |
+
"model.layers.18.self_attn.k_norm.weight": "model-00002-of-00004.safetensors",
|
197 |
+
"model.layers.18.self_attn.k_proj.weight": "model-00002-of-00004.safetensors",
|
198 |
+
"model.layers.18.self_attn.k_proj.weight_scale": "model-00002-of-00004.safetensors",
|
199 |
+
"model.layers.18.self_attn.o_proj.weight": "model-00002-of-00004.safetensors",
|
200 |
+
"model.layers.18.self_attn.o_proj.weight_scale": "model-00002-of-00004.safetensors",
|
201 |
+
"model.layers.18.self_attn.q_norm.weight": "model-00002-of-00004.safetensors",
|
202 |
+
"model.layers.18.self_attn.q_proj.weight": "model-00002-of-00004.safetensors",
|
203 |
+
"model.layers.18.self_attn.q_proj.weight_scale": "model-00002-of-00004.safetensors",
|
204 |
+
"model.layers.18.self_attn.v_proj.weight": "model-00002-of-00004.safetensors",
|
205 |
+
"model.layers.18.self_attn.v_proj.weight_scale": "model-00002-of-00004.safetensors",
|
206 |
+
"model.layers.19.input_layernorm.weight": "model-00002-of-00004.safetensors",
|
207 |
+
"model.layers.19.mlp.down_proj.weight": "model-00002-of-00004.safetensors",
|
208 |
+
"model.layers.19.mlp.down_proj.weight_scale": "model-00002-of-00004.safetensors",
|
209 |
+
"model.layers.19.mlp.gate_proj.weight": "model-00002-of-00004.safetensors",
|
210 |
+
"model.layers.19.mlp.gate_proj.weight_scale": "model-00002-of-00004.safetensors",
|
211 |
+
"model.layers.19.mlp.up_proj.weight": "model-00002-of-00004.safetensors",
|
212 |
+
"model.layers.19.mlp.up_proj.weight_scale": "model-00002-of-00004.safetensors",
|
213 |
+
"model.layers.19.post_attention_layernorm.weight": "model-00002-of-00004.safetensors",
|
214 |
+
"model.layers.19.self_attn.k_norm.weight": "model-00002-of-00004.safetensors",
|
215 |
+
"model.layers.19.self_attn.k_proj.weight": "model-00002-of-00004.safetensors",
|
216 |
+
"model.layers.19.self_attn.k_proj.weight_scale": "model-00002-of-00004.safetensors",
|
217 |
+
"model.layers.19.self_attn.o_proj.weight": "model-00002-of-00004.safetensors",
|
218 |
+
"model.layers.19.self_attn.o_proj.weight_scale": "model-00002-of-00004.safetensors",
|
219 |
+
"model.layers.19.self_attn.q_norm.weight": "model-00002-of-00004.safetensors",
|
220 |
+
"model.layers.19.self_attn.q_proj.weight": "model-00002-of-00004.safetensors",
|
221 |
+
"model.layers.19.self_attn.q_proj.weight_scale": "model-00002-of-00004.safetensors",
|
222 |
+
"model.layers.19.self_attn.v_proj.weight": "model-00002-of-00004.safetensors",
|
223 |
+
"model.layers.19.self_attn.v_proj.weight_scale": "model-00002-of-00004.safetensors",
|
224 |
+
"model.layers.2.input_layernorm.weight": "model-00001-of-00004.safetensors",
|
225 |
+
"model.layers.2.mlp.down_proj.weight": "model-00001-of-00004.safetensors",
|
226 |
+
"model.layers.2.mlp.down_proj.weight_scale": "model-00001-of-00004.safetensors",
|
227 |
+
"model.layers.2.mlp.gate_proj.weight": "model-00001-of-00004.safetensors",
|
228 |
+
"model.layers.2.mlp.gate_proj.weight_scale": "model-00001-of-00004.safetensors",
|
229 |
+
"model.layers.2.mlp.up_proj.weight": "model-00001-of-00004.safetensors",
|
230 |
+
"model.layers.2.mlp.up_proj.weight_scale": "model-00001-of-00004.safetensors",
|
231 |
+
"model.layers.2.post_attention_layernorm.weight": "model-00001-of-00004.safetensors",
|
232 |
+
"model.layers.2.self_attn.k_norm.weight": "model-00001-of-00004.safetensors",
|
233 |
+
"model.layers.2.self_attn.k_proj.weight": "model-00001-of-00004.safetensors",
|
234 |
+
"model.layers.2.self_attn.k_proj.weight_scale": "model-00001-of-00004.safetensors",
|
235 |
+
"model.layers.2.self_attn.o_proj.weight": "model-00001-of-00004.safetensors",
|
236 |
+
"model.layers.2.self_attn.o_proj.weight_scale": "model-00001-of-00004.safetensors",
|
237 |
+
"model.layers.2.self_attn.q_norm.weight": "model-00001-of-00004.safetensors",
|
238 |
+
"model.layers.2.self_attn.q_proj.weight": "model-00001-of-00004.safetensors",
|
239 |
+
"model.layers.2.self_attn.q_proj.weight_scale": "model-00001-of-00004.safetensors",
|
240 |
+
"model.layers.2.self_attn.v_proj.weight": "model-00001-of-00004.safetensors",
|
241 |
+
"model.layers.2.self_attn.v_proj.weight_scale": "model-00001-of-00004.safetensors",
|
242 |
+
"model.layers.20.input_layernorm.weight": "model-00002-of-00004.safetensors",
|
243 |
+
"model.layers.20.mlp.down_proj.weight": "model-00002-of-00004.safetensors",
|
244 |
+
"model.layers.20.mlp.down_proj.weight_scale": "model-00002-of-00004.safetensors",
|
245 |
+
"model.layers.20.mlp.gate_proj.weight": "model-00002-of-00004.safetensors",
|
246 |
+
"model.layers.20.mlp.gate_proj.weight_scale": "model-00002-of-00004.safetensors",
|
247 |
+
"model.layers.20.mlp.up_proj.weight": "model-00002-of-00004.safetensors",
|
248 |
+
"model.layers.20.mlp.up_proj.weight_scale": "model-00002-of-00004.safetensors",
|
249 |
+
"model.layers.20.post_attention_layernorm.weight": "model-00002-of-00004.safetensors",
|
250 |
+
"model.layers.20.self_attn.k_norm.weight": "model-00002-of-00004.safetensors",
|
251 |
+
"model.layers.20.self_attn.k_proj.weight": "model-00002-of-00004.safetensors",
|
252 |
+
"model.layers.20.self_attn.k_proj.weight_scale": "model-00002-of-00004.safetensors",
|
253 |
+
"model.layers.20.self_attn.o_proj.weight": "model-00002-of-00004.safetensors",
|
254 |
+
"model.layers.20.self_attn.o_proj.weight_scale": "model-00002-of-00004.safetensors",
|
255 |
+
"model.layers.20.self_attn.q_norm.weight": "model-00002-of-00004.safetensors",
|
256 |
+
"model.layers.20.self_attn.q_proj.weight": "model-00002-of-00004.safetensors",
|
257 |
+
"model.layers.20.self_attn.q_proj.weight_scale": "model-00002-of-00004.safetensors",
|
258 |
+
"model.layers.20.self_attn.v_proj.weight": "model-00002-of-00004.safetensors",
|
259 |
+
"model.layers.20.self_attn.v_proj.weight_scale": "model-00002-of-00004.safetensors",
|
260 |
+
"model.layers.21.input_layernorm.weight": "model-00002-of-00004.safetensors",
|
261 |
+
"model.layers.21.mlp.down_proj.weight": "model-00002-of-00004.safetensors",
|
262 |
+
"model.layers.21.mlp.down_proj.weight_scale": "model-00002-of-00004.safetensors",
|
263 |
+
"model.layers.21.mlp.gate_proj.weight": "model-00002-of-00004.safetensors",
|
264 |
+
"model.layers.21.mlp.gate_proj.weight_scale": "model-00002-of-00004.safetensors",
|
265 |
+
"model.layers.21.mlp.up_proj.weight": "model-00002-of-00004.safetensors",
|
266 |
+
"model.layers.21.mlp.up_proj.weight_scale": "model-00002-of-00004.safetensors",
|
267 |
+
"model.layers.21.post_attention_layernorm.weight": "model-00002-of-00004.safetensors",
|
268 |
+
"model.layers.21.self_attn.k_norm.weight": "model-00002-of-00004.safetensors",
|
269 |
+
"model.layers.21.self_attn.k_proj.weight": "model-00002-of-00004.safetensors",
|
270 |
+
"model.layers.21.self_attn.k_proj.weight_scale": "model-00002-of-00004.safetensors",
|
271 |
+
"model.layers.21.self_attn.o_proj.weight": "model-00002-of-00004.safetensors",
|
272 |
+
"model.layers.21.self_attn.o_proj.weight_scale": "model-00002-of-00004.safetensors",
|
273 |
+
"model.layers.21.self_attn.q_norm.weight": "model-00002-of-00004.safetensors",
|
274 |
+
"model.layers.21.self_attn.q_proj.weight": "model-00002-of-00004.safetensors",
|
275 |
+
"model.layers.21.self_attn.q_proj.weight_scale": "model-00002-of-00004.safetensors",
|
276 |
+
"model.layers.21.self_attn.v_proj.weight": "model-00002-of-00004.safetensors",
|
277 |
+
"model.layers.21.self_attn.v_proj.weight_scale": "model-00002-of-00004.safetensors",
|
278 |
+
"model.layers.22.input_layernorm.weight": "model-00002-of-00004.safetensors",
|
279 |
+
"model.layers.22.mlp.down_proj.weight": "model-00002-of-00004.safetensors",
|
280 |
+
"model.layers.22.mlp.down_proj.weight_scale": "model-00002-of-00004.safetensors",
|
281 |
+
"model.layers.22.mlp.gate_proj.weight": "model-00002-of-00004.safetensors",
|
282 |
+
"model.layers.22.mlp.gate_proj.weight_scale": "model-00002-of-00004.safetensors",
|
283 |
+
"model.layers.22.mlp.up_proj.weight": "model-00002-of-00004.safetensors",
|
284 |
+
"model.layers.22.mlp.up_proj.weight_scale": "model-00002-of-00004.safetensors",
|
285 |
+
"model.layers.22.post_attention_layernorm.weight": "model-00002-of-00004.safetensors",
|
286 |
+
"model.layers.22.self_attn.k_norm.weight": "model-00002-of-00004.safetensors",
|
287 |
+
"model.layers.22.self_attn.k_proj.weight": "model-00002-of-00004.safetensors",
|
288 |
+
"model.layers.22.self_attn.k_proj.weight_scale": "model-00002-of-00004.safetensors",
|
289 |
+
"model.layers.22.self_attn.o_proj.weight": "model-00002-of-00004.safetensors",
|
290 |
+
"model.layers.22.self_attn.o_proj.weight_scale": "model-00002-of-00004.safetensors",
|
291 |
+
"model.layers.22.self_attn.q_norm.weight": "model-00002-of-00004.safetensors",
|
292 |
+
"model.layers.22.self_attn.q_proj.weight": "model-00002-of-00004.safetensors",
|
293 |
+
"model.layers.22.self_attn.q_proj.weight_scale": "model-00002-of-00004.safetensors",
|
294 |
+
"model.layers.22.self_attn.v_proj.weight": "model-00002-of-00004.safetensors",
|
295 |
+
"model.layers.22.self_attn.v_proj.weight_scale": "model-00002-of-00004.safetensors",
|
296 |
+
"model.layers.23.input_layernorm.weight": "model-00002-of-00004.safetensors",
|
297 |
+
"model.layers.23.mlp.down_proj.weight": "model-00002-of-00004.safetensors",
|
298 |
+
"model.layers.23.mlp.down_proj.weight_scale": "model-00002-of-00004.safetensors",
|
299 |
+
"model.layers.23.mlp.gate_proj.weight": "model-00002-of-00004.safetensors",
|
300 |
+
"model.layers.23.mlp.gate_proj.weight_scale": "model-00002-of-00004.safetensors",
|
301 |
+
"model.layers.23.mlp.up_proj.weight": "model-00002-of-00004.safetensors",
|
302 |
+
"model.layers.23.mlp.up_proj.weight_scale": "model-00002-of-00004.safetensors",
|
303 |
+
"model.layers.23.post_attention_layernorm.weight": "model-00002-of-00004.safetensors",
|
304 |
+
"model.layers.23.self_attn.k_norm.weight": "model-00002-of-00004.safetensors",
|
305 |
+
"model.layers.23.self_attn.k_proj.weight": "model-00002-of-00004.safetensors",
|
306 |
+
"model.layers.23.self_attn.k_proj.weight_scale": "model-00002-of-00004.safetensors",
|
307 |
+
"model.layers.23.self_attn.o_proj.weight": "model-00002-of-00004.safetensors",
|
308 |
+
"model.layers.23.self_attn.o_proj.weight_scale": "model-00002-of-00004.safetensors",
|
309 |
+
"model.layers.23.self_attn.q_norm.weight": "model-00002-of-00004.safetensors",
|
310 |
+
"model.layers.23.self_attn.q_proj.weight": "model-00002-of-00004.safetensors",
|
311 |
+
"model.layers.23.self_attn.q_proj.weight_scale": "model-00002-of-00004.safetensors",
|
312 |
+
"model.layers.23.self_attn.v_proj.weight": "model-00002-of-00004.safetensors",
|
313 |
+
"model.layers.23.self_attn.v_proj.weight_scale": "model-00002-of-00004.safetensors",
|
314 |
+
"model.layers.24.input_layernorm.weight": "model-00002-of-00004.safetensors",
|
315 |
+
"model.layers.24.mlp.down_proj.weight": "model-00002-of-00004.safetensors",
|
316 |
+
"model.layers.24.mlp.down_proj.weight_scale": "model-00002-of-00004.safetensors",
|
317 |
+
"model.layers.24.mlp.gate_proj.weight": "model-00002-of-00004.safetensors",
|
318 |
+
"model.layers.24.mlp.gate_proj.weight_scale": "model-00002-of-00004.safetensors",
|
319 |
+
"model.layers.24.mlp.up_proj.weight": "model-00002-of-00004.safetensors",
|
320 |
+
"model.layers.24.mlp.up_proj.weight_scale": "model-00002-of-00004.safetensors",
|
321 |
+
"model.layers.24.post_attention_layernorm.weight": "model-00002-of-00004.safetensors",
|
322 |
+
"model.layers.24.self_attn.k_norm.weight": "model-00002-of-00004.safetensors",
|
323 |
+
"model.layers.24.self_attn.k_proj.weight": "model-00002-of-00004.safetensors",
|
324 |
+
"model.layers.24.self_attn.k_proj.weight_scale": "model-00002-of-00004.safetensors",
|
325 |
+
"model.layers.24.self_attn.o_proj.weight": "model-00002-of-00004.safetensors",
|
326 |
+
"model.layers.24.self_attn.o_proj.weight_scale": "model-00002-of-00004.safetensors",
|
327 |
+
"model.layers.24.self_attn.q_norm.weight": "model-00002-of-00004.safetensors",
|
328 |
+
"model.layers.24.self_attn.q_proj.weight": "model-00002-of-00004.safetensors",
|
329 |
+
"model.layers.24.self_attn.q_proj.weight_scale": "model-00002-of-00004.safetensors",
|
330 |
+
"model.layers.24.self_attn.v_proj.weight": "model-00002-of-00004.safetensors",
|
331 |
+
"model.layers.24.self_attn.v_proj.weight_scale": "model-00002-of-00004.safetensors",
|
332 |
+
"model.layers.25.input_layernorm.weight": "model-00003-of-00004.safetensors",
|
333 |
+
"model.layers.25.mlp.down_proj.weight": "model-00003-of-00004.safetensors",
|
334 |
+
"model.layers.25.mlp.down_proj.weight_scale": "model-00003-of-00004.safetensors",
|
335 |
+
"model.layers.25.mlp.gate_proj.weight": "model-00003-of-00004.safetensors",
|
336 |
+
"model.layers.25.mlp.gate_proj.weight_scale": "model-00003-of-00004.safetensors",
|
337 |
+
"model.layers.25.mlp.up_proj.weight": "model-00003-of-00004.safetensors",
|
338 |
+
"model.layers.25.mlp.up_proj.weight_scale": "model-00003-of-00004.safetensors",
|
339 |
+
"model.layers.25.post_attention_layernorm.weight": "model-00003-of-00004.safetensors",
|
340 |
+
"model.layers.25.self_attn.k_norm.weight": "model-00002-of-00004.safetensors",
|
341 |
+
"model.layers.25.self_attn.k_proj.weight": "model-00002-of-00004.safetensors",
|
342 |
+
"model.layers.25.self_attn.k_proj.weight_scale": "model-00002-of-00004.safetensors",
|
343 |
+
"model.layers.25.self_attn.o_proj.weight": "model-00002-of-00004.safetensors",
|
344 |
+
"model.layers.25.self_attn.o_proj.weight_scale": "model-00002-of-00004.safetensors",
|
345 |
+
"model.layers.25.self_attn.q_norm.weight": "model-00002-of-00004.safetensors",
|
346 |
+
"model.layers.25.self_attn.q_proj.weight": "model-00002-of-00004.safetensors",
|
347 |
+
"model.layers.25.self_attn.q_proj.weight_scale": "model-00002-of-00004.safetensors",
|
348 |
+
"model.layers.25.self_attn.v_proj.weight": "model-00002-of-00004.safetensors",
|
349 |
+
"model.layers.25.self_attn.v_proj.weight_scale": "model-00002-of-00004.safetensors",
|
350 |
+
"model.layers.26.input_layernorm.weight": "model-00003-of-00004.safetensors",
|
351 |
+
"model.layers.26.mlp.down_proj.weight": "model-00003-of-00004.safetensors",
|
352 |
+
"model.layers.26.mlp.down_proj.weight_scale": "model-00003-of-00004.safetensors",
|
353 |
+
"model.layers.26.mlp.gate_proj.weight": "model-00003-of-00004.safetensors",
|
354 |
+
"model.layers.26.mlp.gate_proj.weight_scale": "model-00003-of-00004.safetensors",
|
355 |
+
"model.layers.26.mlp.up_proj.weight": "model-00003-of-00004.safetensors",
|
356 |
+
"model.layers.26.mlp.up_proj.weight_scale": "model-00003-of-00004.safetensors",
|
357 |
+
"model.layers.26.post_attention_layernorm.weight": "model-00003-of-00004.safetensors",
|
358 |
+
"model.layers.26.self_attn.k_norm.weight": "model-00003-of-00004.safetensors",
|
359 |
+
"model.layers.26.self_attn.k_proj.weight": "model-00003-of-00004.safetensors",
|
360 |
+
"model.layers.26.self_attn.k_proj.weight_scale": "model-00003-of-00004.safetensors",
|
361 |
+
"model.layers.26.self_attn.o_proj.weight": "model-00003-of-00004.safetensors",
|
362 |
+
"model.layers.26.self_attn.o_proj.weight_scale": "model-00003-of-00004.safetensors",
|
363 |
+
"model.layers.26.self_attn.q_norm.weight": "model-00003-of-00004.safetensors",
|
364 |
+
"model.layers.26.self_attn.q_proj.weight": "model-00003-of-00004.safetensors",
|
365 |
+
"model.layers.26.self_attn.q_proj.weight_scale": "model-00003-of-00004.safetensors",
|
366 |
+
"model.layers.26.self_attn.v_proj.weight": "model-00003-of-00004.safetensors",
|
367 |
+
"model.layers.26.self_attn.v_proj.weight_scale": "model-00003-of-00004.safetensors",
|
368 |
+
"model.layers.27.input_layernorm.weight": "model-00003-of-00004.safetensors",
|
369 |
+
"model.layers.27.mlp.down_proj.weight": "model-00003-of-00004.safetensors",
|
370 |
+
"model.layers.27.mlp.down_proj.weight_scale": "model-00003-of-00004.safetensors",
|
371 |
+
"model.layers.27.mlp.gate_proj.weight": "model-00003-of-00004.safetensors",
|
372 |
+
"model.layers.27.mlp.gate_proj.weight_scale": "model-00003-of-00004.safetensors",
|
373 |
+
"model.layers.27.mlp.up_proj.weight": "model-00003-of-00004.safetensors",
|
374 |
+
"model.layers.27.mlp.up_proj.weight_scale": "model-00003-of-00004.safetensors",
|
375 |
+
"model.layers.27.post_attention_layernorm.weight": "model-00003-of-00004.safetensors",
|
376 |
+
"model.layers.27.self_attn.k_norm.weight": "model-00003-of-00004.safetensors",
|
377 |
+
"model.layers.27.self_attn.k_proj.weight": "model-00003-of-00004.safetensors",
|
378 |
+
"model.layers.27.self_attn.k_proj.weight_scale": "model-00003-of-00004.safetensors",
|
379 |
+
"model.layers.27.self_attn.o_proj.weight": "model-00003-of-00004.safetensors",
|
380 |
+
"model.layers.27.self_attn.o_proj.weight_scale": "model-00003-of-00004.safetensors",
|
381 |
+
"model.layers.27.self_attn.q_norm.weight": "model-00003-of-00004.safetensors",
|
382 |
+
"model.layers.27.self_attn.q_proj.weight": "model-00003-of-00004.safetensors",
|
383 |
+
"model.layers.27.self_attn.q_proj.weight_scale": "model-00003-of-00004.safetensors",
|
384 |
+
"model.layers.27.self_attn.v_proj.weight": "model-00003-of-00004.safetensors",
|
385 |
+
"model.layers.27.self_attn.v_proj.weight_scale": "model-00003-of-00004.safetensors",
|
386 |
+
"model.layers.28.input_layernorm.weight": "model-00003-of-00004.safetensors",
|
387 |
+
"model.layers.28.mlp.down_proj.weight": "model-00003-of-00004.safetensors",
|
388 |
+
"model.layers.28.mlp.down_proj.weight_scale": "model-00003-of-00004.safetensors",
|
389 |
+
"model.layers.28.mlp.gate_proj.weight": "model-00003-of-00004.safetensors",
|
390 |
+
"model.layers.28.mlp.gate_proj.weight_scale": "model-00003-of-00004.safetensors",
|
391 |
+
"model.layers.28.mlp.up_proj.weight": "model-00003-of-00004.safetensors",
|
392 |
+
"model.layers.28.mlp.up_proj.weight_scale": "model-00003-of-00004.safetensors",
|
393 |
+
"model.layers.28.post_attention_layernorm.weight": "model-00003-of-00004.safetensors",
|
394 |
+
"model.layers.28.self_attn.k_norm.weight": "model-00003-of-00004.safetensors",
|
395 |
+
"model.layers.28.self_attn.k_proj.weight": "model-00003-of-00004.safetensors",
|
396 |
+
"model.layers.28.self_attn.k_proj.weight_scale": "model-00003-of-00004.safetensors",
|
397 |
+
"model.layers.28.self_attn.o_proj.weight": "model-00003-of-00004.safetensors",
|
398 |
+
"model.layers.28.self_attn.o_proj.weight_scale": "model-00003-of-00004.safetensors",
|
399 |
+
"model.layers.28.self_attn.q_norm.weight": "model-00003-of-00004.safetensors",
|
400 |
+
"model.layers.28.self_attn.q_proj.weight": "model-00003-of-00004.safetensors",
|
401 |
+
"model.layers.28.self_attn.q_proj.weight_scale": "model-00003-of-00004.safetensors",
|
402 |
+
"model.layers.28.self_attn.v_proj.weight": "model-00003-of-00004.safetensors",
|
403 |
+
"model.layers.28.self_attn.v_proj.weight_scale": "model-00003-of-00004.safetensors",
|
404 |
+
"model.layers.29.input_layernorm.weight": "model-00003-of-00004.safetensors",
|
405 |
+
"model.layers.29.mlp.down_proj.weight": "model-00003-of-00004.safetensors",
|
406 |
+
"model.layers.29.mlp.down_proj.weight_scale": "model-00003-of-00004.safetensors",
|
407 |
+
"model.layers.29.mlp.gate_proj.weight": "model-00003-of-00004.safetensors",
|
408 |
+
"model.layers.29.mlp.gate_proj.weight_scale": "model-00003-of-00004.safetensors",
|
409 |
+
"model.layers.29.mlp.up_proj.weight": "model-00003-of-00004.safetensors",
|
410 |
+
"model.layers.29.mlp.up_proj.weight_scale": "model-00003-of-00004.safetensors",
|
411 |
+
"model.layers.29.post_attention_layernorm.weight": "model-00003-of-00004.safetensors",
|
412 |
+
"model.layers.29.self_attn.k_norm.weight": "model-00003-of-00004.safetensors",
|
413 |
+
"model.layers.29.self_attn.k_proj.weight": "model-00003-of-00004.safetensors",
|
414 |
+
"model.layers.29.self_attn.k_proj.weight_scale": "model-00003-of-00004.safetensors",
|
415 |
+
"model.layers.29.self_attn.o_proj.weight": "model-00003-of-00004.safetensors",
|
416 |
+
"model.layers.29.self_attn.o_proj.weight_scale": "model-00003-of-00004.safetensors",
|
417 |
+
"model.layers.29.self_attn.q_norm.weight": "model-00003-of-00004.safetensors",
|
418 |
+
"model.layers.29.self_attn.q_proj.weight": "model-00003-of-00004.safetensors",
|
419 |
+
"model.layers.29.self_attn.q_proj.weight_scale": "model-00003-of-00004.safetensors",
|
420 |
+
"model.layers.29.self_attn.v_proj.weight": "model-00003-of-00004.safetensors",
|
421 |
+
"model.layers.29.self_attn.v_proj.weight_scale": "model-00003-of-00004.safetensors",
|
422 |
+
"model.layers.3.input_layernorm.weight": "model-00001-of-00004.safetensors",
|
423 |
+
"model.layers.3.mlp.down_proj.weight": "model-00001-of-00004.safetensors",
|
424 |
+
"model.layers.3.mlp.down_proj.weight_scale": "model-00001-of-00004.safetensors",
|
425 |
+
"model.layers.3.mlp.gate_proj.weight": "model-00001-of-00004.safetensors",
|
426 |
+
"model.layers.3.mlp.gate_proj.weight_scale": "model-00001-of-00004.safetensors",
|
427 |
+
"model.layers.3.mlp.up_proj.weight": "model-00001-of-00004.safetensors",
|
428 |
+
"model.layers.3.mlp.up_proj.weight_scale": "model-00001-of-00004.safetensors",
|
429 |
+
"model.layers.3.post_attention_layernorm.weight": "model-00001-of-00004.safetensors",
|
430 |
+
"model.layers.3.self_attn.k_norm.weight": "model-00001-of-00004.safetensors",
|
431 |
+
"model.layers.3.self_attn.k_proj.weight": "model-00001-of-00004.safetensors",
|
432 |
+
"model.layers.3.self_attn.k_proj.weight_scale": "model-00001-of-00004.safetensors",
|
433 |
+
"model.layers.3.self_attn.o_proj.weight": "model-00001-of-00004.safetensors",
|
434 |
+
"model.layers.3.self_attn.o_proj.weight_scale": "model-00001-of-00004.safetensors",
|
435 |
+
"model.layers.3.self_attn.q_norm.weight": "model-00001-of-00004.safetensors",
|
436 |
+
"model.layers.3.self_attn.q_proj.weight": "model-00001-of-00004.safetensors",
|
437 |
+
"model.layers.3.self_attn.q_proj.weight_scale": "model-00001-of-00004.safetensors",
|
438 |
+
"model.layers.3.self_attn.v_proj.weight": "model-00001-of-00004.safetensors",
|
439 |
+
"model.layers.3.self_attn.v_proj.weight_scale": "model-00001-of-00004.safetensors",
|
440 |
+
"model.layers.30.input_layernorm.weight": "model-00003-of-00004.safetensors",
|
441 |
+
"model.layers.30.mlp.down_proj.weight": "model-00003-of-00004.safetensors",
|
442 |
+
"model.layers.30.mlp.down_proj.weight_scale": "model-00003-of-00004.safetensors",
|
443 |
+
"model.layers.30.mlp.gate_proj.weight": "model-00003-of-00004.safetensors",
|
444 |
+
"model.layers.30.mlp.gate_proj.weight_scale": "model-00003-of-00004.safetensors",
|
445 |
+
"model.layers.30.mlp.up_proj.weight": "model-00003-of-00004.safetensors",
|
446 |
+
"model.layers.30.mlp.up_proj.weight_scale": "model-00003-of-00004.safetensors",
|
447 |
+
"model.layers.30.post_attention_layernorm.weight": "model-00003-of-00004.safetensors",
|
448 |
+
"model.layers.30.self_attn.k_norm.weight": "model-00003-of-00004.safetensors",
|
449 |
+
"model.layers.30.self_attn.k_proj.weight": "model-00003-of-00004.safetensors",
|
450 |
+
"model.layers.30.self_attn.k_proj.weight_scale": "model-00003-of-00004.safetensors",
|
451 |
+
"model.layers.30.self_attn.o_proj.weight": "model-00003-of-00004.safetensors",
|
452 |
+
"model.layers.30.self_attn.o_proj.weight_scale": "model-00003-of-00004.safetensors",
|
453 |
+
"model.layers.30.self_attn.q_norm.weight": "model-00003-of-00004.safetensors",
|
454 |
+
"model.layers.30.self_attn.q_proj.weight": "model-00003-of-00004.safetensors",
|
455 |
+
"model.layers.30.self_attn.q_proj.weight_scale": "model-00003-of-00004.safetensors",
|
456 |
+
"model.layers.30.self_attn.v_proj.weight": "model-00003-of-00004.safetensors",
|
457 |
+
"model.layers.30.self_attn.v_proj.weight_scale": "model-00003-of-00004.safetensors",
|
458 |
+
"model.layers.31.input_layernorm.weight": "model-00003-of-00004.safetensors",
|
459 |
+
"model.layers.31.mlp.down_proj.weight": "model-00003-of-00004.safetensors",
|
460 |
+
"model.layers.31.mlp.down_proj.weight_scale": "model-00003-of-00004.safetensors",
|
461 |
+
"model.layers.31.mlp.gate_proj.weight": "model-00003-of-00004.safetensors",
|
462 |
+
"model.layers.31.mlp.gate_proj.weight_scale": "model-00003-of-00004.safetensors",
|
463 |
+
"model.layers.31.mlp.up_proj.weight": "model-00003-of-00004.safetensors",
|
464 |
+
"model.layers.31.mlp.up_proj.weight_scale": "model-00003-of-00004.safetensors",
|
465 |
+
"model.layers.31.post_attention_layernorm.weight": "model-00003-of-00004.safetensors",
|
466 |
+
"model.layers.31.self_attn.k_norm.weight": "model-00003-of-00004.safetensors",
|
467 |
+
"model.layers.31.self_attn.k_proj.weight": "model-00003-of-00004.safetensors",
|
468 |
+
"model.layers.31.self_attn.k_proj.weight_scale": "model-00003-of-00004.safetensors",
|
469 |
+
"model.layers.31.self_attn.o_proj.weight": "model-00003-of-00004.safetensors",
|
470 |
+
"model.layers.31.self_attn.o_proj.weight_scale": "model-00003-of-00004.safetensors",
|
471 |
+
"model.layers.31.self_attn.q_norm.weight": "model-00003-of-00004.safetensors",
|
472 |
+
"model.layers.31.self_attn.q_proj.weight": "model-00003-of-00004.safetensors",
|
473 |
+
"model.layers.31.self_attn.q_proj.weight_scale": "model-00003-of-00004.safetensors",
|
474 |
+
"model.layers.31.self_attn.v_proj.weight": "model-00003-of-00004.safetensors",
|
475 |
+
"model.layers.31.self_attn.v_proj.weight_scale": "model-00003-of-00004.safetensors",
|
476 |
+
"model.layers.32.input_layernorm.weight": "model-00003-of-00004.safetensors",
|
477 |
+
"model.layers.32.mlp.down_proj.weight": "model-00003-of-00004.safetensors",
|
478 |
+
"model.layers.32.mlp.down_proj.weight_scale": "model-00003-of-00004.safetensors",
|
479 |
+
"model.layers.32.mlp.gate_proj.weight": "model-00003-of-00004.safetensors",
|
480 |
+
"model.layers.32.mlp.gate_proj.weight_scale": "model-00003-of-00004.safetensors",
|
481 |
+
"model.layers.32.mlp.up_proj.weight": "model-00003-of-00004.safetensors",
|
482 |
+
"model.layers.32.mlp.up_proj.weight_scale": "model-00003-of-00004.safetensors",
|
483 |
+
"model.layers.32.post_attention_layernorm.weight": "model-00003-of-00004.safetensors",
|
484 |
+
"model.layers.32.self_attn.k_norm.weight": "model-00003-of-00004.safetensors",
|
485 |
+
"model.layers.32.self_attn.k_proj.weight": "model-00003-of-00004.safetensors",
|
486 |
+
"model.layers.32.self_attn.k_proj.weight_scale": "model-00003-of-00004.safetensors",
|
487 |
+
"model.layers.32.self_attn.o_proj.weight": "model-00003-of-00004.safetensors",
|
488 |
+
"model.layers.32.self_attn.o_proj.weight_scale": "model-00003-of-00004.safetensors",
|
489 |
+
"model.layers.32.self_attn.q_norm.weight": "model-00003-of-00004.safetensors",
|
490 |
+
"model.layers.32.self_attn.q_proj.weight": "model-00003-of-00004.safetensors",
|
491 |
+
"model.layers.32.self_attn.q_proj.weight_scale": "model-00003-of-00004.safetensors",
|
492 |
+
"model.layers.32.self_attn.v_proj.weight": "model-00003-of-00004.safetensors",
|
493 |
+
"model.layers.32.self_attn.v_proj.weight_scale": "model-00003-of-00004.safetensors",
|
494 |
+
"model.layers.33.input_layernorm.weight": "model-00003-of-00004.safetensors",
|
495 |
+
"model.layers.33.mlp.down_proj.weight": "model-00003-of-00004.safetensors",
|
496 |
+
"model.layers.33.mlp.down_proj.weight_scale": "model-00003-of-00004.safetensors",
|
497 |
+
"model.layers.33.mlp.gate_proj.weight": "model-00003-of-00004.safetensors",
|
498 |
+
"model.layers.33.mlp.gate_proj.weight_scale": "model-00003-of-00004.safetensors",
|
499 |
+
"model.layers.33.mlp.up_proj.weight": "model-00003-of-00004.safetensors",
|
500 |
+
"model.layers.33.mlp.up_proj.weight_scale": "model-00003-of-00004.safetensors",
|
501 |
+
"model.layers.33.post_attention_layernorm.weight": "model-00003-of-00004.safetensors",
|
502 |
+
"model.layers.33.self_attn.k_norm.weight": "model-00003-of-00004.safetensors",
|
503 |
+
"model.layers.33.self_attn.k_proj.weight": "model-00003-of-00004.safetensors",
|
504 |
+
"model.layers.33.self_attn.k_proj.weight_scale": "model-00003-of-00004.safetensors",
|
505 |
+
"model.layers.33.self_attn.o_proj.weight": "model-00003-of-00004.safetensors",
|
506 |
+
"model.layers.33.self_attn.o_proj.weight_scale": "model-00003-of-00004.safetensors",
|
507 |
+
"model.layers.33.self_attn.q_norm.weight": "model-00003-of-00004.safetensors",
|
508 |
+
"model.layers.33.self_attn.q_proj.weight": "model-00003-of-00004.safetensors",
|
509 |
+
"model.layers.33.self_attn.q_proj.weight_scale": "model-00003-of-00004.safetensors",
|
510 |
+
"model.layers.33.self_attn.v_proj.weight": "model-00003-of-00004.safetensors",
|
511 |
+
"model.layers.33.self_attn.v_proj.weight_scale": "model-00003-of-00004.safetensors",
|
512 |
+
"model.layers.34.input_layernorm.weight": "model-00003-of-00004.safetensors",
|
513 |
+
"model.layers.34.mlp.down_proj.weight": "model-00003-of-00004.safetensors",
|
514 |
+
"model.layers.34.mlp.down_proj.weight_scale": "model-00003-of-00004.safetensors",
|
515 |
+
"model.layers.34.mlp.gate_proj.weight": "model-00003-of-00004.safetensors",
|
516 |
+
"model.layers.34.mlp.gate_proj.weight_scale": "model-00003-of-00004.safetensors",
|
517 |
+
"model.layers.34.mlp.up_proj.weight": "model-00003-of-00004.safetensors",
|
518 |
+
"model.layers.34.mlp.up_proj.weight_scale": "model-00003-of-00004.safetensors",
|
519 |
+
"model.layers.34.post_attention_layernorm.weight": "model-00003-of-00004.safetensors",
|
520 |
+
"model.layers.34.self_attn.k_norm.weight": "model-00003-of-00004.safetensors",
|
521 |
+
"model.layers.34.self_attn.k_proj.weight": "model-00003-of-00004.safetensors",
|
522 |
+
"model.layers.34.self_attn.k_proj.weight_scale": "model-00003-of-00004.safetensors",
|
523 |
+
"model.layers.34.self_attn.o_proj.weight": "model-00003-of-00004.safetensors",
|
524 |
+
"model.layers.34.self_attn.o_proj.weight_scale": "model-00003-of-00004.safetensors",
|
525 |
+
"model.layers.34.self_attn.q_norm.weight": "model-00003-of-00004.safetensors",
|
526 |
+
"model.layers.34.self_attn.q_proj.weight": "model-00003-of-00004.safetensors",
|
527 |
+
"model.layers.34.self_attn.q_proj.weight_scale": "model-00003-of-00004.safetensors",
|
528 |
+
"model.layers.34.self_attn.v_proj.weight": "model-00003-of-00004.safetensors",
|
529 |
+
"model.layers.34.self_attn.v_proj.weight_scale": "model-00003-of-00004.safetensors",
|
530 |
+
"model.layers.35.input_layernorm.weight": "model-00003-of-00004.safetensors",
|
531 |
+
"model.layers.35.mlp.down_proj.weight": "model-00003-of-00004.safetensors",
|
532 |
+
"model.layers.35.mlp.down_proj.weight_scale": "model-00003-of-00004.safetensors",
|
533 |
+
"model.layers.35.mlp.gate_proj.weight": "model-00003-of-00004.safetensors",
|
534 |
+
"model.layers.35.mlp.gate_proj.weight_scale": "model-00003-of-00004.safetensors",
|
535 |
+
"model.layers.35.mlp.up_proj.weight": "model-00003-of-00004.safetensors",
|
536 |
+
"model.layers.35.mlp.up_proj.weight_scale": "model-00003-of-00004.safetensors",
|
537 |
+
"model.layers.35.post_attention_layernorm.weight": "model-00003-of-00004.safetensors",
|
538 |
+
"model.layers.35.self_attn.k_norm.weight": "model-00003-of-00004.safetensors",
|
539 |
+
"model.layers.35.self_attn.k_proj.weight": "model-00003-of-00004.safetensors",
|
540 |
+
"model.layers.35.self_attn.k_proj.weight_scale": "model-00003-of-00004.safetensors",
|
541 |
+
"model.layers.35.self_attn.o_proj.weight": "model-00003-of-00004.safetensors",
|
542 |
+
"model.layers.35.self_attn.o_proj.weight_scale": "model-00003-of-00004.safetensors",
|
543 |
+
"model.layers.35.self_attn.q_norm.weight": "model-00003-of-00004.safetensors",
|
544 |
+
"model.layers.35.self_attn.q_proj.weight": "model-00003-of-00004.safetensors",
|
545 |
+
"model.layers.35.self_attn.q_proj.weight_scale": "model-00003-of-00004.safetensors",
|
546 |
+
"model.layers.35.self_attn.v_proj.weight": "model-00003-of-00004.safetensors",
|
547 |
+
"model.layers.35.self_attn.v_proj.weight_scale": "model-00003-of-00004.safetensors",
|
548 |
+
"model.layers.36.input_layernorm.weight": "model-00003-of-00004.safetensors",
|
549 |
+
"model.layers.36.mlp.down_proj.weight": "model-00003-of-00004.safetensors",
|
550 |
+
"model.layers.36.mlp.down_proj.weight_scale": "model-00003-of-00004.safetensors",
|
551 |
+
"model.layers.36.mlp.gate_proj.weight": "model-00003-of-00004.safetensors",
|
552 |
+
"model.layers.36.mlp.gate_proj.weight_scale": "model-00003-of-00004.safetensors",
|
553 |
+
"model.layers.36.mlp.up_proj.weight": "model-00003-of-00004.safetensors",
|
554 |
+
"model.layers.36.mlp.up_proj.weight_scale": "model-00003-of-00004.safetensors",
|
555 |
+
"model.layers.36.post_attention_layernorm.weight": "model-00003-of-00004.safetensors",
|
556 |
+
"model.layers.36.self_attn.k_norm.weight": "model-00003-of-00004.safetensors",
|
557 |
+
"model.layers.36.self_attn.k_proj.weight": "model-00003-of-00004.safetensors",
|
558 |
+
"model.layers.36.self_attn.k_proj.weight_scale": "model-00003-of-00004.safetensors",
|
559 |
+
"model.layers.36.self_attn.o_proj.weight": "model-00003-of-00004.safetensors",
|
560 |
+
"model.layers.36.self_attn.o_proj.weight_scale": "model-00003-of-00004.safetensors",
|
561 |
+
"model.layers.36.self_attn.q_norm.weight": "model-00003-of-00004.safetensors",
|
562 |
+
"model.layers.36.self_attn.q_proj.weight": "model-00003-of-00004.safetensors",
|
563 |
+
"model.layers.36.self_attn.q_proj.weight_scale": "model-00003-of-00004.safetensors",
|
564 |
+
"model.layers.36.self_attn.v_proj.weight": "model-00003-of-00004.safetensors",
|
565 |
+
"model.layers.36.self_attn.v_proj.weight_scale": "model-00003-of-00004.safetensors",
|
566 |
+
"model.layers.37.input_layernorm.weight": "model-00003-of-00004.safetensors",
|
567 |
+
"model.layers.37.mlp.down_proj.weight": "model-00003-of-00004.safetensors",
|
568 |
+
"model.layers.37.mlp.down_proj.weight_scale": "model-00003-of-00004.safetensors",
|
569 |
+
"model.layers.37.mlp.gate_proj.weight": "model-00003-of-00004.safetensors",
|
570 |
+
"model.layers.37.mlp.gate_proj.weight_scale": "model-00003-of-00004.safetensors",
|
571 |
+
"model.layers.37.mlp.up_proj.weight": "model-00003-of-00004.safetensors",
|
572 |
+
"model.layers.37.mlp.up_proj.weight_scale": "model-00003-of-00004.safetensors",
|
573 |
+
"model.layers.37.post_attention_layernorm.weight": "model-00003-of-00004.safetensors",
|
574 |
+
"model.layers.37.self_attn.k_norm.weight": "model-00003-of-00004.safetensors",
|
575 |
+
"model.layers.37.self_attn.k_proj.weight": "model-00003-of-00004.safetensors",
|
576 |
+
"model.layers.37.self_attn.k_proj.weight_scale": "model-00003-of-00004.safetensors",
|
577 |
+
"model.layers.37.self_attn.o_proj.weight": "model-00003-of-00004.safetensors",
|
578 |
+
"model.layers.37.self_attn.o_proj.weight_scale": "model-00003-of-00004.safetensors",
|
579 |
+
"model.layers.37.self_attn.q_norm.weight": "model-00003-of-00004.safetensors",
|
580 |
+
"model.layers.37.self_attn.q_proj.weight": "model-00003-of-00004.safetensors",
|
581 |
+
"model.layers.37.self_attn.q_proj.weight_scale": "model-00003-of-00004.safetensors",
|
582 |
+
"model.layers.37.self_attn.v_proj.weight": "model-00003-of-00004.safetensors",
|
583 |
+
"model.layers.37.self_attn.v_proj.weight_scale": "model-00003-of-00004.safetensors",
|
584 |
+
"model.layers.38.input_layernorm.weight": "model-00003-of-00004.safetensors",
|
585 |
+
"model.layers.38.mlp.down_proj.weight": "model-00003-of-00004.safetensors",
|
586 |
+
"model.layers.38.mlp.down_proj.weight_scale": "model-00003-of-00004.safetensors",
|
587 |
+
"model.layers.38.mlp.gate_proj.weight": "model-00003-of-00004.safetensors",
|
588 |
+
"model.layers.38.mlp.gate_proj.weight_scale": "model-00003-of-00004.safetensors",
|
589 |
+
"model.layers.38.mlp.up_proj.weight": "model-00003-of-00004.safetensors",
|
590 |
+
"model.layers.38.mlp.up_proj.weight_scale": "model-00003-of-00004.safetensors",
|
591 |
+
"model.layers.38.post_attention_layernorm.weight": "model-00003-of-00004.safetensors",
|
592 |
+
"model.layers.38.self_attn.k_norm.weight": "model-00003-of-00004.safetensors",
|
593 |
+
"model.layers.38.self_attn.k_proj.weight": "model-00003-of-00004.safetensors",
|
594 |
+
"model.layers.38.self_attn.k_proj.weight_scale": "model-00003-of-00004.safetensors",
|
595 |
+
"model.layers.38.self_attn.o_proj.weight": "model-00003-of-00004.safetensors",
|
596 |
+
"model.layers.38.self_attn.o_proj.weight_scale": "model-00003-of-00004.safetensors",
|
597 |
+
"model.layers.38.self_attn.q_norm.weight": "model-00003-of-00004.safetensors",
|
598 |
+
"model.layers.38.self_attn.q_proj.weight": "model-00003-of-00004.safetensors",
|
599 |
+
"model.layers.38.self_attn.q_proj.weight_scale": "model-00003-of-00004.safetensors",
|
600 |
+
"model.layers.38.self_attn.v_proj.weight": "model-00003-of-00004.safetensors",
|
601 |
+
"model.layers.38.self_attn.v_proj.weight_scale": "model-00003-of-00004.safetensors",
|
602 |
+
"model.layers.39.input_layernorm.weight": "model-00003-of-00004.safetensors",
|
603 |
+
"model.layers.39.mlp.down_proj.weight": "model-00003-of-00004.safetensors",
|
604 |
+
"model.layers.39.mlp.down_proj.weight_scale": "model-00003-of-00004.safetensors",
|
605 |
+
"model.layers.39.mlp.gate_proj.weight": "model-00003-of-00004.safetensors",
|
606 |
+
"model.layers.39.mlp.gate_proj.weight_scale": "model-00003-of-00004.safetensors",
|
607 |
+
"model.layers.39.mlp.up_proj.weight": "model-00003-of-00004.safetensors",
|
608 |
+
"model.layers.39.mlp.up_proj.weight_scale": "model-00003-of-00004.safetensors",
|
609 |
+
"model.layers.39.post_attention_layernorm.weight": "model-00003-of-00004.safetensors",
|
610 |
+
"model.layers.39.self_attn.k_norm.weight": "model-00003-of-00004.safetensors",
|
611 |
+
"model.layers.39.self_attn.k_proj.weight": "model-00003-of-00004.safetensors",
|
612 |
+
"model.layers.39.self_attn.k_proj.weight_scale": "model-00003-of-00004.safetensors",
|
613 |
+
"model.layers.39.self_attn.o_proj.weight": "model-00003-of-00004.safetensors",
|
614 |
+
"model.layers.39.self_attn.o_proj.weight_scale": "model-00003-of-00004.safetensors",
|
615 |
+
"model.layers.39.self_attn.q_norm.weight": "model-00003-of-00004.safetensors",
|
616 |
+
"model.layers.39.self_attn.q_proj.weight": "model-00003-of-00004.safetensors",
|
617 |
+
"model.layers.39.self_attn.q_proj.weight_scale": "model-00003-of-00004.safetensors",
|
618 |
+
"model.layers.39.self_attn.v_proj.weight": "model-00003-of-00004.safetensors",
|
619 |
+
"model.layers.39.self_attn.v_proj.weight_scale": "model-00003-of-00004.safetensors",
|
620 |
+
"model.layers.4.input_layernorm.weight": "model-00001-of-00004.safetensors",
|
621 |
+
"model.layers.4.mlp.down_proj.weight": "model-00001-of-00004.safetensors",
|
622 |
+
"model.layers.4.mlp.down_proj.weight_scale": "model-00001-of-00004.safetensors",
|
623 |
+
"model.layers.4.mlp.gate_proj.weight": "model-00001-of-00004.safetensors",
|
624 |
+
"model.layers.4.mlp.gate_proj.weight_scale": "model-00001-of-00004.safetensors",
|
625 |
+
"model.layers.4.mlp.up_proj.weight": "model-00001-of-00004.safetensors",
|
626 |
+
"model.layers.4.mlp.up_proj.weight_scale": "model-00001-of-00004.safetensors",
|
627 |
+
"model.layers.4.post_attention_layernorm.weight": "model-00001-of-00004.safetensors",
|
628 |
+
"model.layers.4.self_attn.k_norm.weight": "model-00001-of-00004.safetensors",
|
629 |
+
"model.layers.4.self_attn.k_proj.weight": "model-00001-of-00004.safetensors",
|
630 |
+
"model.layers.4.self_attn.k_proj.weight_scale": "model-00001-of-00004.safetensors",
|
631 |
+
"model.layers.4.self_attn.o_proj.weight": "model-00001-of-00004.safetensors",
|
632 |
+
"model.layers.4.self_attn.o_proj.weight_scale": "model-00001-of-00004.safetensors",
|
633 |
+
"model.layers.4.self_attn.q_norm.weight": "model-00001-of-00004.safetensors",
|
634 |
+
"model.layers.4.self_attn.q_proj.weight": "model-00001-of-00004.safetensors",
|
635 |
+
"model.layers.4.self_attn.q_proj.weight_scale": "model-00001-of-00004.safetensors",
|
636 |
+
"model.layers.4.self_attn.v_proj.weight": "model-00001-of-00004.safetensors",
|
637 |
+
"model.layers.4.self_attn.v_proj.weight_scale": "model-00001-of-00004.safetensors",
|
638 |
+
"model.layers.5.input_layernorm.weight": "model-00001-of-00004.safetensors",
|
639 |
+
"model.layers.5.mlp.down_proj.weight": "model-00001-of-00004.safetensors",
|
640 |
+
"model.layers.5.mlp.down_proj.weight_scale": "model-00001-of-00004.safetensors",
|
641 |
+
"model.layers.5.mlp.gate_proj.weight": "model-00001-of-00004.safetensors",
|
642 |
+
"model.layers.5.mlp.gate_proj.weight_scale": "model-00001-of-00004.safetensors",
|
643 |
+
"model.layers.5.mlp.up_proj.weight": "model-00001-of-00004.safetensors",
|
644 |
+
"model.layers.5.mlp.up_proj.weight_scale": "model-00001-of-00004.safetensors",
|
645 |
+
"model.layers.5.post_attention_layernorm.weight": "model-00001-of-00004.safetensors",
|
646 |
+
"model.layers.5.self_attn.k_norm.weight": "model-00001-of-00004.safetensors",
|
647 |
+
"model.layers.5.self_attn.k_proj.weight": "model-00001-of-00004.safetensors",
|
648 |
+
"model.layers.5.self_attn.k_proj.weight_scale": "model-00001-of-00004.safetensors",
|
649 |
+
"model.layers.5.self_attn.o_proj.weight": "model-00001-of-00004.safetensors",
|
650 |
+
"model.layers.5.self_attn.o_proj.weight_scale": "model-00001-of-00004.safetensors",
|
651 |
+
"model.layers.5.self_attn.q_norm.weight": "model-00001-of-00004.safetensors",
|
652 |
+
"model.layers.5.self_attn.q_proj.weight": "model-00001-of-00004.safetensors",
|
653 |
+
"model.layers.5.self_attn.q_proj.weight_scale": "model-00001-of-00004.safetensors",
|
654 |
+
"model.layers.5.self_attn.v_proj.weight": "model-00001-of-00004.safetensors",
|
655 |
+
"model.layers.5.self_attn.v_proj.weight_scale": "model-00001-of-00004.safetensors",
|
656 |
+
"model.layers.6.input_layernorm.weight": "model-00001-of-00004.safetensors",
|
657 |
+
"model.layers.6.mlp.down_proj.weight": "model-00001-of-00004.safetensors",
|
658 |
+
"model.layers.6.mlp.down_proj.weight_scale": "model-00001-of-00004.safetensors",
|
659 |
+
"model.layers.6.mlp.gate_proj.weight": "model-00001-of-00004.safetensors",
|
660 |
+
"model.layers.6.mlp.gate_proj.weight_scale": "model-00001-of-00004.safetensors",
|
661 |
+
"model.layers.6.mlp.up_proj.weight": "model-00001-of-00004.safetensors",
|
662 |
+
"model.layers.6.mlp.up_proj.weight_scale": "model-00001-of-00004.safetensors",
|
663 |
+
"model.layers.6.post_attention_layernorm.weight": "model-00001-of-00004.safetensors",
|
664 |
+
"model.layers.6.self_attn.k_norm.weight": "model-00001-of-00004.safetensors",
|
665 |
+
"model.layers.6.self_attn.k_proj.weight": "model-00001-of-00004.safetensors",
|
666 |
+
"model.layers.6.self_attn.k_proj.weight_scale": "model-00001-of-00004.safetensors",
|
667 |
+
"model.layers.6.self_attn.o_proj.weight": "model-00001-of-00004.safetensors",
|
668 |
+
"model.layers.6.self_attn.o_proj.weight_scale": "model-00001-of-00004.safetensors",
|
669 |
+
"model.layers.6.self_attn.q_norm.weight": "model-00001-of-00004.safetensors",
|
670 |
+
"model.layers.6.self_attn.q_proj.weight": "model-00001-of-00004.safetensors",
|
671 |
+
"model.layers.6.self_attn.q_proj.weight_scale": "model-00001-of-00004.safetensors",
|
672 |
+
"model.layers.6.self_attn.v_proj.weight": "model-00001-of-00004.safetensors",
|
673 |
+
"model.layers.6.self_attn.v_proj.weight_scale": "model-00001-of-00004.safetensors",
|
674 |
+
"model.layers.7.input_layernorm.weight": "model-00001-of-00004.safetensors",
|
675 |
+
"model.layers.7.mlp.down_proj.weight": "model-00001-of-00004.safetensors",
|
676 |
+
"model.layers.7.mlp.down_proj.weight_scale": "model-00001-of-00004.safetensors",
|
677 |
+
"model.layers.7.mlp.gate_proj.weight": "model-00001-of-00004.safetensors",
|
678 |
+
"model.layers.7.mlp.gate_proj.weight_scale": "model-00001-of-00004.safetensors",
|
679 |
+
"model.layers.7.mlp.up_proj.weight": "model-00001-of-00004.safetensors",
|
680 |
+
"model.layers.7.mlp.up_proj.weight_scale": "model-00001-of-00004.safetensors",
|
681 |
+
"model.layers.7.post_attention_layernorm.weight": "model-00001-of-00004.safetensors",
|
682 |
+
"model.layers.7.self_attn.k_norm.weight": "model-00001-of-00004.safetensors",
|
683 |
+
"model.layers.7.self_attn.k_proj.weight": "model-00001-of-00004.safetensors",
|
684 |
+
"model.layers.7.self_attn.k_proj.weight_scale": "model-00001-of-00004.safetensors",
|
685 |
+
"model.layers.7.self_attn.o_proj.weight": "model-00001-of-00004.safetensors",
|
686 |
+
"model.layers.7.self_attn.o_proj.weight_scale": "model-00001-of-00004.safetensors",
|
687 |
+
"model.layers.7.self_attn.q_norm.weight": "model-00001-of-00004.safetensors",
|
688 |
+
"model.layers.7.self_attn.q_proj.weight": "model-00001-of-00004.safetensors",
|
689 |
+
"model.layers.7.self_attn.q_proj.weight_scale": "model-00001-of-00004.safetensors",
|
690 |
+
"model.layers.7.self_attn.v_proj.weight": "model-00001-of-00004.safetensors",
|
691 |
+
"model.layers.7.self_attn.v_proj.weight_scale": "model-00001-of-00004.safetensors",
|
692 |
+
"model.layers.8.input_layernorm.weight": "model-00001-of-00004.safetensors",
|
693 |
+
"model.layers.8.mlp.down_proj.weight": "model-00001-of-00004.safetensors",
|
694 |
+
"model.layers.8.mlp.down_proj.weight_scale": "model-00001-of-00004.safetensors",
|
695 |
+
"model.layers.8.mlp.gate_proj.weight": "model-00001-of-00004.safetensors",
|
696 |
+
"model.layers.8.mlp.gate_proj.weight_scale": "model-00001-of-00004.safetensors",
|
697 |
+
"model.layers.8.mlp.up_proj.weight": "model-00001-of-00004.safetensors",
|
698 |
+
"model.layers.8.mlp.up_proj.weight_scale": "model-00001-of-00004.safetensors",
|
699 |
+
"model.layers.8.post_attention_layernorm.weight": "model-00001-of-00004.safetensors",
|
700 |
+
"model.layers.8.self_attn.k_norm.weight": "model-00001-of-00004.safetensors",
|
701 |
+
"model.layers.8.self_attn.k_proj.weight": "model-00001-of-00004.safetensors",
|
702 |
+
"model.layers.8.self_attn.k_proj.weight_scale": "model-00001-of-00004.safetensors",
|
703 |
+
"model.layers.8.self_attn.o_proj.weight": "model-00001-of-00004.safetensors",
|
704 |
+
"model.layers.8.self_attn.o_proj.weight_scale": "model-00001-of-00004.safetensors",
|
705 |
+
"model.layers.8.self_attn.q_norm.weight": "model-00001-of-00004.safetensors",
|
706 |
+
"model.layers.8.self_attn.q_proj.weight": "model-00001-of-00004.safetensors",
|
707 |
+
"model.layers.8.self_attn.q_proj.weight_scale": "model-00001-of-00004.safetensors",
|
708 |
+
"model.layers.8.self_attn.v_proj.weight": "model-00001-of-00004.safetensors",
|
709 |
+
"model.layers.8.self_attn.v_proj.weight_scale": "model-00001-of-00004.safetensors",
|
710 |
+
"model.layers.9.input_layernorm.weight": "model-00001-of-00004.safetensors",
|
711 |
+
"model.layers.9.mlp.down_proj.weight": "model-00001-of-00004.safetensors",
|
712 |
+
"model.layers.9.mlp.down_proj.weight_scale": "model-00001-of-00004.safetensors",
|
713 |
+
"model.layers.9.mlp.gate_proj.weight": "model-00001-of-00004.safetensors",
|
714 |
+
"model.layers.9.mlp.gate_proj.weight_scale": "model-00001-of-00004.safetensors",
|
715 |
+
"model.layers.9.mlp.up_proj.weight": "model-00001-of-00004.safetensors",
|
716 |
+
"model.layers.9.mlp.up_proj.weight_scale": "model-00001-of-00004.safetensors",
|
717 |
+
"model.layers.9.post_attention_layernorm.weight": "model-00001-of-00004.safetensors",
|
718 |
+
"model.layers.9.self_attn.k_norm.weight": "model-00001-of-00004.safetensors",
|
719 |
+
"model.layers.9.self_attn.k_proj.weight": "model-00001-of-00004.safetensors",
|
720 |
+
"model.layers.9.self_attn.k_proj.weight_scale": "model-00001-of-00004.safetensors",
|
721 |
+
"model.layers.9.self_attn.o_proj.weight": "model-00001-of-00004.safetensors",
|
722 |
+
"model.layers.9.self_attn.o_proj.weight_scale": "model-00001-of-00004.safetensors",
|
723 |
+
"model.layers.9.self_attn.q_norm.weight": "model-00001-of-00004.safetensors",
|
724 |
+
"model.layers.9.self_attn.q_proj.weight": "model-00001-of-00004.safetensors",
|
725 |
+
"model.layers.9.self_attn.q_proj.weight_scale": "model-00001-of-00004.safetensors",
|
726 |
+
"model.layers.9.self_attn.v_proj.weight": "model-00001-of-00004.safetensors",
|
727 |
+
"model.layers.9.self_attn.v_proj.weight_scale": "model-00001-of-00004.safetensors",
|
728 |
+
"model.norm.weight": "model-00003-of-00004.safetensors"
|
729 |
+
}
|
730 |
+
}
|
recipe.yaml
ADDED
@@ -0,0 +1,6 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
default_stage:
|
2 |
+
default_modifiers:
|
3 |
+
QuantizationModifier:
|
4 |
+
ignore: [lm_head]
|
5 |
+
targets: [Linear]
|
6 |
+
scheme: FP8_DYNAMIC
|
special_tokens_map.json
ADDED
@@ -0,0 +1,1009 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"additional_special_tokens": [
|
3 |
+
"[control_10]",
|
4 |
+
"[control_11]",
|
5 |
+
"[control_12]",
|
6 |
+
"[control_13]",
|
7 |
+
"[control_14]",
|
8 |
+
"[control_15]",
|
9 |
+
"[control_16]",
|
10 |
+
"[control_17]",
|
11 |
+
"[control_18]",
|
12 |
+
"[control_19]",
|
13 |
+
"[control_20]",
|
14 |
+
"[control_21]",
|
15 |
+
"[control_22]",
|
16 |
+
"[control_23]",
|
17 |
+
"[control_24]",
|
18 |
+
"[control_25]",
|
19 |
+
"[control_26]",
|
20 |
+
"[control_27]",
|
21 |
+
"[control_28]",
|
22 |
+
"[control_29]",
|
23 |
+
"[control_30]",
|
24 |
+
"[control_31]",
|
25 |
+
"[control_32]",
|
26 |
+
"[control_33]",
|
27 |
+
"[control_34]",
|
28 |
+
"[control_35]",
|
29 |
+
"[control_36]",
|
30 |
+
"[control_37]",
|
31 |
+
"[control_38]",
|
32 |
+
"[control_39]",
|
33 |
+
"[control_40]",
|
34 |
+
"[control_41]",
|
35 |
+
"[control_42]",
|
36 |
+
"[control_43]",
|
37 |
+
"[control_44]",
|
38 |
+
"[control_45]",
|
39 |
+
"[control_46]",
|
40 |
+
"[control_47]",
|
41 |
+
"[control_48]",
|
42 |
+
"[control_49]",
|
43 |
+
"[control_50]",
|
44 |
+
"[control_51]",
|
45 |
+
"[control_52]",
|
46 |
+
"[control_53]",
|
47 |
+
"[control_54]",
|
48 |
+
"[control_55]",
|
49 |
+
"[control_56]",
|
50 |
+
"[control_57]",
|
51 |
+
"[control_58]",
|
52 |
+
"[control_59]",
|
53 |
+
"[control_60]",
|
54 |
+
"[control_61]",
|
55 |
+
"[control_62]",
|
56 |
+
"[control_63]",
|
57 |
+
"[control_64]",
|
58 |
+
"[control_65]",
|
59 |
+
"[control_66]",
|
60 |
+
"[control_67]",
|
61 |
+
"[control_68]",
|
62 |
+
"[control_69]",
|
63 |
+
"[control_70]",
|
64 |
+
"[control_71]",
|
65 |
+
"[control_72]",
|
66 |
+
"[control_73]",
|
67 |
+
"[control_74]",
|
68 |
+
"[control_75]",
|
69 |
+
"[control_76]",
|
70 |
+
"[control_77]",
|
71 |
+
"[control_78]",
|
72 |
+
"[control_79]",
|
73 |
+
"[control_80]",
|
74 |
+
"[control_81]",
|
75 |
+
"[control_82]",
|
76 |
+
"[control_83]",
|
77 |
+
"[control_84]",
|
78 |
+
"[control_85]",
|
79 |
+
"[control_86]",
|
80 |
+
"[control_87]",
|
81 |
+
"[control_88]",
|
82 |
+
"[control_89]",
|
83 |
+
"[control_90]",
|
84 |
+
"[control_91]",
|
85 |
+
"[control_92]",
|
86 |
+
"[control_93]",
|
87 |
+
"[control_94]",
|
88 |
+
"[control_95]",
|
89 |
+
"[control_96]",
|
90 |
+
"[control_97]",
|
91 |
+
"[control_98]",
|
92 |
+
"[control_99]",
|
93 |
+
"[control_100]",
|
94 |
+
"[control_101]",
|
95 |
+
"[control_102]",
|
96 |
+
"[control_103]",
|
97 |
+
"[control_104]",
|
98 |
+
"[control_105]",
|
99 |
+
"[control_106]",
|
100 |
+
"[control_107]",
|
101 |
+
"[control_108]",
|
102 |
+
"[control_109]",
|
103 |
+
"[control_110]",
|
104 |
+
"[control_111]",
|
105 |
+
"[control_112]",
|
106 |
+
"[control_113]",
|
107 |
+
"[control_114]",
|
108 |
+
"[control_115]",
|
109 |
+
"[control_116]",
|
110 |
+
"[control_117]",
|
111 |
+
"[control_118]",
|
112 |
+
"[control_119]",
|
113 |
+
"[control_120]",
|
114 |
+
"[control_121]",
|
115 |
+
"[control_122]",
|
116 |
+
"[control_123]",
|
117 |
+
"[control_124]",
|
118 |
+
"[control_125]",
|
119 |
+
"[control_126]",
|
120 |
+
"[control_127]",
|
121 |
+
"[control_128]",
|
122 |
+
"[control_129]",
|
123 |
+
"[control_130]",
|
124 |
+
"[control_131]",
|
125 |
+
"[control_132]",
|
126 |
+
"[control_133]",
|
127 |
+
"[control_134]",
|
128 |
+
"[control_135]",
|
129 |
+
"[control_136]",
|
130 |
+
"[control_137]",
|
131 |
+
"[control_138]",
|
132 |
+
"[control_139]",
|
133 |
+
"[control_140]",
|
134 |
+
"[control_141]",
|
135 |
+
"[control_142]",
|
136 |
+
"[control_143]",
|
137 |
+
"[control_144]",
|
138 |
+
"[control_145]",
|
139 |
+
"[control_146]",
|
140 |
+
"[control_147]",
|
141 |
+
"[control_148]",
|
142 |
+
"[control_149]",
|
143 |
+
"[control_150]",
|
144 |
+
"[control_151]",
|
145 |
+
"[control_152]",
|
146 |
+
"[control_153]",
|
147 |
+
"[control_154]",
|
148 |
+
"[control_155]",
|
149 |
+
"[control_156]",
|
150 |
+
"[control_157]",
|
151 |
+
"[control_158]",
|
152 |
+
"[control_159]",
|
153 |
+
"[control_160]",
|
154 |
+
"[control_161]",
|
155 |
+
"[control_162]",
|
156 |
+
"[control_163]",
|
157 |
+
"[control_164]",
|
158 |
+
"[control_165]",
|
159 |
+
"[control_166]",
|
160 |
+
"[control_167]",
|
161 |
+
"[control_168]",
|
162 |
+
"[control_169]",
|
163 |
+
"[control_170]",
|
164 |
+
"[control_171]",
|
165 |
+
"[control_172]",
|
166 |
+
"[control_173]",
|
167 |
+
"[control_174]",
|
168 |
+
"[control_175]",
|
169 |
+
"[control_176]",
|
170 |
+
"[control_177]",
|
171 |
+
"[control_178]",
|
172 |
+
"[control_179]",
|
173 |
+
"[control_180]",
|
174 |
+
"[control_181]",
|
175 |
+
"[control_182]",
|
176 |
+
"[control_183]",
|
177 |
+
"[control_184]",
|
178 |
+
"[control_185]",
|
179 |
+
"[control_186]",
|
180 |
+
"[control_187]",
|
181 |
+
"[control_188]",
|
182 |
+
"[control_189]",
|
183 |
+
"[control_190]",
|
184 |
+
"[control_191]",
|
185 |
+
"[control_192]",
|
186 |
+
"[control_193]",
|
187 |
+
"[control_194]",
|
188 |
+
"[control_195]",
|
189 |
+
"[control_196]",
|
190 |
+
"[control_197]",
|
191 |
+
"[control_198]",
|
192 |
+
"[control_199]",
|
193 |
+
"[control_200]",
|
194 |
+
"[control_201]",
|
195 |
+
"[control_202]",
|
196 |
+
"[control_203]",
|
197 |
+
"[control_204]",
|
198 |
+
"[control_205]",
|
199 |
+
"[control_206]",
|
200 |
+
"[control_207]",
|
201 |
+
"[control_208]",
|
202 |
+
"[control_209]",
|
203 |
+
"[control_210]",
|
204 |
+
"[control_211]",
|
205 |
+
"[control_212]",
|
206 |
+
"[control_213]",
|
207 |
+
"[control_214]",
|
208 |
+
"[control_215]",
|
209 |
+
"[control_216]",
|
210 |
+
"[control_217]",
|
211 |
+
"[control_218]",
|
212 |
+
"[control_219]",
|
213 |
+
"[control_220]",
|
214 |
+
"[control_221]",
|
215 |
+
"[control_222]",
|
216 |
+
"[control_223]",
|
217 |
+
"[control_224]",
|
218 |
+
"[control_225]",
|
219 |
+
"[control_226]",
|
220 |
+
"[control_227]",
|
221 |
+
"[control_228]",
|
222 |
+
"[control_229]",
|
223 |
+
"[control_230]",
|
224 |
+
"[control_231]",
|
225 |
+
"[control_232]",
|
226 |
+
"[control_233]",
|
227 |
+
"[control_234]",
|
228 |
+
"[control_235]",
|
229 |
+
"[control_236]",
|
230 |
+
"[control_237]",
|
231 |
+
"[control_238]",
|
232 |
+
"[control_239]",
|
233 |
+
"[control_240]",
|
234 |
+
"[control_241]",
|
235 |
+
"[control_242]",
|
236 |
+
"[control_243]",
|
237 |
+
"[control_244]",
|
238 |
+
"[control_245]",
|
239 |
+
"[control_246]",
|
240 |
+
"[control_247]",
|
241 |
+
"[control_248]",
|
242 |
+
"[control_249]",
|
243 |
+
"[control_250]",
|
244 |
+
"[control_251]",
|
245 |
+
"[control_252]",
|
246 |
+
"[control_253]",
|
247 |
+
"[control_254]",
|
248 |
+
"[control_255]",
|
249 |
+
"[control_256]",
|
250 |
+
"[control_257]",
|
251 |
+
"[control_258]",
|
252 |
+
"[control_259]",
|
253 |
+
"[control_260]",
|
254 |
+
"[control_261]",
|
255 |
+
"[control_262]",
|
256 |
+
"[control_263]",
|
257 |
+
"[control_264]",
|
258 |
+
"[control_265]",
|
259 |
+
"[control_266]",
|
260 |
+
"[control_267]",
|
261 |
+
"[control_268]",
|
262 |
+
"[control_269]",
|
263 |
+
"[control_270]",
|
264 |
+
"[control_271]",
|
265 |
+
"[control_272]",
|
266 |
+
"[control_273]",
|
267 |
+
"[control_274]",
|
268 |
+
"[control_275]",
|
269 |
+
"[control_276]",
|
270 |
+
"[control_277]",
|
271 |
+
"[control_278]",
|
272 |
+
"[control_279]",
|
273 |
+
"[control_280]",
|
274 |
+
"[control_281]",
|
275 |
+
"[control_282]",
|
276 |
+
"[control_283]",
|
277 |
+
"[control_284]",
|
278 |
+
"[control_285]",
|
279 |
+
"[control_286]",
|
280 |
+
"[control_287]",
|
281 |
+
"[control_288]",
|
282 |
+
"[control_289]",
|
283 |
+
"[control_290]",
|
284 |
+
"[control_291]",
|
285 |
+
"[control_292]",
|
286 |
+
"[control_293]",
|
287 |
+
"[control_294]",
|
288 |
+
"[control_295]",
|
289 |
+
"[control_296]",
|
290 |
+
"[control_297]",
|
291 |
+
"[control_298]",
|
292 |
+
"[control_299]",
|
293 |
+
"[control_300]",
|
294 |
+
"[control_301]",
|
295 |
+
"[control_302]",
|
296 |
+
"[control_303]",
|
297 |
+
"[control_304]",
|
298 |
+
"[control_305]",
|
299 |
+
"[control_306]",
|
300 |
+
"[control_307]",
|
301 |
+
"[control_308]",
|
302 |
+
"[control_309]",
|
303 |
+
"[control_310]",
|
304 |
+
"[control_311]",
|
305 |
+
"[control_312]",
|
306 |
+
"[control_313]",
|
307 |
+
"[control_314]",
|
308 |
+
"[control_315]",
|
309 |
+
"[control_316]",
|
310 |
+
"[control_317]",
|
311 |
+
"[control_318]",
|
312 |
+
"[control_319]",
|
313 |
+
"[control_320]",
|
314 |
+
"[control_321]",
|
315 |
+
"[control_322]",
|
316 |
+
"[control_323]",
|
317 |
+
"[control_324]",
|
318 |
+
"[control_325]",
|
319 |
+
"[control_326]",
|
320 |
+
"[control_327]",
|
321 |
+
"[control_328]",
|
322 |
+
"[control_329]",
|
323 |
+
"[control_330]",
|
324 |
+
"[control_331]",
|
325 |
+
"[control_332]",
|
326 |
+
"[control_333]",
|
327 |
+
"[control_334]",
|
328 |
+
"[control_335]",
|
329 |
+
"[control_336]",
|
330 |
+
"[control_337]",
|
331 |
+
"[control_338]",
|
332 |
+
"[control_339]",
|
333 |
+
"[control_340]",
|
334 |
+
"[control_341]",
|
335 |
+
"[control_342]",
|
336 |
+
"[control_343]",
|
337 |
+
"[control_344]",
|
338 |
+
"[control_345]",
|
339 |
+
"[control_346]",
|
340 |
+
"[control_347]",
|
341 |
+
"[control_348]",
|
342 |
+
"[control_349]",
|
343 |
+
"[control_350]",
|
344 |
+
"[control_351]",
|
345 |
+
"[control_352]",
|
346 |
+
"[control_353]",
|
347 |
+
"[control_354]",
|
348 |
+
"[control_355]",
|
349 |
+
"[control_356]",
|
350 |
+
"[control_357]",
|
351 |
+
"[control_358]",
|
352 |
+
"[control_359]",
|
353 |
+
"[control_360]",
|
354 |
+
"[control_361]",
|
355 |
+
"[control_362]",
|
356 |
+
"[control_363]",
|
357 |
+
"[control_364]",
|
358 |
+
"[control_365]",
|
359 |
+
"[control_366]",
|
360 |
+
"[control_367]",
|
361 |
+
"[control_368]",
|
362 |
+
"[control_369]",
|
363 |
+
"[control_370]",
|
364 |
+
"[control_371]",
|
365 |
+
"[control_372]",
|
366 |
+
"[control_373]",
|
367 |
+
"[control_374]",
|
368 |
+
"[control_375]",
|
369 |
+
"[control_376]",
|
370 |
+
"[control_377]",
|
371 |
+
"[control_378]",
|
372 |
+
"[control_379]",
|
373 |
+
"[control_380]",
|
374 |
+
"[control_381]",
|
375 |
+
"[control_382]",
|
376 |
+
"[control_383]",
|
377 |
+
"[control_384]",
|
378 |
+
"[control_385]",
|
379 |
+
"[control_386]",
|
380 |
+
"[control_387]",
|
381 |
+
"[control_388]",
|
382 |
+
"[control_389]",
|
383 |
+
"[control_390]",
|
384 |
+
"[control_391]",
|
385 |
+
"[control_392]",
|
386 |
+
"[control_393]",
|
387 |
+
"[control_394]",
|
388 |
+
"[control_395]",
|
389 |
+
"[control_396]",
|
390 |
+
"[control_397]",
|
391 |
+
"[control_398]",
|
392 |
+
"[control_399]",
|
393 |
+
"[control_400]",
|
394 |
+
"[control_401]",
|
395 |
+
"[control_402]",
|
396 |
+
"[control_403]",
|
397 |
+
"[control_404]",
|
398 |
+
"[control_405]",
|
399 |
+
"[control_406]",
|
400 |
+
"[control_407]",
|
401 |
+
"[control_408]",
|
402 |
+
"[control_409]",
|
403 |
+
"[control_410]",
|
404 |
+
"[control_411]",
|
405 |
+
"[control_412]",
|
406 |
+
"[control_413]",
|
407 |
+
"[control_414]",
|
408 |
+
"[control_415]",
|
409 |
+
"[control_416]",
|
410 |
+
"[control_417]",
|
411 |
+
"[control_418]",
|
412 |
+
"[control_419]",
|
413 |
+
"[control_420]",
|
414 |
+
"[control_421]",
|
415 |
+
"[control_422]",
|
416 |
+
"[control_423]",
|
417 |
+
"[control_424]",
|
418 |
+
"[control_425]",
|
419 |
+
"[control_426]",
|
420 |
+
"[control_427]",
|
421 |
+
"[control_428]",
|
422 |
+
"[control_429]",
|
423 |
+
"[control_430]",
|
424 |
+
"[control_431]",
|
425 |
+
"[control_432]",
|
426 |
+
"[control_433]",
|
427 |
+
"[control_434]",
|
428 |
+
"[control_435]",
|
429 |
+
"[control_436]",
|
430 |
+
"[control_437]",
|
431 |
+
"[control_438]",
|
432 |
+
"[control_439]",
|
433 |
+
"[control_440]",
|
434 |
+
"[control_441]",
|
435 |
+
"[control_442]",
|
436 |
+
"[control_443]",
|
437 |
+
"[control_444]",
|
438 |
+
"[control_445]",
|
439 |
+
"[control_446]",
|
440 |
+
"[control_447]",
|
441 |
+
"[control_448]",
|
442 |
+
"[control_449]",
|
443 |
+
"[control_450]",
|
444 |
+
"[control_451]",
|
445 |
+
"[control_452]",
|
446 |
+
"[control_453]",
|
447 |
+
"[control_454]",
|
448 |
+
"[control_455]",
|
449 |
+
"[control_456]",
|
450 |
+
"[control_457]",
|
451 |
+
"[control_458]",
|
452 |
+
"[control_459]",
|
453 |
+
"[control_460]",
|
454 |
+
"[control_461]",
|
455 |
+
"[control_462]",
|
456 |
+
"[control_463]",
|
457 |
+
"[control_464]",
|
458 |
+
"[control_465]",
|
459 |
+
"[control_466]",
|
460 |
+
"[control_467]",
|
461 |
+
"[control_468]",
|
462 |
+
"[control_469]",
|
463 |
+
"[control_470]",
|
464 |
+
"[control_471]",
|
465 |
+
"[control_472]",
|
466 |
+
"[control_473]",
|
467 |
+
"[control_474]",
|
468 |
+
"[control_475]",
|
469 |
+
"[control_476]",
|
470 |
+
"[control_477]",
|
471 |
+
"[control_478]",
|
472 |
+
"[control_479]",
|
473 |
+
"[control_480]",
|
474 |
+
"[control_481]",
|
475 |
+
"[control_482]",
|
476 |
+
"[control_483]",
|
477 |
+
"[control_484]",
|
478 |
+
"[control_485]",
|
479 |
+
"[control_486]",
|
480 |
+
"[control_487]",
|
481 |
+
"[control_488]",
|
482 |
+
"[control_489]",
|
483 |
+
"[control_490]",
|
484 |
+
"[control_491]",
|
485 |
+
"[control_492]",
|
486 |
+
"[control_493]",
|
487 |
+
"[control_494]",
|
488 |
+
"[control_495]",
|
489 |
+
"[control_496]",
|
490 |
+
"[control_497]",
|
491 |
+
"[control_498]",
|
492 |
+
"[control_499]",
|
493 |
+
"[control_500]",
|
494 |
+
"[control_501]",
|
495 |
+
"[control_502]",
|
496 |
+
"[control_503]",
|
497 |
+
"[control_504]",
|
498 |
+
"[control_505]",
|
499 |
+
"[control_506]",
|
500 |
+
"[control_507]",
|
501 |
+
"[control_508]",
|
502 |
+
"[control_509]",
|
503 |
+
"[control_510]",
|
504 |
+
"[control_511]",
|
505 |
+
"[control_512]",
|
506 |
+
"[control_513]",
|
507 |
+
"[control_514]",
|
508 |
+
"[control_515]",
|
509 |
+
"[control_516]",
|
510 |
+
"[control_517]",
|
511 |
+
"[control_518]",
|
512 |
+
"[control_519]",
|
513 |
+
"[control_520]",
|
514 |
+
"[control_521]",
|
515 |
+
"[control_522]",
|
516 |
+
"[control_523]",
|
517 |
+
"[control_524]",
|
518 |
+
"[control_525]",
|
519 |
+
"[control_526]",
|
520 |
+
"[control_527]",
|
521 |
+
"[control_528]",
|
522 |
+
"[control_529]",
|
523 |
+
"[control_530]",
|
524 |
+
"[control_531]",
|
525 |
+
"[control_532]",
|
526 |
+
"[control_533]",
|
527 |
+
"[control_534]",
|
528 |
+
"[control_535]",
|
529 |
+
"[control_536]",
|
530 |
+
"[control_537]",
|
531 |
+
"[control_538]",
|
532 |
+
"[control_539]",
|
533 |
+
"[control_540]",
|
534 |
+
"[control_541]",
|
535 |
+
"[control_542]",
|
536 |
+
"[control_543]",
|
537 |
+
"[control_544]",
|
538 |
+
"[control_545]",
|
539 |
+
"[control_546]",
|
540 |
+
"[control_547]",
|
541 |
+
"[control_548]",
|
542 |
+
"[control_549]",
|
543 |
+
"[control_550]",
|
544 |
+
"[control_551]",
|
545 |
+
"[control_552]",
|
546 |
+
"[control_553]",
|
547 |
+
"[control_554]",
|
548 |
+
"[control_555]",
|
549 |
+
"[control_556]",
|
550 |
+
"[control_557]",
|
551 |
+
"[control_558]",
|
552 |
+
"[control_559]",
|
553 |
+
"[control_560]",
|
554 |
+
"[control_561]",
|
555 |
+
"[control_562]",
|
556 |
+
"[control_563]",
|
557 |
+
"[control_564]",
|
558 |
+
"[control_565]",
|
559 |
+
"[control_566]",
|
560 |
+
"[control_567]",
|
561 |
+
"[control_568]",
|
562 |
+
"[control_569]",
|
563 |
+
"[control_570]",
|
564 |
+
"[control_571]",
|
565 |
+
"[control_572]",
|
566 |
+
"[control_573]",
|
567 |
+
"[control_574]",
|
568 |
+
"[control_575]",
|
569 |
+
"[control_576]",
|
570 |
+
"[control_577]",
|
571 |
+
"[control_578]",
|
572 |
+
"[control_579]",
|
573 |
+
"[control_580]",
|
574 |
+
"[control_581]",
|
575 |
+
"[control_582]",
|
576 |
+
"[control_583]",
|
577 |
+
"[control_584]",
|
578 |
+
"[control_585]",
|
579 |
+
"[control_586]",
|
580 |
+
"[control_587]",
|
581 |
+
"[control_588]",
|
582 |
+
"[control_589]",
|
583 |
+
"[control_590]",
|
584 |
+
"[control_591]",
|
585 |
+
"[control_592]",
|
586 |
+
"[control_593]",
|
587 |
+
"[control_594]",
|
588 |
+
"[control_595]",
|
589 |
+
"[control_596]",
|
590 |
+
"[control_597]",
|
591 |
+
"[control_598]",
|
592 |
+
"[control_599]",
|
593 |
+
"[control_600]",
|
594 |
+
"[control_601]",
|
595 |
+
"[control_602]",
|
596 |
+
"[control_603]",
|
597 |
+
"[control_604]",
|
598 |
+
"[control_605]",
|
599 |
+
"[control_606]",
|
600 |
+
"[control_607]",
|
601 |
+
"[control_608]",
|
602 |
+
"[control_609]",
|
603 |
+
"[control_610]",
|
604 |
+
"[control_611]",
|
605 |
+
"[control_612]",
|
606 |
+
"[control_613]",
|
607 |
+
"[control_614]",
|
608 |
+
"[control_615]",
|
609 |
+
"[control_616]",
|
610 |
+
"[control_617]",
|
611 |
+
"[control_618]",
|
612 |
+
"[control_619]",
|
613 |
+
"[control_620]",
|
614 |
+
"[control_621]",
|
615 |
+
"[control_622]",
|
616 |
+
"[control_623]",
|
617 |
+
"[control_624]",
|
618 |
+
"[control_625]",
|
619 |
+
"[control_626]",
|
620 |
+
"[control_627]",
|
621 |
+
"[control_628]",
|
622 |
+
"[control_629]",
|
623 |
+
"[control_630]",
|
624 |
+
"[control_631]",
|
625 |
+
"[control_632]",
|
626 |
+
"[control_633]",
|
627 |
+
"[control_634]",
|
628 |
+
"[control_635]",
|
629 |
+
"[control_636]",
|
630 |
+
"[control_637]",
|
631 |
+
"[control_638]",
|
632 |
+
"[control_639]",
|
633 |
+
"[control_640]",
|
634 |
+
"[control_641]",
|
635 |
+
"[control_642]",
|
636 |
+
"[control_643]",
|
637 |
+
"[control_644]",
|
638 |
+
"[control_645]",
|
639 |
+
"[control_646]",
|
640 |
+
"[control_647]",
|
641 |
+
"[control_648]",
|
642 |
+
"[control_649]",
|
643 |
+
"[control_650]",
|
644 |
+
"[control_651]",
|
645 |
+
"[control_652]",
|
646 |
+
"[control_653]",
|
647 |
+
"[control_654]",
|
648 |
+
"[control_655]",
|
649 |
+
"[control_656]",
|
650 |
+
"[control_657]",
|
651 |
+
"[control_658]",
|
652 |
+
"[control_659]",
|
653 |
+
"[control_660]",
|
654 |
+
"[control_661]",
|
655 |
+
"[control_662]",
|
656 |
+
"[control_663]",
|
657 |
+
"[control_664]",
|
658 |
+
"[control_665]",
|
659 |
+
"[control_666]",
|
660 |
+
"[control_667]",
|
661 |
+
"[control_668]",
|
662 |
+
"[control_669]",
|
663 |
+
"[control_670]",
|
664 |
+
"[control_671]",
|
665 |
+
"[control_672]",
|
666 |
+
"[control_673]",
|
667 |
+
"[control_674]",
|
668 |
+
"[control_675]",
|
669 |
+
"[control_676]",
|
670 |
+
"[control_677]",
|
671 |
+
"[control_678]",
|
672 |
+
"[control_679]",
|
673 |
+
"[control_680]",
|
674 |
+
"[control_681]",
|
675 |
+
"[control_682]",
|
676 |
+
"[control_683]",
|
677 |
+
"[control_684]",
|
678 |
+
"[control_685]",
|
679 |
+
"[control_686]",
|
680 |
+
"[control_687]",
|
681 |
+
"[control_688]",
|
682 |
+
"[control_689]",
|
683 |
+
"[control_690]",
|
684 |
+
"[control_691]",
|
685 |
+
"[control_692]",
|
686 |
+
"[control_693]",
|
687 |
+
"[control_694]",
|
688 |
+
"[control_695]",
|
689 |
+
"[control_696]",
|
690 |
+
"[control_697]",
|
691 |
+
"[control_698]",
|
692 |
+
"[control_699]",
|
693 |
+
"[control_700]",
|
694 |
+
"[control_701]",
|
695 |
+
"[control_702]",
|
696 |
+
"[control_703]",
|
697 |
+
"[control_704]",
|
698 |
+
"[control_705]",
|
699 |
+
"[control_706]",
|
700 |
+
"[control_707]",
|
701 |
+
"[control_708]",
|
702 |
+
"[control_709]",
|
703 |
+
"[control_710]",
|
704 |
+
"[control_711]",
|
705 |
+
"[control_712]",
|
706 |
+
"[control_713]",
|
707 |
+
"[control_714]",
|
708 |
+
"[control_715]",
|
709 |
+
"[control_716]",
|
710 |
+
"[control_717]",
|
711 |
+
"[control_718]",
|
712 |
+
"[control_719]",
|
713 |
+
"[control_720]",
|
714 |
+
"[control_721]",
|
715 |
+
"[control_722]",
|
716 |
+
"[control_723]",
|
717 |
+
"[control_724]",
|
718 |
+
"[control_725]",
|
719 |
+
"[control_726]",
|
720 |
+
"[control_727]",
|
721 |
+
"[control_728]",
|
722 |
+
"[control_729]",
|
723 |
+
"[control_730]",
|
724 |
+
"[control_731]",
|
725 |
+
"[control_732]",
|
726 |
+
"[control_733]",
|
727 |
+
"[control_734]",
|
728 |
+
"[control_735]",
|
729 |
+
"[control_736]",
|
730 |
+
"[control_737]",
|
731 |
+
"[control_738]",
|
732 |
+
"[control_739]",
|
733 |
+
"[control_740]",
|
734 |
+
"[control_741]",
|
735 |
+
"[control_742]",
|
736 |
+
"[control_743]",
|
737 |
+
"[control_744]",
|
738 |
+
"[control_745]",
|
739 |
+
"[control_746]",
|
740 |
+
"[control_747]",
|
741 |
+
"[control_748]",
|
742 |
+
"[control_749]",
|
743 |
+
"[control_750]",
|
744 |
+
"[control_751]",
|
745 |
+
"[control_752]",
|
746 |
+
"[control_753]",
|
747 |
+
"[control_754]",
|
748 |
+
"[control_755]",
|
749 |
+
"[control_756]",
|
750 |
+
"[control_757]",
|
751 |
+
"[control_758]",
|
752 |
+
"[control_759]",
|
753 |
+
"[control_760]",
|
754 |
+
"[control_761]",
|
755 |
+
"[control_762]",
|
756 |
+
"[control_763]",
|
757 |
+
"[control_764]",
|
758 |
+
"[control_765]",
|
759 |
+
"[control_766]",
|
760 |
+
"[control_767]",
|
761 |
+
"[control_768]",
|
762 |
+
"[control_769]",
|
763 |
+
"[control_770]",
|
764 |
+
"[control_771]",
|
765 |
+
"[control_772]",
|
766 |
+
"[control_773]",
|
767 |
+
"[control_774]",
|
768 |
+
"[control_775]",
|
769 |
+
"[control_776]",
|
770 |
+
"[control_777]",
|
771 |
+
"[control_778]",
|
772 |
+
"[control_779]",
|
773 |
+
"[control_780]",
|
774 |
+
"[control_781]",
|
775 |
+
"[control_782]",
|
776 |
+
"[control_783]",
|
777 |
+
"[control_784]",
|
778 |
+
"[control_785]",
|
779 |
+
"[control_786]",
|
780 |
+
"[control_787]",
|
781 |
+
"[control_788]",
|
782 |
+
"[control_789]",
|
783 |
+
"[control_790]",
|
784 |
+
"[control_791]",
|
785 |
+
"[control_792]",
|
786 |
+
"[control_793]",
|
787 |
+
"[control_794]",
|
788 |
+
"[control_795]",
|
789 |
+
"[control_796]",
|
790 |
+
"[control_797]",
|
791 |
+
"[control_798]",
|
792 |
+
"[control_799]",
|
793 |
+
"[control_800]",
|
794 |
+
"[control_801]",
|
795 |
+
"[control_802]",
|
796 |
+
"[control_803]",
|
797 |
+
"[control_804]",
|
798 |
+
"[control_805]",
|
799 |
+
"[control_806]",
|
800 |
+
"[control_807]",
|
801 |
+
"[control_808]",
|
802 |
+
"[control_809]",
|
803 |
+
"[control_810]",
|
804 |
+
"[control_811]",
|
805 |
+
"[control_812]",
|
806 |
+
"[control_813]",
|
807 |
+
"[control_814]",
|
808 |
+
"[control_815]",
|
809 |
+
"[control_816]",
|
810 |
+
"[control_817]",
|
811 |
+
"[control_818]",
|
812 |
+
"[control_819]",
|
813 |
+
"[control_820]",
|
814 |
+
"[control_821]",
|
815 |
+
"[control_822]",
|
816 |
+
"[control_823]",
|
817 |
+
"[control_824]",
|
818 |
+
"[control_825]",
|
819 |
+
"[control_826]",
|
820 |
+
"[control_827]",
|
821 |
+
"[control_828]",
|
822 |
+
"[control_829]",
|
823 |
+
"[control_830]",
|
824 |
+
"[control_831]",
|
825 |
+
"[control_832]",
|
826 |
+
"[control_833]",
|
827 |
+
"[control_834]",
|
828 |
+
"[control_835]",
|
829 |
+
"[control_836]",
|
830 |
+
"[control_837]",
|
831 |
+
"[control_838]",
|
832 |
+
"[control_839]",
|
833 |
+
"[control_840]",
|
834 |
+
"[control_841]",
|
835 |
+
"[control_842]",
|
836 |
+
"[control_843]",
|
837 |
+
"[control_844]",
|
838 |
+
"[control_845]",
|
839 |
+
"[control_846]",
|
840 |
+
"[control_847]",
|
841 |
+
"[control_848]",
|
842 |
+
"[control_849]",
|
843 |
+
"[control_850]",
|
844 |
+
"[control_851]",
|
845 |
+
"[control_852]",
|
846 |
+
"[control_853]",
|
847 |
+
"[control_854]",
|
848 |
+
"[control_855]",
|
849 |
+
"[control_856]",
|
850 |
+
"[control_857]",
|
851 |
+
"[control_858]",
|
852 |
+
"[control_859]",
|
853 |
+
"[control_860]",
|
854 |
+
"[control_861]",
|
855 |
+
"[control_862]",
|
856 |
+
"[control_863]",
|
857 |
+
"[control_864]",
|
858 |
+
"[control_865]",
|
859 |
+
"[control_866]",
|
860 |
+
"[control_867]",
|
861 |
+
"[control_868]",
|
862 |
+
"[control_869]",
|
863 |
+
"[control_870]",
|
864 |
+
"[control_871]",
|
865 |
+
"[control_872]",
|
866 |
+
"[control_873]",
|
867 |
+
"[control_874]",
|
868 |
+
"[control_875]",
|
869 |
+
"[control_876]",
|
870 |
+
"[control_877]",
|
871 |
+
"[control_878]",
|
872 |
+
"[control_879]",
|
873 |
+
"[control_880]",
|
874 |
+
"[control_881]",
|
875 |
+
"[control_882]",
|
876 |
+
"[control_883]",
|
877 |
+
"[control_884]",
|
878 |
+
"[control_885]",
|
879 |
+
"[control_886]",
|
880 |
+
"[control_887]",
|
881 |
+
"[control_888]",
|
882 |
+
"[control_889]",
|
883 |
+
"[control_890]",
|
884 |
+
"[control_891]",
|
885 |
+
"[control_892]",
|
886 |
+
"[control_893]",
|
887 |
+
"[control_894]",
|
888 |
+
"[control_895]",
|
889 |
+
"[control_896]",
|
890 |
+
"[control_897]",
|
891 |
+
"[control_898]",
|
892 |
+
"[control_899]",
|
893 |
+
"[control_900]",
|
894 |
+
"[control_901]",
|
895 |
+
"[control_902]",
|
896 |
+
"[control_903]",
|
897 |
+
"[control_904]",
|
898 |
+
"[control_905]",
|
899 |
+
"[control_906]",
|
900 |
+
"[control_907]",
|
901 |
+
"[control_908]",
|
902 |
+
"[control_909]",
|
903 |
+
"[control_910]",
|
904 |
+
"[control_911]",
|
905 |
+
"[control_912]",
|
906 |
+
"[control_913]",
|
907 |
+
"[control_914]",
|
908 |
+
"[control_915]",
|
909 |
+
"[control_916]",
|
910 |
+
"[control_917]",
|
911 |
+
"[control_918]",
|
912 |
+
"[control_919]",
|
913 |
+
"[control_920]",
|
914 |
+
"[control_921]",
|
915 |
+
"[control_922]",
|
916 |
+
"[control_923]",
|
917 |
+
"[control_924]",
|
918 |
+
"[control_925]",
|
919 |
+
"[control_926]",
|
920 |
+
"[control_927]",
|
921 |
+
"[control_928]",
|
922 |
+
"[control_929]",
|
923 |
+
"[control_930]",
|
924 |
+
"[control_931]",
|
925 |
+
"[control_932]",
|
926 |
+
"[control_933]",
|
927 |
+
"[control_934]",
|
928 |
+
"[control_935]",
|
929 |
+
"[control_936]",
|
930 |
+
"[control_937]",
|
931 |
+
"[control_938]",
|
932 |
+
"[control_939]",
|
933 |
+
"[control_940]",
|
934 |
+
"[control_941]",
|
935 |
+
"[control_942]",
|
936 |
+
"[control_943]",
|
937 |
+
"[control_944]",
|
938 |
+
"[control_945]",
|
939 |
+
"[control_946]",
|
940 |
+
"[control_947]",
|
941 |
+
"[control_948]",
|
942 |
+
"[control_949]",
|
943 |
+
"[control_950]",
|
944 |
+
"[control_951]",
|
945 |
+
"[control_952]",
|
946 |
+
"[control_953]",
|
947 |
+
"[control_954]",
|
948 |
+
"[control_955]",
|
949 |
+
"[control_956]",
|
950 |
+
"[control_957]",
|
951 |
+
"[control_958]",
|
952 |
+
"[control_959]",
|
953 |
+
"[control_960]",
|
954 |
+
"[control_961]",
|
955 |
+
"[control_962]",
|
956 |
+
"[control_963]",
|
957 |
+
"[control_964]",
|
958 |
+
"[control_965]",
|
959 |
+
"[control_966]",
|
960 |
+
"[control_967]",
|
961 |
+
"[control_968]",
|
962 |
+
"[control_969]",
|
963 |
+
"[control_970]",
|
964 |
+
"[control_971]",
|
965 |
+
"[control_972]",
|
966 |
+
"[control_973]",
|
967 |
+
"[control_974]",
|
968 |
+
"[control_975]",
|
969 |
+
"[control_976]",
|
970 |
+
"[control_977]",
|
971 |
+
"[control_978]",
|
972 |
+
"[control_979]",
|
973 |
+
"[control_980]",
|
974 |
+
"[control_981]",
|
975 |
+
"[control_982]",
|
976 |
+
"[control_983]",
|
977 |
+
"[control_984]",
|
978 |
+
"[control_985]",
|
979 |
+
"[control_986]",
|
980 |
+
"[control_987]",
|
981 |
+
"[control_988]",
|
982 |
+
"[control_989]",
|
983 |
+
"[control_990]",
|
984 |
+
"[control_991]",
|
985 |
+
"[control_992]",
|
986 |
+
"[control_993]",
|
987 |
+
"[control_994]",
|
988 |
+
"[control_995]",
|
989 |
+
"[control_996]",
|
990 |
+
"[control_997]",
|
991 |
+
"[control_998]",
|
992 |
+
"[control_999]",
|
993 |
+
"[control_1000]"
|
994 |
+
],
|
995 |
+
"eos_token": {
|
996 |
+
"content": "<|im_end|>",
|
997 |
+
"lstrip": false,
|
998 |
+
"normalized": false,
|
999 |
+
"rstrip": false,
|
1000 |
+
"single_word": false
|
1001 |
+
},
|
1002 |
+
"pad_token": {
|
1003 |
+
"content": "<|endoftext|>",
|
1004 |
+
"lstrip": false,
|
1005 |
+
"normalized": false,
|
1006 |
+
"rstrip": false,
|
1007 |
+
"single_word": false
|
1008 |
+
}
|
1009 |
+
}
|
tokenizer.json
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:328a87eddd7f2f23d04dd5f861aceadf75f9af4563253644a1b6c11bc09a19ef
|
3 |
+
size 11610855
|
tokenizer_config.json
ADDED
The diff for this file is too large to render.
See raw diff
|
|
vocab.json
ADDED
The diff for this file is too large to render.
See raw diff
|
|