Citaman commited on
Commit
33d544d
·
verified ·
1 Parent(s): e563423

Upload folder using huggingface_hub

Browse files
README.md CHANGED
@@ -29,9 +29,9 @@ The following YAML configuration was used to produce this model:
29
  slices:
30
  - sources:
31
  - model: Citaman/command-r-35-layer
32
- layer_range: [0, 33]
33
  - model: Citaman/command-r-35-layer
34
- layer_range: [1, 34]
35
  merge_method: slerp
36
  base_model: Citaman/command-r-35-layer
37
  parameters:
 
29
  slices:
30
  - sources:
31
  - model: Citaman/command-r-35-layer
32
+ layer_range: [0, 34]
33
  - model: Citaman/command-r-35-layer
34
+ layer_range: [1, 35]
35
  merge_method: slerp
36
  base_model: Citaman/command-r-35-layer
37
  parameters:
config.json CHANGED
@@ -17,7 +17,7 @@
17
  "model_max_length": 131072,
18
  "model_type": "cohere",
19
  "num_attention_heads": 64,
20
- "num_hidden_layers": 33,
21
  "num_key_value_heads": 64,
22
  "pad_token_id": 0,
23
  "pretraining_tp": 1,
 
17
  "model_max_length": 131072,
18
  "model_type": "cohere",
19
  "num_attention_heads": 64,
20
+ "num_hidden_layers": 34,
21
  "num_key_value_heads": 64,
22
  "pad_token_id": 0,
23
  "pretraining_tp": 1,
mergekit_config.yml CHANGED
@@ -1,9 +1,9 @@
1
  slices:
2
  - sources:
3
  - model: Citaman/command-r-35-layer
4
- layer_range: [0, 33]
5
  - model: Citaman/command-r-35-layer
6
- layer_range: [1, 34]
7
  merge_method: slerp
8
  base_model: Citaman/command-r-35-layer
9
  parameters:
 
1
  slices:
2
  - sources:
3
  - model: Citaman/command-r-35-layer
4
+ layer_range: [0, 34]
5
  - model: Citaman/command-r-35-layer
6
+ layer_range: [1, 35]
7
  merge_method: slerp
8
  base_model: Citaman/command-r-35-layer
9
  parameters:
model-00002-of-00007.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:163e75793e98e7b3814f99f4621538d04e83e6ac3c035a1e9ebb6ad043081335
3
- size 9663698896
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:eb0766adc00e62f800e46a48df9ddd669938694b6ba03944eeebae5dd67744fb
3
+ size 9831471408
model-00003-of-00007.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:760d90c0be21495a2e37dc618e5d3fd4a6d58f3cda5112de7fe9f15b40b871b3
3
- size 9932281344
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:cbae3928ae12b49c0d225df583b6fad1339ec166e0fe9853ab430f8c4a742cad
3
+ size 9663812656
model-00004-of-00007.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:fb4f75c39ed55aae32619d905d84d514037d4d3cfbe265c622adf0d0f69f4dad
3
- size 9898661568
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4bdd6226fc2544437d2eddfdc6771f63f279d55ab11c04ce1f24fbadf9bbb744
3
+ size 9697350880
model-00005-of-00007.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:44c12d58f3db9c1cb2b5e566b4f0b1ce97f07067dc3efe77df68a99839a15bc3
3
- size 9965770088
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:017af2683ec2dd41ef922fc7ef91214258b5574d22b36964a309a82f677ff833
3
+ size 9697350896
model-00006-of-00007.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:a0ad6773833855f84f05855f854765e0d7f80d9ad1c09f2ea23541e223efdbeb
3
- size 9865139072
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:88ae891e3468519d9b89c3ed448f9bf6d23f21aa0e54a6e0a3a23153e2c7953c
3
+ size 9697350896
model-00007-of-00007.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:9fc98aa85759f00543d23cc79e789ed68e62efd8c32782ece8c05309fca1628b
3
- size 3321939312
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:831bd1a8b0891b80a15b5b9b7e6df816643fb3e6fc0c0226b861f36cb6116652
3
+ size 5704338032
model.safetensors.index.json CHANGED
@@ -1 +1 @@
1
- {"metadata": {"mergekit_version": "0.0.4.2", "total_size": 62646681600}, "weight_map": {"lm_head.weight": "model-00001-of-00007.safetensors", "model.layers.15.self_attn.o_proj.weight": "model-00001-of-00007.safetensors", "model.layers.15.self_attn.v_proj.weight": "model-00001-of-00007.safetensors", "model.layers.15.self_attn.k_proj.weight": "model-00001-of-00007.safetensors", "model.layers.15.self_attn.q_proj.weight": "model-00001-of-00007.safetensors", "model.layers.9.self_attn.o_proj.weight": "model-00001-of-00007.safetensors", "model.layers.9.self_attn.v_proj.weight": "model-00001-of-00007.safetensors", "model.layers.9.self_attn.k_proj.weight": "model-00001-of-00007.safetensors", "model.layers.9.self_attn.q_proj.weight": "model-00001-of-00007.safetensors", "model.layers.8.self_attn.o_proj.weight": "model-00001-of-00007.safetensors", "model.layers.8.self_attn.v_proj.weight": "model-00001-of-00007.safetensors", "model.layers.8.self_attn.k_proj.weight": "model-00001-of-00007.safetensors", "model.layers.3.self_attn.o_proj.weight": "model-00001-of-00007.safetensors", "model.embed_tokens.weight": "model-00001-of-00007.safetensors", "model.layers.27.self_attn.o_proj.weight": "model-00002-of-00007.safetensors", "model.layers.27.self_attn.v_proj.weight": "model-00002-of-00007.safetensors", "model.layers.27.self_attn.k_proj.weight": "model-00002-of-00007.safetensors", "model.layers.27.self_attn.q_proj.weight": "model-00002-of-00007.safetensors", "model.layers.26.self_attn.o_proj.weight": "model-00002-of-00007.safetensors", "model.layers.26.self_attn.v_proj.weight": "model-00002-of-00007.safetensors", "model.layers.26.self_attn.k_proj.weight": "model-00002-of-00007.safetensors", "model.layers.26.self_attn.q_proj.weight": "model-00002-of-00007.safetensors", "model.layers.26.mlp.up_proj.weight": "model-00002-of-00007.safetensors", "model.layers.21.self_attn.o_proj.weight": "model-00002-of-00007.safetensors", "model.layers.21.self_attn.v_proj.weight": "model-00002-of-00007.safetensors", "model.layers.21.self_attn.k_proj.weight": "model-00002-of-00007.safetensors", "model.layers.21.self_attn.q_proj.weight": "model-00002-of-00007.safetensors", "model.layers.20.self_attn.o_proj.weight": "model-00002-of-00007.safetensors", "model.layers.20.self_attn.v_proj.weight": "model-00002-of-00007.safetensors", "model.layers.20.self_attn.k_proj.weight": "model-00002-of-00007.safetensors", "model.layers.20.self_attn.q_proj.weight": "model-00002-of-00007.safetensors", "model.layers.20.mlp.up_proj.weight": "model-00002-of-00007.safetensors", "model.layers.14.self_attn.o_proj.weight": "model-00002-of-00007.safetensors", "model.layers.14.self_attn.v_proj.weight": "model-00002-of-00007.safetensors", "model.layers.14.self_attn.k_proj.weight": "model-00002-of-00007.safetensors", "model.layers.14.self_attn.q_proj.weight": "model-00002-of-00007.safetensors", "model.layers.14.mlp.up_proj.weight": "model-00002-of-00007.safetensors", "model.layers.14.mlp.gate_proj.weight": "model-00002-of-00007.safetensors", "model.layers.8.self_attn.q_proj.weight": "model-00002-of-00007.safetensors", "model.layers.8.mlp.up_proj.weight": "model-00002-of-00007.safetensors", "model.layers.8.mlp.gate_proj.weight": "model-00002-of-00007.safetensors", "model.layers.3.self_attn.v_proj.weight": "model-00002-of-00007.safetensors", "model.layers.3.self_attn.k_proj.weight": "model-00002-of-00007.safetensors", "model.layers.3.self_attn.q_proj.weight": "model-00002-of-00007.safetensors", "model.layers.2.self_attn.o_proj.weight": "model-00002-of-00007.safetensors", "model.layers.2.self_attn.v_proj.weight": "model-00002-of-00007.safetensors", "model.layers.2.self_attn.k_proj.weight": "model-00002-of-00007.safetensors", "model.layers.2.self_attn.q_proj.weight": "model-00002-of-00007.safetensors", "model.layers.2.mlp.up_proj.weight": "model-00002-of-00007.safetensors", "model.layers.2.mlp.gate_proj.weight": "model-00002-of-00007.safetensors", "model.layers.2.mlp.down_proj.weight": "model-00002-of-00007.safetensors", "model.layers.2.input_layernorm.weight": "model-00002-of-00007.safetensors", "model.layers.1.self_attn.o_proj.weight": "model-00002-of-00007.safetensors", "model.layers.1.self_attn.v_proj.weight": "model-00002-of-00007.safetensors", "model.layers.1.self_attn.k_proj.weight": "model-00002-of-00007.safetensors", "model.layers.0.self_attn.o_proj.weight": "model-00002-of-00007.safetensors", "model.layers.0.self_attn.v_proj.weight": "model-00002-of-00007.safetensors", "model.layers.0.self_attn.k_proj.weight": "model-00002-of-00007.safetensors", "model.layers.0.self_attn.q_proj.weight": "model-00002-of-00007.safetensors", "model.layers.32.self_attn.o_proj.weight": "model-00002-of-00007.safetensors", "model.layers.32.self_attn.v_proj.weight": "model-00002-of-00007.safetensors", "model.layers.32.self_attn.k_proj.weight": "model-00002-of-00007.safetensors", "model.layers.32.self_attn.q_proj.weight": "model-00002-of-00007.safetensors", "model.layers.32.mlp.up_proj.weight": "model-00002-of-00007.safetensors", "model.layers.26.mlp.gate_proj.weight": "model-00002-of-00007.safetensors", "model.layers.20.mlp.gate_proj.weight": "model-00002-of-00007.safetensors", "model.layers.8.mlp.down_proj.weight": "model-00003-of-00007.safetensors", "model.layers.8.input_layernorm.weight": "model-00003-of-00007.safetensors", "model.layers.7.self_attn.o_proj.weight": "model-00003-of-00007.safetensors", "model.layers.7.self_attn.v_proj.weight": "model-00003-of-00007.safetensors", "model.layers.7.self_attn.k_proj.weight": "model-00003-of-00007.safetensors", "model.layers.7.self_attn.q_proj.weight": "model-00003-of-00007.safetensors", "model.layers.6.self_attn.o_proj.weight": "model-00003-of-00007.safetensors", "model.layers.6.self_attn.v_proj.weight": "model-00003-of-00007.safetensors", "model.layers.6.self_attn.k_proj.weight": "model-00003-of-00007.safetensors", "model.layers.6.self_attn.q_proj.weight": "model-00003-of-00007.safetensors", "model.layers.5.self_attn.o_proj.weight": "model-00003-of-00007.safetensors", "model.layers.5.self_attn.v_proj.weight": "model-00003-of-00007.safetensors", "model.layers.5.self_attn.k_proj.weight": "model-00003-of-00007.safetensors", "model.layers.5.self_attn.q_proj.weight": "model-00003-of-00007.safetensors", "model.layers.5.mlp.up_proj.weight": "model-00003-of-00007.safetensors", "model.layers.5.mlp.gate_proj.weight": "model-00003-of-00007.safetensors", "model.layers.5.mlp.down_proj.weight": "model-00003-of-00007.safetensors", "model.layers.5.input_layernorm.weight": "model-00003-of-00007.safetensors", "model.layers.4.self_attn.o_proj.weight": "model-00003-of-00007.safetensors", "model.layers.4.self_attn.v_proj.weight": "model-00003-of-00007.safetensors", "model.layers.4.self_attn.k_proj.weight": "model-00003-of-00007.safetensors", "model.layers.4.self_attn.q_proj.weight": "model-00003-of-00007.safetensors", "model.layers.4.mlp.up_proj.weight": "model-00003-of-00007.safetensors", "model.layers.4.mlp.gate_proj.weight": "model-00003-of-00007.safetensors", "model.layers.4.mlp.down_proj.weight": "model-00003-of-00007.safetensors", "model.layers.4.input_layernorm.weight": "model-00003-of-00007.safetensors", "model.layers.3.mlp.up_proj.weight": "model-00003-of-00007.safetensors", "model.layers.3.mlp.gate_proj.weight": "model-00003-of-00007.safetensors", "model.layers.3.mlp.down_proj.weight": "model-00003-of-00007.safetensors", "model.layers.3.input_layernorm.weight": "model-00003-of-00007.safetensors", "model.layers.1.self_attn.q_proj.weight": "model-00003-of-00007.safetensors", "model.layers.1.mlp.up_proj.weight": "model-00003-of-00007.safetensors", "model.layers.1.mlp.gate_proj.weight": "model-00003-of-00007.safetensors", "model.layers.1.mlp.down_proj.weight": "model-00003-of-00007.safetensors", "model.layers.1.input_layernorm.weight": "model-00003-of-00007.safetensors", "model.layers.0.mlp.up_proj.weight": "model-00003-of-00007.safetensors", "model.layers.0.mlp.gate_proj.weight": "model-00003-of-00007.safetensors", "model.layers.0.mlp.down_proj.weight": "model-00003-of-00007.safetensors", "model.layers.0.input_layernorm.weight": "model-00003-of-00007.safetensors", "model.norm.weight": "model-00003-of-00007.safetensors", "model.layers.32.mlp.gate_proj.weight": "model-00003-of-00007.safetensors", "model.layers.32.mlp.down_proj.weight": "model-00003-of-00007.safetensors", "model.layers.32.input_layernorm.weight": "model-00003-of-00007.safetensors", "model.layers.26.mlp.down_proj.weight": "model-00003-of-00007.safetensors", "model.layers.26.input_layernorm.weight": "model-00003-of-00007.safetensors", "model.layers.25.self_attn.o_proj.weight": "model-00003-of-00007.safetensors", "model.layers.20.mlp.down_proj.weight": "model-00003-of-00007.safetensors", "model.layers.20.input_layernorm.weight": "model-00003-of-00007.safetensors", "model.layers.19.self_attn.o_proj.weight": "model-00003-of-00007.safetensors", "model.layers.19.self_attn.v_proj.weight": "model-00004-of-00007.safetensors", "model.layers.14.mlp.down_proj.weight": "model-00004-of-00007.safetensors", "model.layers.14.input_layernorm.weight": "model-00004-of-00007.safetensors", "model.layers.13.self_attn.o_proj.weight": "model-00004-of-00007.safetensors", "model.layers.13.self_attn.v_proj.weight": "model-00004-of-00007.safetensors", "model.layers.13.self_attn.k_proj.weight": "model-00004-of-00007.safetensors", "model.layers.13.self_attn.q_proj.weight": "model-00004-of-00007.safetensors", "model.layers.12.self_attn.o_proj.weight": "model-00004-of-00007.safetensors", "model.layers.12.self_attn.v_proj.weight": "model-00004-of-00007.safetensors", "model.layers.12.self_attn.k_proj.weight": "model-00004-of-00007.safetensors", "model.layers.12.self_attn.q_proj.weight": "model-00004-of-00007.safetensors", "model.layers.11.self_attn.o_proj.weight": "model-00004-of-00007.safetensors", "model.layers.11.self_attn.v_proj.weight": "model-00004-of-00007.safetensors", "model.layers.11.self_attn.k_proj.weight": "model-00004-of-00007.safetensors", "model.layers.11.self_attn.q_proj.weight": "model-00004-of-00007.safetensors", "model.layers.11.mlp.up_proj.weight": "model-00004-of-00007.safetensors", "model.layers.11.mlp.gate_proj.weight": "model-00004-of-00007.safetensors", "model.layers.11.mlp.down_proj.weight": "model-00004-of-00007.safetensors", "model.layers.11.input_layernorm.weight": "model-00004-of-00007.safetensors", "model.layers.10.self_attn.o_proj.weight": "model-00004-of-00007.safetensors", "model.layers.10.self_attn.v_proj.weight": "model-00004-of-00007.safetensors", "model.layers.10.self_attn.k_proj.weight": "model-00004-of-00007.safetensors", "model.layers.10.self_attn.q_proj.weight": "model-00004-of-00007.safetensors", "model.layers.10.mlp.up_proj.weight": "model-00004-of-00007.safetensors", "model.layers.10.mlp.gate_proj.weight": "model-00004-of-00007.safetensors", "model.layers.10.mlp.down_proj.weight": "model-00004-of-00007.safetensors", "model.layers.10.input_layernorm.weight": "model-00004-of-00007.safetensors", "model.layers.9.mlp.up_proj.weight": "model-00004-of-00007.safetensors", "model.layers.9.mlp.gate_proj.weight": "model-00004-of-00007.safetensors", "model.layers.9.mlp.down_proj.weight": "model-00004-of-00007.safetensors", "model.layers.9.input_layernorm.weight": "model-00004-of-00007.safetensors", "model.layers.7.mlp.up_proj.weight": "model-00004-of-00007.safetensors", "model.layers.7.mlp.gate_proj.weight": "model-00004-of-00007.safetensors", "model.layers.7.mlp.down_proj.weight": "model-00004-of-00007.safetensors", "model.layers.7.input_layernorm.weight": "model-00004-of-00007.safetensors", "model.layers.6.mlp.up_proj.weight": "model-00004-of-00007.safetensors", "model.layers.6.mlp.gate_proj.weight": "model-00004-of-00007.safetensors", "model.layers.6.mlp.down_proj.weight": "model-00004-of-00007.safetensors", "model.layers.6.input_layernorm.weight": "model-00004-of-00007.safetensors", "model.layers.31.self_attn.o_proj.weight": "model-00004-of-00007.safetensors", "model.layers.31.self_attn.v_proj.weight": "model-00004-of-00007.safetensors", "model.layers.31.self_attn.k_proj.weight": "model-00004-of-00007.safetensors", "model.layers.31.self_attn.q_proj.weight": "model-00004-of-00007.safetensors", "model.layers.31.mlp.up_proj.weight": "model-00004-of-00007.safetensors", "model.layers.25.self_attn.v_proj.weight": "model-00004-of-00007.safetensors", "model.layers.25.self_attn.k_proj.weight": "model-00004-of-00007.safetensors", "model.layers.25.self_attn.q_proj.weight": "model-00004-of-00007.safetensors", "model.layers.24.self_attn.o_proj.weight": "model-00004-of-00007.safetensors", "model.layers.24.self_attn.v_proj.weight": "model-00004-of-00007.safetensors", "model.layers.24.self_attn.k_proj.weight": "model-00004-of-00007.safetensors", "model.layers.24.self_attn.q_proj.weight": "model-00005-of-00007.safetensors", "model.layers.23.self_attn.o_proj.weight": "model-00005-of-00007.safetensors", "model.layers.23.self_attn.v_proj.weight": "model-00005-of-00007.safetensors", "model.layers.23.self_attn.k_proj.weight": "model-00005-of-00007.safetensors", "model.layers.19.self_attn.k_proj.weight": "model-00005-of-00007.safetensors", "model.layers.19.self_attn.q_proj.weight": "model-00005-of-00007.safetensors", "model.layers.18.self_attn.o_proj.weight": "model-00005-of-00007.safetensors", "model.layers.18.self_attn.v_proj.weight": "model-00005-of-00007.safetensors", "model.layers.18.self_attn.k_proj.weight": "model-00005-of-00007.safetensors", "model.layers.18.self_attn.q_proj.weight": "model-00005-of-00007.safetensors", "model.layers.17.self_attn.o_proj.weight": "model-00005-of-00007.safetensors", "model.layers.17.self_attn.v_proj.weight": "model-00005-of-00007.safetensors", "model.layers.17.self_attn.k_proj.weight": "model-00005-of-00007.safetensors", "model.layers.17.self_attn.q_proj.weight": "model-00005-of-00007.safetensors", "model.layers.17.mlp.up_proj.weight": "model-00005-of-00007.safetensors", "model.layers.17.mlp.gate_proj.weight": "model-00005-of-00007.safetensors", "model.layers.17.mlp.down_proj.weight": "model-00005-of-00007.safetensors", "model.layers.17.input_layernorm.weight": "model-00005-of-00007.safetensors", "model.layers.16.self_attn.o_proj.weight": "model-00005-of-00007.safetensors", "model.layers.16.self_attn.v_proj.weight": "model-00005-of-00007.safetensors", "model.layers.16.self_attn.k_proj.weight": "model-00005-of-00007.safetensors", "model.layers.16.self_attn.q_proj.weight": "model-00005-of-00007.safetensors", "model.layers.16.mlp.up_proj.weight": "model-00005-of-00007.safetensors", "model.layers.16.mlp.gate_proj.weight": "model-00005-of-00007.safetensors", "model.layers.16.mlp.down_proj.weight": "model-00005-of-00007.safetensors", "model.layers.16.input_layernorm.weight": "model-00005-of-00007.safetensors", "model.layers.15.mlp.up_proj.weight": "model-00005-of-00007.safetensors", "model.layers.15.mlp.gate_proj.weight": "model-00005-of-00007.safetensors", "model.layers.15.mlp.down_proj.weight": "model-00005-of-00007.safetensors", "model.layers.15.input_layernorm.weight": "model-00005-of-00007.safetensors", "model.layers.13.mlp.up_proj.weight": "model-00005-of-00007.safetensors", "model.layers.13.mlp.gate_proj.weight": "model-00005-of-00007.safetensors", "model.layers.13.mlp.down_proj.weight": "model-00005-of-00007.safetensors", "model.layers.13.input_layernorm.weight": "model-00005-of-00007.safetensors", "model.layers.12.mlp.up_proj.weight": "model-00005-of-00007.safetensors", "model.layers.12.mlp.gate_proj.weight": "model-00005-of-00007.safetensors", "model.layers.12.mlp.down_proj.weight": "model-00005-of-00007.safetensors", "model.layers.12.input_layernorm.weight": "model-00005-of-00007.safetensors", "model.layers.31.mlp.gate_proj.weight": "model-00005-of-00007.safetensors", "model.layers.31.mlp.down_proj.weight": "model-00005-of-00007.safetensors", "model.layers.31.input_layernorm.weight": "model-00005-of-00007.safetensors", "model.layers.30.self_attn.o_proj.weight": "model-00005-of-00007.safetensors", "model.layers.30.self_attn.v_proj.weight": "model-00005-of-00007.safetensors", "model.layers.30.self_attn.k_proj.weight": "model-00005-of-00007.safetensors", "model.layers.30.self_attn.q_proj.weight": "model-00005-of-00007.safetensors", "model.layers.30.mlp.up_proj.weight": "model-00005-of-00007.safetensors", "model.layers.30.mlp.gate_proj.weight": "model-00005-of-00007.safetensors", "model.layers.30.mlp.down_proj.weight": "model-00006-of-00007.safetensors", "model.layers.30.input_layernorm.weight": "model-00006-of-00007.safetensors", "model.layers.29.self_attn.o_proj.weight": "model-00006-of-00007.safetensors", "model.layers.29.self_attn.v_proj.weight": "model-00006-of-00007.safetensors", "model.layers.29.self_attn.k_proj.weight": "model-00006-of-00007.safetensors", "model.layers.29.self_attn.q_proj.weight": "model-00006-of-00007.safetensors", "model.layers.29.mlp.up_proj.weight": "model-00006-of-00007.safetensors", "model.layers.29.mlp.gate_proj.weight": "model-00006-of-00007.safetensors", "model.layers.23.self_attn.q_proj.weight": "model-00006-of-00007.safetensors", "model.layers.23.mlp.up_proj.weight": "model-00006-of-00007.safetensors", "model.layers.23.mlp.gate_proj.weight": "model-00006-of-00007.safetensors", "model.layers.23.mlp.down_proj.weight": "model-00006-of-00007.safetensors", "model.layers.23.input_layernorm.weight": "model-00006-of-00007.safetensors", "model.layers.22.self_attn.o_proj.weight": "model-00006-of-00007.safetensors", "model.layers.22.self_attn.v_proj.weight": "model-00006-of-00007.safetensors", "model.layers.22.self_attn.k_proj.weight": "model-00006-of-00007.safetensors", "model.layers.22.self_attn.q_proj.weight": "model-00006-of-00007.safetensors", "model.layers.22.mlp.up_proj.weight": "model-00006-of-00007.safetensors", "model.layers.22.mlp.gate_proj.weight": "model-00006-of-00007.safetensors", "model.layers.22.mlp.down_proj.weight": "model-00006-of-00007.safetensors", "model.layers.22.input_layernorm.weight": "model-00006-of-00007.safetensors", "model.layers.21.mlp.up_proj.weight": "model-00006-of-00007.safetensors", "model.layers.21.mlp.gate_proj.weight": "model-00006-of-00007.safetensors", "model.layers.21.mlp.down_proj.weight": "model-00006-of-00007.safetensors", "model.layers.21.input_layernorm.weight": "model-00006-of-00007.safetensors", "model.layers.19.mlp.up_proj.weight": "model-00006-of-00007.safetensors", "model.layers.19.mlp.gate_proj.weight": "model-00006-of-00007.safetensors", "model.layers.19.mlp.down_proj.weight": "model-00006-of-00007.safetensors", "model.layers.19.input_layernorm.weight": "model-00006-of-00007.safetensors", "model.layers.18.mlp.up_proj.weight": "model-00006-of-00007.safetensors", "model.layers.18.mlp.gate_proj.weight": "model-00006-of-00007.safetensors", "model.layers.18.mlp.down_proj.weight": "model-00006-of-00007.safetensors", "model.layers.18.input_layernorm.weight": "model-00006-of-00007.safetensors", "model.layers.29.mlp.down_proj.weight": "model-00006-of-00007.safetensors", "model.layers.29.input_layernorm.weight": "model-00006-of-00007.safetensors", "model.layers.28.self_attn.o_proj.weight": "model-00006-of-00007.safetensors", "model.layers.28.self_attn.v_proj.weight": "model-00006-of-00007.safetensors", "model.layers.28.self_attn.k_proj.weight": "model-00006-of-00007.safetensors", "model.layers.28.self_attn.q_proj.weight": "model-00006-of-00007.safetensors", "model.layers.28.mlp.up_proj.weight": "model-00006-of-00007.safetensors", "model.layers.28.mlp.gate_proj.weight": "model-00006-of-00007.safetensors", "model.layers.28.mlp.down_proj.weight": "model-00006-of-00007.safetensors", "model.layers.28.input_layernorm.weight": "model-00006-of-00007.safetensors", "model.layers.27.mlp.up_proj.weight": "model-00007-of-00007.safetensors", "model.layers.27.mlp.gate_proj.weight": "model-00007-of-00007.safetensors", "model.layers.27.mlp.down_proj.weight": "model-00007-of-00007.safetensors", "model.layers.27.input_layernorm.weight": "model-00007-of-00007.safetensors", "model.layers.25.mlp.up_proj.weight": "model-00007-of-00007.safetensors", "model.layers.25.mlp.gate_proj.weight": "model-00007-of-00007.safetensors", "model.layers.25.mlp.down_proj.weight": "model-00007-of-00007.safetensors", "model.layers.25.input_layernorm.weight": "model-00007-of-00007.safetensors", "model.layers.24.mlp.up_proj.weight": "model-00007-of-00007.safetensors", "model.layers.24.mlp.gate_proj.weight": "model-00007-of-00007.safetensors", "model.layers.24.mlp.down_proj.weight": "model-00007-of-00007.safetensors", "model.layers.24.input_layernorm.weight": "model-00007-of-00007.safetensors"}}
 
1
+ {"metadata": {"mergekit_version": "0.0.4.2", "total_size": 64290865152}, "weight_map": {"lm_head.weight": "model-00001-of-00007.safetensors", "model.layers.15.self_attn.o_proj.weight": "model-00001-of-00007.safetensors", "model.layers.15.self_attn.v_proj.weight": "model-00001-of-00007.safetensors", "model.layers.15.self_attn.k_proj.weight": "model-00001-of-00007.safetensors", "model.layers.15.self_attn.q_proj.weight": "model-00001-of-00007.safetensors", "model.layers.9.self_attn.o_proj.weight": "model-00001-of-00007.safetensors", "model.layers.9.self_attn.v_proj.weight": "model-00001-of-00007.safetensors", "model.layers.9.self_attn.k_proj.weight": "model-00001-of-00007.safetensors", "model.layers.9.self_attn.q_proj.weight": "model-00001-of-00007.safetensors", "model.layers.8.self_attn.o_proj.weight": "model-00001-of-00007.safetensors", "model.layers.8.self_attn.v_proj.weight": "model-00001-of-00007.safetensors", "model.layers.8.self_attn.k_proj.weight": "model-00001-of-00007.safetensors", "model.layers.3.self_attn.o_proj.weight": "model-00001-of-00007.safetensors", "model.embed_tokens.weight": "model-00001-of-00007.safetensors", "model.layers.27.self_attn.o_proj.weight": "model-00002-of-00007.safetensors", "model.layers.27.self_attn.v_proj.weight": "model-00002-of-00007.safetensors", "model.layers.27.self_attn.k_proj.weight": "model-00002-of-00007.safetensors", "model.layers.27.self_attn.q_proj.weight": "model-00002-of-00007.safetensors", "model.layers.26.self_attn.o_proj.weight": "model-00002-of-00007.safetensors", "model.layers.26.self_attn.v_proj.weight": "model-00002-of-00007.safetensors", "model.layers.26.self_attn.k_proj.weight": "model-00002-of-00007.safetensors", "model.layers.26.self_attn.q_proj.weight": "model-00002-of-00007.safetensors", "model.layers.26.mlp.up_proj.weight": "model-00002-of-00007.safetensors", "model.layers.21.self_attn.o_proj.weight": "model-00002-of-00007.safetensors", "model.layers.21.self_attn.v_proj.weight": "model-00002-of-00007.safetensors", "model.layers.21.self_attn.k_proj.weight": "model-00002-of-00007.safetensors", "model.layers.21.self_attn.q_proj.weight": "model-00002-of-00007.safetensors", "model.layers.20.self_attn.o_proj.weight": "model-00002-of-00007.safetensors", "model.layers.20.self_attn.v_proj.weight": "model-00002-of-00007.safetensors", "model.layers.20.self_attn.k_proj.weight": "model-00002-of-00007.safetensors", "model.layers.20.self_attn.q_proj.weight": "model-00002-of-00007.safetensors", "model.layers.20.mlp.up_proj.weight": "model-00002-of-00007.safetensors", "model.layers.14.self_attn.o_proj.weight": "model-00002-of-00007.safetensors", "model.layers.14.self_attn.v_proj.weight": "model-00002-of-00007.safetensors", "model.layers.14.self_attn.k_proj.weight": "model-00002-of-00007.safetensors", "model.layers.14.self_attn.q_proj.weight": "model-00002-of-00007.safetensors", "model.layers.14.mlp.up_proj.weight": "model-00002-of-00007.safetensors", "model.layers.14.mlp.gate_proj.weight": "model-00002-of-00007.safetensors", "model.layers.8.self_attn.q_proj.weight": "model-00002-of-00007.safetensors", "model.layers.8.mlp.up_proj.weight": "model-00002-of-00007.safetensors", "model.layers.8.mlp.gate_proj.weight": "model-00002-of-00007.safetensors", "model.layers.3.self_attn.v_proj.weight": "model-00002-of-00007.safetensors", "model.layers.3.self_attn.k_proj.weight": "model-00002-of-00007.safetensors", "model.layers.3.self_attn.q_proj.weight": "model-00002-of-00007.safetensors", "model.layers.2.self_attn.o_proj.weight": "model-00002-of-00007.safetensors", "model.layers.2.self_attn.v_proj.weight": "model-00002-of-00007.safetensors", "model.layers.2.self_attn.k_proj.weight": "model-00002-of-00007.safetensors", "model.layers.2.self_attn.q_proj.weight": "model-00002-of-00007.safetensors", "model.layers.2.mlp.up_proj.weight": "model-00002-of-00007.safetensors", "model.layers.2.mlp.gate_proj.weight": "model-00002-of-00007.safetensors", "model.layers.2.mlp.down_proj.weight": "model-00002-of-00007.safetensors", "model.layers.2.input_layernorm.weight": "model-00002-of-00007.safetensors", "model.layers.1.self_attn.o_proj.weight": "model-00002-of-00007.safetensors", "model.layers.1.self_attn.v_proj.weight": "model-00002-of-00007.safetensors", "model.layers.1.self_attn.k_proj.weight": "model-00002-of-00007.safetensors", "model.layers.0.self_attn.o_proj.weight": "model-00002-of-00007.safetensors", "model.layers.0.self_attn.v_proj.weight": "model-00002-of-00007.safetensors", "model.layers.0.self_attn.k_proj.weight": "model-00002-of-00007.safetensors", "model.layers.0.self_attn.q_proj.weight": "model-00002-of-00007.safetensors", "model.layers.33.self_attn.o_proj.weight": "model-00002-of-00007.safetensors", "model.layers.33.self_attn.v_proj.weight": "model-00002-of-00007.safetensors", "model.layers.33.self_attn.k_proj.weight": "model-00002-of-00007.safetensors", "model.layers.33.self_attn.q_proj.weight": "model-00002-of-00007.safetensors", "model.layers.32.self_attn.o_proj.weight": "model-00002-of-00007.safetensors", "model.layers.32.self_attn.v_proj.weight": "model-00002-of-00007.safetensors", "model.layers.32.self_attn.k_proj.weight": "model-00002-of-00007.safetensors", "model.layers.32.self_attn.q_proj.weight": "model-00002-of-00007.safetensors", "model.layers.32.mlp.up_proj.weight": "model-00002-of-00007.safetensors", "model.layers.20.mlp.gate_proj.weight": "model-00002-of-00007.safetensors", "model.layers.8.mlp.down_proj.weight": "model-00003-of-00007.safetensors", "model.layers.8.input_layernorm.weight": "model-00003-of-00007.safetensors", "model.layers.7.self_attn.o_proj.weight": "model-00003-of-00007.safetensors", "model.layers.7.self_attn.v_proj.weight": "model-00003-of-00007.safetensors", "model.layers.7.self_attn.k_proj.weight": "model-00003-of-00007.safetensors", "model.layers.7.self_attn.q_proj.weight": "model-00003-of-00007.safetensors", "model.layers.6.self_attn.o_proj.weight": "model-00003-of-00007.safetensors", "model.layers.6.self_attn.v_proj.weight": "model-00003-of-00007.safetensors", "model.layers.6.self_attn.k_proj.weight": "model-00003-of-00007.safetensors", "model.layers.6.self_attn.q_proj.weight": "model-00003-of-00007.safetensors", "model.layers.5.self_attn.o_proj.weight": "model-00003-of-00007.safetensors", "model.layers.5.self_attn.v_proj.weight": "model-00003-of-00007.safetensors", "model.layers.5.self_attn.k_proj.weight": "model-00003-of-00007.safetensors", "model.layers.5.self_attn.q_proj.weight": "model-00003-of-00007.safetensors", "model.layers.5.mlp.up_proj.weight": "model-00003-of-00007.safetensors", "model.layers.5.mlp.gate_proj.weight": "model-00003-of-00007.safetensors", "model.layers.5.mlp.down_proj.weight": "model-00003-of-00007.safetensors", "model.layers.5.input_layernorm.weight": "model-00003-of-00007.safetensors", "model.layers.4.self_attn.o_proj.weight": "model-00003-of-00007.safetensors", "model.layers.4.self_attn.v_proj.weight": "model-00003-of-00007.safetensors", "model.layers.4.self_attn.k_proj.weight": "model-00003-of-00007.safetensors", "model.layers.4.self_attn.q_proj.weight": "model-00003-of-00007.safetensors", "model.layers.4.mlp.up_proj.weight": "model-00003-of-00007.safetensors", "model.layers.4.mlp.gate_proj.weight": "model-00003-of-00007.safetensors", "model.layers.4.mlp.down_proj.weight": "model-00003-of-00007.safetensors", "model.layers.4.input_layernorm.weight": "model-00003-of-00007.safetensors", "model.layers.3.mlp.up_proj.weight": "model-00003-of-00007.safetensors", "model.layers.3.mlp.gate_proj.weight": "model-00003-of-00007.safetensors", "model.layers.3.mlp.down_proj.weight": "model-00003-of-00007.safetensors", "model.layers.3.input_layernorm.weight": "model-00003-of-00007.safetensors", "model.layers.1.self_attn.q_proj.weight": "model-00003-of-00007.safetensors", "model.layers.1.mlp.up_proj.weight": "model-00003-of-00007.safetensors", "model.layers.1.mlp.gate_proj.weight": "model-00003-of-00007.safetensors", "model.layers.1.mlp.down_proj.weight": "model-00003-of-00007.safetensors", "model.layers.1.input_layernorm.weight": "model-00003-of-00007.safetensors", "model.layers.0.mlp.up_proj.weight": "model-00003-of-00007.safetensors", "model.layers.0.mlp.gate_proj.weight": "model-00003-of-00007.safetensors", "model.layers.0.mlp.down_proj.weight": "model-00003-of-00007.safetensors", "model.layers.0.input_layernorm.weight": "model-00003-of-00007.safetensors", "model.norm.weight": "model-00003-of-00007.safetensors", "model.layers.33.mlp.up_proj.weight": "model-00003-of-00007.safetensors", "model.layers.33.mlp.gate_proj.weight": "model-00003-of-00007.safetensors", "model.layers.33.mlp.down_proj.weight": "model-00003-of-00007.safetensors", "model.layers.33.input_layernorm.weight": "model-00003-of-00007.safetensors", "model.layers.32.mlp.gate_proj.weight": "model-00003-of-00007.safetensors", "model.layers.26.mlp.gate_proj.weight": "model-00004-of-00007.safetensors", "model.layers.14.mlp.down_proj.weight": "model-00004-of-00007.safetensors", "model.layers.14.input_layernorm.weight": "model-00004-of-00007.safetensors", "model.layers.13.self_attn.o_proj.weight": "model-00004-of-00007.safetensors", "model.layers.13.self_attn.v_proj.weight": "model-00004-of-00007.safetensors", "model.layers.13.self_attn.k_proj.weight": "model-00004-of-00007.safetensors", "model.layers.13.self_attn.q_proj.weight": "model-00004-of-00007.safetensors", "model.layers.12.self_attn.o_proj.weight": "model-00004-of-00007.safetensors", "model.layers.12.self_attn.v_proj.weight": "model-00004-of-00007.safetensors", "model.layers.12.self_attn.k_proj.weight": "model-00004-of-00007.safetensors", "model.layers.12.self_attn.q_proj.weight": "model-00004-of-00007.safetensors", "model.layers.11.self_attn.o_proj.weight": "model-00004-of-00007.safetensors", "model.layers.11.self_attn.v_proj.weight": "model-00004-of-00007.safetensors", "model.layers.11.self_attn.k_proj.weight": "model-00004-of-00007.safetensors", "model.layers.11.self_attn.q_proj.weight": "model-00004-of-00007.safetensors", "model.layers.11.mlp.up_proj.weight": "model-00004-of-00007.safetensors", "model.layers.11.mlp.gate_proj.weight": "model-00004-of-00007.safetensors", "model.layers.11.mlp.down_proj.weight": "model-00004-of-00007.safetensors", "model.layers.11.input_layernorm.weight": "model-00004-of-00007.safetensors", "model.layers.10.self_attn.o_proj.weight": "model-00004-of-00007.safetensors", "model.layers.10.self_attn.v_proj.weight": "model-00004-of-00007.safetensors", "model.layers.10.self_attn.k_proj.weight": "model-00004-of-00007.safetensors", "model.layers.10.self_attn.q_proj.weight": "model-00004-of-00007.safetensors", "model.layers.10.mlp.up_proj.weight": "model-00004-of-00007.safetensors", "model.layers.10.mlp.gate_proj.weight": "model-00004-of-00007.safetensors", "model.layers.10.mlp.down_proj.weight": "model-00004-of-00007.safetensors", "model.layers.10.input_layernorm.weight": "model-00004-of-00007.safetensors", "model.layers.9.mlp.up_proj.weight": "model-00004-of-00007.safetensors", "model.layers.9.mlp.gate_proj.weight": "model-00004-of-00007.safetensors", "model.layers.9.mlp.down_proj.weight": "model-00004-of-00007.safetensors", "model.layers.9.input_layernorm.weight": "model-00004-of-00007.safetensors", "model.layers.7.mlp.up_proj.weight": "model-00004-of-00007.safetensors", "model.layers.7.mlp.gate_proj.weight": "model-00004-of-00007.safetensors", "model.layers.7.mlp.down_proj.weight": "model-00004-of-00007.safetensors", "model.layers.7.input_layernorm.weight": "model-00004-of-00007.safetensors", "model.layers.6.mlp.up_proj.weight": "model-00004-of-00007.safetensors", "model.layers.6.mlp.gate_proj.weight": "model-00004-of-00007.safetensors", "model.layers.6.mlp.down_proj.weight": "model-00004-of-00007.safetensors", "model.layers.6.input_layernorm.weight": "model-00004-of-00007.safetensors", "model.layers.32.mlp.down_proj.weight": "model-00004-of-00007.safetensors", "model.layers.32.input_layernorm.weight": "model-00004-of-00007.safetensors", "model.layers.31.self_attn.o_proj.weight": "model-00004-of-00007.safetensors", "model.layers.31.self_attn.v_proj.weight": "model-00004-of-00007.safetensors", "model.layers.31.self_attn.k_proj.weight": "model-00004-of-00007.safetensors", "model.layers.31.self_attn.q_proj.weight": "model-00004-of-00007.safetensors", "model.layers.31.mlp.up_proj.weight": "model-00004-of-00007.safetensors", "model.layers.31.mlp.gate_proj.weight": "model-00005-of-00007.safetensors", "model.layers.20.mlp.down_proj.weight": "model-00005-of-00007.safetensors", "model.layers.20.input_layernorm.weight": "model-00005-of-00007.safetensors", "model.layers.19.self_attn.o_proj.weight": "model-00005-of-00007.safetensors", "model.layers.19.self_attn.v_proj.weight": "model-00005-of-00007.safetensors", "model.layers.19.self_attn.k_proj.weight": "model-00005-of-00007.safetensors", "model.layers.19.self_attn.q_proj.weight": "model-00005-of-00007.safetensors", "model.layers.18.self_attn.o_proj.weight": "model-00005-of-00007.safetensors", "model.layers.18.self_attn.v_proj.weight": "model-00005-of-00007.safetensors", "model.layers.18.self_attn.k_proj.weight": "model-00005-of-00007.safetensors", "model.layers.18.self_attn.q_proj.weight": "model-00005-of-00007.safetensors", "model.layers.17.self_attn.o_proj.weight": "model-00005-of-00007.safetensors", "model.layers.17.self_attn.v_proj.weight": "model-00005-of-00007.safetensors", "model.layers.17.self_attn.k_proj.weight": "model-00005-of-00007.safetensors", "model.layers.17.self_attn.q_proj.weight": "model-00005-of-00007.safetensors", "model.layers.17.mlp.up_proj.weight": "model-00005-of-00007.safetensors", "model.layers.17.mlp.gate_proj.weight": "model-00005-of-00007.safetensors", "model.layers.17.mlp.down_proj.weight": "model-00005-of-00007.safetensors", "model.layers.17.input_layernorm.weight": "model-00005-of-00007.safetensors", "model.layers.16.self_attn.o_proj.weight": "model-00005-of-00007.safetensors", "model.layers.16.self_attn.v_proj.weight": "model-00005-of-00007.safetensors", "model.layers.16.self_attn.k_proj.weight": "model-00005-of-00007.safetensors", "model.layers.16.self_attn.q_proj.weight": "model-00005-of-00007.safetensors", "model.layers.16.mlp.up_proj.weight": "model-00005-of-00007.safetensors", "model.layers.16.mlp.gate_proj.weight": "model-00005-of-00007.safetensors", "model.layers.16.mlp.down_proj.weight": "model-00005-of-00007.safetensors", "model.layers.16.input_layernorm.weight": "model-00005-of-00007.safetensors", "model.layers.15.mlp.up_proj.weight": "model-00005-of-00007.safetensors", "model.layers.15.mlp.gate_proj.weight": "model-00005-of-00007.safetensors", "model.layers.15.mlp.down_proj.weight": "model-00005-of-00007.safetensors", "model.layers.15.input_layernorm.weight": "model-00005-of-00007.safetensors", "model.layers.13.mlp.up_proj.weight": "model-00005-of-00007.safetensors", "model.layers.13.mlp.gate_proj.weight": "model-00005-of-00007.safetensors", "model.layers.13.mlp.down_proj.weight": "model-00005-of-00007.safetensors", "model.layers.13.input_layernorm.weight": "model-00005-of-00007.safetensors", "model.layers.12.mlp.up_proj.weight": "model-00005-of-00007.safetensors", "model.layers.12.mlp.gate_proj.weight": "model-00005-of-00007.safetensors", "model.layers.12.mlp.down_proj.weight": "model-00005-of-00007.safetensors", "model.layers.12.input_layernorm.weight": "model-00005-of-00007.safetensors", "model.layers.31.mlp.down_proj.weight": "model-00005-of-00007.safetensors", "model.layers.31.input_layernorm.weight": "model-00005-of-00007.safetensors", "model.layers.30.self_attn.o_proj.weight": "model-00005-of-00007.safetensors", "model.layers.30.self_attn.v_proj.weight": "model-00005-of-00007.safetensors", "model.layers.30.self_attn.k_proj.weight": "model-00005-of-00007.safetensors", "model.layers.30.self_attn.q_proj.weight": "model-00005-of-00007.safetensors", "model.layers.30.mlp.up_proj.weight": "model-00005-of-00007.safetensors", "model.layers.30.mlp.gate_proj.weight": "model-00006-of-00007.safetensors", "model.layers.26.mlp.down_proj.weight": "model-00006-of-00007.safetensors", "model.layers.26.input_layernorm.weight": "model-00006-of-00007.safetensors", "model.layers.25.self_attn.o_proj.weight": "model-00006-of-00007.safetensors", "model.layers.25.self_attn.v_proj.weight": "model-00006-of-00007.safetensors", "model.layers.25.self_attn.k_proj.weight": "model-00006-of-00007.safetensors", "model.layers.25.self_attn.q_proj.weight": "model-00006-of-00007.safetensors", "model.layers.24.self_attn.o_proj.weight": "model-00006-of-00007.safetensors", "model.layers.24.self_attn.v_proj.weight": "model-00006-of-00007.safetensors", "model.layers.24.self_attn.k_proj.weight": "model-00006-of-00007.safetensors", "model.layers.24.self_attn.q_proj.weight": "model-00006-of-00007.safetensors", "model.layers.23.self_attn.o_proj.weight": "model-00006-of-00007.safetensors", "model.layers.23.self_attn.v_proj.weight": "model-00006-of-00007.safetensors", "model.layers.23.self_attn.k_proj.weight": "model-00006-of-00007.safetensors", "model.layers.23.self_attn.q_proj.weight": "model-00006-of-00007.safetensors", "model.layers.23.mlp.up_proj.weight": "model-00006-of-00007.safetensors", "model.layers.23.mlp.gate_proj.weight": "model-00006-of-00007.safetensors", "model.layers.23.mlp.down_proj.weight": "model-00006-of-00007.safetensors", "model.layers.23.input_layernorm.weight": "model-00006-of-00007.safetensors", "model.layers.22.self_attn.o_proj.weight": "model-00006-of-00007.safetensors", "model.layers.22.self_attn.v_proj.weight": "model-00006-of-00007.safetensors", "model.layers.22.self_attn.k_proj.weight": "model-00006-of-00007.safetensors", "model.layers.22.self_attn.q_proj.weight": "model-00006-of-00007.safetensors", "model.layers.22.mlp.up_proj.weight": "model-00006-of-00007.safetensors", "model.layers.22.mlp.gate_proj.weight": "model-00006-of-00007.safetensors", "model.layers.22.mlp.down_proj.weight": "model-00006-of-00007.safetensors", "model.layers.22.input_layernorm.weight": "model-00006-of-00007.safetensors", "model.layers.21.mlp.up_proj.weight": "model-00006-of-00007.safetensors", "model.layers.21.mlp.gate_proj.weight": "model-00006-of-00007.safetensors", "model.layers.21.mlp.down_proj.weight": "model-00006-of-00007.safetensors", "model.layers.21.input_layernorm.weight": "model-00006-of-00007.safetensors", "model.layers.19.mlp.up_proj.weight": "model-00006-of-00007.safetensors", "model.layers.19.mlp.gate_proj.weight": "model-00006-of-00007.safetensors", "model.layers.19.mlp.down_proj.weight": "model-00006-of-00007.safetensors", "model.layers.19.input_layernorm.weight": "model-00006-of-00007.safetensors", "model.layers.18.mlp.up_proj.weight": "model-00006-of-00007.safetensors", "model.layers.18.mlp.gate_proj.weight": "model-00006-of-00007.safetensors", "model.layers.18.mlp.down_proj.weight": "model-00006-of-00007.safetensors", "model.layers.18.input_layernorm.weight": "model-00006-of-00007.safetensors", "model.layers.30.mlp.down_proj.weight": "model-00006-of-00007.safetensors", "model.layers.30.input_layernorm.weight": "model-00006-of-00007.safetensors", "model.layers.29.self_attn.o_proj.weight": "model-00006-of-00007.safetensors", "model.layers.29.self_attn.v_proj.weight": "model-00006-of-00007.safetensors", "model.layers.29.self_attn.k_proj.weight": "model-00006-of-00007.safetensors", "model.layers.29.self_attn.q_proj.weight": "model-00006-of-00007.safetensors", "model.layers.29.mlp.up_proj.weight": "model-00006-of-00007.safetensors", "model.layers.29.mlp.gate_proj.weight": "model-00007-of-00007.safetensors", "model.layers.29.mlp.down_proj.weight": "model-00007-of-00007.safetensors", "model.layers.29.input_layernorm.weight": "model-00007-of-00007.safetensors", "model.layers.28.self_attn.o_proj.weight": "model-00007-of-00007.safetensors", "model.layers.28.self_attn.v_proj.weight": "model-00007-of-00007.safetensors", "model.layers.28.self_attn.k_proj.weight": "model-00007-of-00007.safetensors", "model.layers.28.self_attn.q_proj.weight": "model-00007-of-00007.safetensors", "model.layers.28.mlp.up_proj.weight": "model-00007-of-00007.safetensors", "model.layers.28.mlp.gate_proj.weight": "model-00007-of-00007.safetensors", "model.layers.28.mlp.down_proj.weight": "model-00007-of-00007.safetensors", "model.layers.28.input_layernorm.weight": "model-00007-of-00007.safetensors", "model.layers.27.mlp.up_proj.weight": "model-00007-of-00007.safetensors", "model.layers.27.mlp.gate_proj.weight": "model-00007-of-00007.safetensors", "model.layers.27.mlp.down_proj.weight": "model-00007-of-00007.safetensors", "model.layers.27.input_layernorm.weight": "model-00007-of-00007.safetensors", "model.layers.25.mlp.up_proj.weight": "model-00007-of-00007.safetensors", "model.layers.25.mlp.gate_proj.weight": "model-00007-of-00007.safetensors", "model.layers.25.mlp.down_proj.weight": "model-00007-of-00007.safetensors", "model.layers.25.input_layernorm.weight": "model-00007-of-00007.safetensors", "model.layers.24.mlp.up_proj.weight": "model-00007-of-00007.safetensors", "model.layers.24.mlp.gate_proj.weight": "model-00007-of-00007.safetensors", "model.layers.24.mlp.down_proj.weight": "model-00007-of-00007.safetensors", "model.layers.24.input_layernorm.weight": "model-00007-of-00007.safetensors"}}