Custom-5 / model.safetensors.index.json
Sakalti's picture
Upload model
fa8e1f6 verified
{
"metadata": {
"total_size": 6484624384
},
"weight_map": {
"embedding.weight": "model-00001-of-00002.safetensors",
"layer_norm.bias": "model-00002-of-00002.safetensors",
"layer_norm.weight": "model-00002-of-00002.safetensors",
"layers.0.attention.k_proj.bias": "model-00001-of-00002.safetensors",
"layers.0.attention.k_proj.weight": "model-00001-of-00002.safetensors",
"layers.0.attention.out_proj.bias": "model-00001-of-00002.safetensors",
"layers.0.attention.out_proj.weight": "model-00001-of-00002.safetensors",
"layers.0.attention.q_proj.bias": "model-00001-of-00002.safetensors",
"layers.0.attention.q_proj.weight": "model-00001-of-00002.safetensors",
"layers.0.attention.v_proj.bias": "model-00001-of-00002.safetensors",
"layers.0.attention.v_proj.weight": "model-00001-of-00002.safetensors",
"layers.0.layer_norm1.bias": "model-00001-of-00002.safetensors",
"layers.0.layer_norm1.weight": "model-00001-of-00002.safetensors",
"layers.0.layer_norm2.bias": "model-00001-of-00002.safetensors",
"layers.0.layer_norm2.weight": "model-00001-of-00002.safetensors",
"layers.0.mlp.0.bias": "model-00001-of-00002.safetensors",
"layers.0.mlp.0.weight": "model-00001-of-00002.safetensors",
"layers.0.mlp.2.bias": "model-00001-of-00002.safetensors",
"layers.0.mlp.2.weight": "model-00001-of-00002.safetensors",
"layers.1.attention.k_proj.bias": "model-00001-of-00002.safetensors",
"layers.1.attention.k_proj.weight": "model-00001-of-00002.safetensors",
"layers.1.attention.out_proj.bias": "model-00001-of-00002.safetensors",
"layers.1.attention.out_proj.weight": "model-00001-of-00002.safetensors",
"layers.1.attention.q_proj.bias": "model-00001-of-00002.safetensors",
"layers.1.attention.q_proj.weight": "model-00001-of-00002.safetensors",
"layers.1.attention.v_proj.bias": "model-00001-of-00002.safetensors",
"layers.1.attention.v_proj.weight": "model-00001-of-00002.safetensors",
"layers.1.layer_norm1.bias": "model-00001-of-00002.safetensors",
"layers.1.layer_norm1.weight": "model-00001-of-00002.safetensors",
"layers.1.layer_norm2.bias": "model-00001-of-00002.safetensors",
"layers.1.layer_norm2.weight": "model-00001-of-00002.safetensors",
"layers.1.mlp.0.bias": "model-00001-of-00002.safetensors",
"layers.1.mlp.0.weight": "model-00001-of-00002.safetensors",
"layers.1.mlp.2.bias": "model-00001-of-00002.safetensors",
"layers.1.mlp.2.weight": "model-00001-of-00002.safetensors",
"layers.10.attention.k_proj.bias": "model-00001-of-00002.safetensors",
"layers.10.attention.k_proj.weight": "model-00001-of-00002.safetensors",
"layers.10.attention.out_proj.bias": "model-00001-of-00002.safetensors",
"layers.10.attention.out_proj.weight": "model-00001-of-00002.safetensors",
"layers.10.attention.q_proj.bias": "model-00001-of-00002.safetensors",
"layers.10.attention.q_proj.weight": "model-00001-of-00002.safetensors",
"layers.10.attention.v_proj.bias": "model-00001-of-00002.safetensors",
"layers.10.attention.v_proj.weight": "model-00001-of-00002.safetensors",
"layers.10.layer_norm1.bias": "model-00001-of-00002.safetensors",
"layers.10.layer_norm1.weight": "model-00001-of-00002.safetensors",
"layers.10.layer_norm2.bias": "model-00001-of-00002.safetensors",
"layers.10.layer_norm2.weight": "model-00001-of-00002.safetensors",
"layers.10.mlp.0.bias": "model-00001-of-00002.safetensors",
"layers.10.mlp.0.weight": "model-00001-of-00002.safetensors",
"layers.10.mlp.2.bias": "model-00001-of-00002.safetensors",
"layers.10.mlp.2.weight": "model-00001-of-00002.safetensors",
"layers.11.attention.k_proj.bias": "model-00001-of-00002.safetensors",
"layers.11.attention.k_proj.weight": "model-00001-of-00002.safetensors",
"layers.11.attention.out_proj.bias": "model-00001-of-00002.safetensors",
"layers.11.attention.out_proj.weight": "model-00001-of-00002.safetensors",
"layers.11.attention.q_proj.bias": "model-00001-of-00002.safetensors",
"layers.11.attention.q_proj.weight": "model-00001-of-00002.safetensors",
"layers.11.attention.v_proj.bias": "model-00001-of-00002.safetensors",
"layers.11.attention.v_proj.weight": "model-00001-of-00002.safetensors",
"layers.11.layer_norm1.bias": "model-00001-of-00002.safetensors",
"layers.11.layer_norm1.weight": "model-00001-of-00002.safetensors",
"layers.11.layer_norm2.bias": "model-00001-of-00002.safetensors",
"layers.11.layer_norm2.weight": "model-00001-of-00002.safetensors",
"layers.11.mlp.0.bias": "model-00001-of-00002.safetensors",
"layers.11.mlp.0.weight": "model-00001-of-00002.safetensors",
"layers.11.mlp.2.bias": "model-00001-of-00002.safetensors",
"layers.11.mlp.2.weight": "model-00001-of-00002.safetensors",
"layers.12.attention.k_proj.bias": "model-00001-of-00002.safetensors",
"layers.12.attention.k_proj.weight": "model-00001-of-00002.safetensors",
"layers.12.attention.out_proj.bias": "model-00001-of-00002.safetensors",
"layers.12.attention.out_proj.weight": "model-00001-of-00002.safetensors",
"layers.12.attention.q_proj.bias": "model-00001-of-00002.safetensors",
"layers.12.attention.q_proj.weight": "model-00001-of-00002.safetensors",
"layers.12.attention.v_proj.bias": "model-00001-of-00002.safetensors",
"layers.12.attention.v_proj.weight": "model-00001-of-00002.safetensors",
"layers.12.layer_norm1.bias": "model-00001-of-00002.safetensors",
"layers.12.layer_norm1.weight": "model-00001-of-00002.safetensors",
"layers.12.layer_norm2.bias": "model-00001-of-00002.safetensors",
"layers.12.layer_norm2.weight": "model-00001-of-00002.safetensors",
"layers.12.mlp.0.bias": "model-00001-of-00002.safetensors",
"layers.12.mlp.0.weight": "model-00001-of-00002.safetensors",
"layers.12.mlp.2.bias": "model-00001-of-00002.safetensors",
"layers.12.mlp.2.weight": "model-00001-of-00002.safetensors",
"layers.13.attention.k_proj.bias": "model-00001-of-00002.safetensors",
"layers.13.attention.k_proj.weight": "model-00001-of-00002.safetensors",
"layers.13.attention.out_proj.bias": "model-00001-of-00002.safetensors",
"layers.13.attention.out_proj.weight": "model-00001-of-00002.safetensors",
"layers.13.attention.q_proj.bias": "model-00001-of-00002.safetensors",
"layers.13.attention.q_proj.weight": "model-00001-of-00002.safetensors",
"layers.13.attention.v_proj.bias": "model-00001-of-00002.safetensors",
"layers.13.attention.v_proj.weight": "model-00001-of-00002.safetensors",
"layers.13.layer_norm1.bias": "model-00001-of-00002.safetensors",
"layers.13.layer_norm1.weight": "model-00001-of-00002.safetensors",
"layers.13.layer_norm2.bias": "model-00001-of-00002.safetensors",
"layers.13.layer_norm2.weight": "model-00001-of-00002.safetensors",
"layers.13.mlp.0.bias": "model-00001-of-00002.safetensors",
"layers.13.mlp.0.weight": "model-00001-of-00002.safetensors",
"layers.13.mlp.2.bias": "model-00001-of-00002.safetensors",
"layers.13.mlp.2.weight": "model-00001-of-00002.safetensors",
"layers.14.attention.k_proj.bias": "model-00001-of-00002.safetensors",
"layers.14.attention.k_proj.weight": "model-00001-of-00002.safetensors",
"layers.14.attention.out_proj.bias": "model-00001-of-00002.safetensors",
"layers.14.attention.out_proj.weight": "model-00001-of-00002.safetensors",
"layers.14.attention.q_proj.bias": "model-00001-of-00002.safetensors",
"layers.14.attention.q_proj.weight": "model-00001-of-00002.safetensors",
"layers.14.attention.v_proj.bias": "model-00001-of-00002.safetensors",
"layers.14.attention.v_proj.weight": "model-00001-of-00002.safetensors",
"layers.14.layer_norm1.bias": "model-00001-of-00002.safetensors",
"layers.14.layer_norm1.weight": "model-00001-of-00002.safetensors",
"layers.14.layer_norm2.bias": "model-00001-of-00002.safetensors",
"layers.14.layer_norm2.weight": "model-00001-of-00002.safetensors",
"layers.14.mlp.0.bias": "model-00001-of-00002.safetensors",
"layers.14.mlp.0.weight": "model-00001-of-00002.safetensors",
"layers.14.mlp.2.bias": "model-00001-of-00002.safetensors",
"layers.14.mlp.2.weight": "model-00001-of-00002.safetensors",
"layers.15.attention.k_proj.bias": "model-00001-of-00002.safetensors",
"layers.15.attention.k_proj.weight": "model-00001-of-00002.safetensors",
"layers.15.attention.out_proj.bias": "model-00001-of-00002.safetensors",
"layers.15.attention.out_proj.weight": "model-00001-of-00002.safetensors",
"layers.15.attention.q_proj.bias": "model-00001-of-00002.safetensors",
"layers.15.attention.q_proj.weight": "model-00001-of-00002.safetensors",
"layers.15.attention.v_proj.bias": "model-00001-of-00002.safetensors",
"layers.15.attention.v_proj.weight": "model-00001-of-00002.safetensors",
"layers.15.layer_norm1.bias": "model-00001-of-00002.safetensors",
"layers.15.layer_norm1.weight": "model-00001-of-00002.safetensors",
"layers.15.layer_norm2.bias": "model-00001-of-00002.safetensors",
"layers.15.layer_norm2.weight": "model-00001-of-00002.safetensors",
"layers.15.mlp.0.bias": "model-00001-of-00002.safetensors",
"layers.15.mlp.0.weight": "model-00001-of-00002.safetensors",
"layers.15.mlp.2.bias": "model-00001-of-00002.safetensors",
"layers.15.mlp.2.weight": "model-00001-of-00002.safetensors",
"layers.16.attention.k_proj.bias": "model-00001-of-00002.safetensors",
"layers.16.attention.k_proj.weight": "model-00001-of-00002.safetensors",
"layers.16.attention.out_proj.bias": "model-00001-of-00002.safetensors",
"layers.16.attention.out_proj.weight": "model-00001-of-00002.safetensors",
"layers.16.attention.q_proj.bias": "model-00001-of-00002.safetensors",
"layers.16.attention.q_proj.weight": "model-00001-of-00002.safetensors",
"layers.16.attention.v_proj.bias": "model-00001-of-00002.safetensors",
"layers.16.attention.v_proj.weight": "model-00001-of-00002.safetensors",
"layers.16.layer_norm1.bias": "model-00001-of-00002.safetensors",
"layers.16.layer_norm1.weight": "model-00001-of-00002.safetensors",
"layers.16.layer_norm2.bias": "model-00001-of-00002.safetensors",
"layers.16.layer_norm2.weight": "model-00001-of-00002.safetensors",
"layers.16.mlp.0.bias": "model-00001-of-00002.safetensors",
"layers.16.mlp.0.weight": "model-00001-of-00002.safetensors",
"layers.16.mlp.2.bias": "model-00001-of-00002.safetensors",
"layers.16.mlp.2.weight": "model-00001-of-00002.safetensors",
"layers.17.attention.k_proj.bias": "model-00001-of-00002.safetensors",
"layers.17.attention.k_proj.weight": "model-00001-of-00002.safetensors",
"layers.17.attention.out_proj.bias": "model-00001-of-00002.safetensors",
"layers.17.attention.out_proj.weight": "model-00001-of-00002.safetensors",
"layers.17.attention.q_proj.bias": "model-00001-of-00002.safetensors",
"layers.17.attention.q_proj.weight": "model-00001-of-00002.safetensors",
"layers.17.attention.v_proj.bias": "model-00001-of-00002.safetensors",
"layers.17.attention.v_proj.weight": "model-00001-of-00002.safetensors",
"layers.17.layer_norm1.bias": "model-00001-of-00002.safetensors",
"layers.17.layer_norm1.weight": "model-00001-of-00002.safetensors",
"layers.17.layer_norm2.bias": "model-00001-of-00002.safetensors",
"layers.17.layer_norm2.weight": "model-00001-of-00002.safetensors",
"layers.17.mlp.0.bias": "model-00001-of-00002.safetensors",
"layers.17.mlp.0.weight": "model-00001-of-00002.safetensors",
"layers.17.mlp.2.bias": "model-00001-of-00002.safetensors",
"layers.17.mlp.2.weight": "model-00001-of-00002.safetensors",
"layers.18.attention.k_proj.bias": "model-00001-of-00002.safetensors",
"layers.18.attention.k_proj.weight": "model-00001-of-00002.safetensors",
"layers.18.attention.out_proj.bias": "model-00001-of-00002.safetensors",
"layers.18.attention.out_proj.weight": "model-00001-of-00002.safetensors",
"layers.18.attention.q_proj.bias": "model-00001-of-00002.safetensors",
"layers.18.attention.q_proj.weight": "model-00001-of-00002.safetensors",
"layers.18.attention.v_proj.bias": "model-00001-of-00002.safetensors",
"layers.18.attention.v_proj.weight": "model-00001-of-00002.safetensors",
"layers.18.layer_norm1.bias": "model-00001-of-00002.safetensors",
"layers.18.layer_norm1.weight": "model-00001-of-00002.safetensors",
"layers.18.layer_norm2.bias": "model-00001-of-00002.safetensors",
"layers.18.layer_norm2.weight": "model-00001-of-00002.safetensors",
"layers.18.mlp.0.bias": "model-00001-of-00002.safetensors",
"layers.18.mlp.0.weight": "model-00001-of-00002.safetensors",
"layers.18.mlp.2.bias": "model-00001-of-00002.safetensors",
"layers.18.mlp.2.weight": "model-00001-of-00002.safetensors",
"layers.19.attention.k_proj.bias": "model-00001-of-00002.safetensors",
"layers.19.attention.k_proj.weight": "model-00001-of-00002.safetensors",
"layers.19.attention.out_proj.bias": "model-00001-of-00002.safetensors",
"layers.19.attention.out_proj.weight": "model-00001-of-00002.safetensors",
"layers.19.attention.q_proj.bias": "model-00001-of-00002.safetensors",
"layers.19.attention.q_proj.weight": "model-00001-of-00002.safetensors",
"layers.19.attention.v_proj.bias": "model-00001-of-00002.safetensors",
"layers.19.attention.v_proj.weight": "model-00001-of-00002.safetensors",
"layers.19.layer_norm1.bias": "model-00001-of-00002.safetensors",
"layers.19.layer_norm1.weight": "model-00001-of-00002.safetensors",
"layers.19.layer_norm2.bias": "model-00001-of-00002.safetensors",
"layers.19.layer_norm2.weight": "model-00001-of-00002.safetensors",
"layers.19.mlp.0.bias": "model-00001-of-00002.safetensors",
"layers.19.mlp.0.weight": "model-00001-of-00002.safetensors",
"layers.19.mlp.2.bias": "model-00001-of-00002.safetensors",
"layers.19.mlp.2.weight": "model-00001-of-00002.safetensors",
"layers.2.attention.k_proj.bias": "model-00001-of-00002.safetensors",
"layers.2.attention.k_proj.weight": "model-00001-of-00002.safetensors",
"layers.2.attention.out_proj.bias": "model-00001-of-00002.safetensors",
"layers.2.attention.out_proj.weight": "model-00001-of-00002.safetensors",
"layers.2.attention.q_proj.bias": "model-00001-of-00002.safetensors",
"layers.2.attention.q_proj.weight": "model-00001-of-00002.safetensors",
"layers.2.attention.v_proj.bias": "model-00001-of-00002.safetensors",
"layers.2.attention.v_proj.weight": "model-00001-of-00002.safetensors",
"layers.2.layer_norm1.bias": "model-00001-of-00002.safetensors",
"layers.2.layer_norm1.weight": "model-00001-of-00002.safetensors",
"layers.2.layer_norm2.bias": "model-00001-of-00002.safetensors",
"layers.2.layer_norm2.weight": "model-00001-of-00002.safetensors",
"layers.2.mlp.0.bias": "model-00001-of-00002.safetensors",
"layers.2.mlp.0.weight": "model-00001-of-00002.safetensors",
"layers.2.mlp.2.bias": "model-00001-of-00002.safetensors",
"layers.2.mlp.2.weight": "model-00001-of-00002.safetensors",
"layers.20.attention.k_proj.bias": "model-00001-of-00002.safetensors",
"layers.20.attention.k_proj.weight": "model-00001-of-00002.safetensors",
"layers.20.attention.out_proj.bias": "model-00001-of-00002.safetensors",
"layers.20.attention.out_proj.weight": "model-00001-of-00002.safetensors",
"layers.20.attention.q_proj.bias": "model-00001-of-00002.safetensors",
"layers.20.attention.q_proj.weight": "model-00001-of-00002.safetensors",
"layers.20.attention.v_proj.bias": "model-00001-of-00002.safetensors",
"layers.20.attention.v_proj.weight": "model-00001-of-00002.safetensors",
"layers.20.layer_norm1.bias": "model-00001-of-00002.safetensors",
"layers.20.layer_norm1.weight": "model-00001-of-00002.safetensors",
"layers.20.layer_norm2.bias": "model-00001-of-00002.safetensors",
"layers.20.layer_norm2.weight": "model-00001-of-00002.safetensors",
"layers.20.mlp.0.bias": "model-00001-of-00002.safetensors",
"layers.20.mlp.0.weight": "model-00001-of-00002.safetensors",
"layers.20.mlp.2.bias": "model-00001-of-00002.safetensors",
"layers.20.mlp.2.weight": "model-00001-of-00002.safetensors",
"layers.21.attention.k_proj.bias": "model-00001-of-00002.safetensors",
"layers.21.attention.k_proj.weight": "model-00001-of-00002.safetensors",
"layers.21.attention.out_proj.bias": "model-00001-of-00002.safetensors",
"layers.21.attention.out_proj.weight": "model-00001-of-00002.safetensors",
"layers.21.attention.q_proj.bias": "model-00001-of-00002.safetensors",
"layers.21.attention.q_proj.weight": "model-00001-of-00002.safetensors",
"layers.21.attention.v_proj.bias": "model-00001-of-00002.safetensors",
"layers.21.attention.v_proj.weight": "model-00001-of-00002.safetensors",
"layers.21.layer_norm1.bias": "model-00002-of-00002.safetensors",
"layers.21.layer_norm1.weight": "model-00002-of-00002.safetensors",
"layers.21.layer_norm2.bias": "model-00002-of-00002.safetensors",
"layers.21.layer_norm2.weight": "model-00002-of-00002.safetensors",
"layers.21.mlp.0.bias": "model-00002-of-00002.safetensors",
"layers.21.mlp.0.weight": "model-00002-of-00002.safetensors",
"layers.21.mlp.2.bias": "model-00002-of-00002.safetensors",
"layers.21.mlp.2.weight": "model-00002-of-00002.safetensors",
"layers.22.attention.k_proj.bias": "model-00002-of-00002.safetensors",
"layers.22.attention.k_proj.weight": "model-00002-of-00002.safetensors",
"layers.22.attention.out_proj.bias": "model-00002-of-00002.safetensors",
"layers.22.attention.out_proj.weight": "model-00002-of-00002.safetensors",
"layers.22.attention.q_proj.bias": "model-00002-of-00002.safetensors",
"layers.22.attention.q_proj.weight": "model-00002-of-00002.safetensors",
"layers.22.attention.v_proj.bias": "model-00002-of-00002.safetensors",
"layers.22.attention.v_proj.weight": "model-00002-of-00002.safetensors",
"layers.22.layer_norm1.bias": "model-00002-of-00002.safetensors",
"layers.22.layer_norm1.weight": "model-00002-of-00002.safetensors",
"layers.22.layer_norm2.bias": "model-00002-of-00002.safetensors",
"layers.22.layer_norm2.weight": "model-00002-of-00002.safetensors",
"layers.22.mlp.0.bias": "model-00002-of-00002.safetensors",
"layers.22.mlp.0.weight": "model-00002-of-00002.safetensors",
"layers.22.mlp.2.bias": "model-00002-of-00002.safetensors",
"layers.22.mlp.2.weight": "model-00002-of-00002.safetensors",
"layers.23.attention.k_proj.bias": "model-00002-of-00002.safetensors",
"layers.23.attention.k_proj.weight": "model-00002-of-00002.safetensors",
"layers.23.attention.out_proj.bias": "model-00002-of-00002.safetensors",
"layers.23.attention.out_proj.weight": "model-00002-of-00002.safetensors",
"layers.23.attention.q_proj.bias": "model-00002-of-00002.safetensors",
"layers.23.attention.q_proj.weight": "model-00002-of-00002.safetensors",
"layers.23.attention.v_proj.bias": "model-00002-of-00002.safetensors",
"layers.23.attention.v_proj.weight": "model-00002-of-00002.safetensors",
"layers.23.layer_norm1.bias": "model-00002-of-00002.safetensors",
"layers.23.layer_norm1.weight": "model-00002-of-00002.safetensors",
"layers.23.layer_norm2.bias": "model-00002-of-00002.safetensors",
"layers.23.layer_norm2.weight": "model-00002-of-00002.safetensors",
"layers.23.mlp.0.bias": "model-00002-of-00002.safetensors",
"layers.23.mlp.0.weight": "model-00002-of-00002.safetensors",
"layers.23.mlp.2.bias": "model-00002-of-00002.safetensors",
"layers.23.mlp.2.weight": "model-00002-of-00002.safetensors",
"layers.3.attention.k_proj.bias": "model-00001-of-00002.safetensors",
"layers.3.attention.k_proj.weight": "model-00001-of-00002.safetensors",
"layers.3.attention.out_proj.bias": "model-00001-of-00002.safetensors",
"layers.3.attention.out_proj.weight": "model-00001-of-00002.safetensors",
"layers.3.attention.q_proj.bias": "model-00001-of-00002.safetensors",
"layers.3.attention.q_proj.weight": "model-00001-of-00002.safetensors",
"layers.3.attention.v_proj.bias": "model-00001-of-00002.safetensors",
"layers.3.attention.v_proj.weight": "model-00001-of-00002.safetensors",
"layers.3.layer_norm1.bias": "model-00001-of-00002.safetensors",
"layers.3.layer_norm1.weight": "model-00001-of-00002.safetensors",
"layers.3.layer_norm2.bias": "model-00001-of-00002.safetensors",
"layers.3.layer_norm2.weight": "model-00001-of-00002.safetensors",
"layers.3.mlp.0.bias": "model-00001-of-00002.safetensors",
"layers.3.mlp.0.weight": "model-00001-of-00002.safetensors",
"layers.3.mlp.2.bias": "model-00001-of-00002.safetensors",
"layers.3.mlp.2.weight": "model-00001-of-00002.safetensors",
"layers.4.attention.k_proj.bias": "model-00001-of-00002.safetensors",
"layers.4.attention.k_proj.weight": "model-00001-of-00002.safetensors",
"layers.4.attention.out_proj.bias": "model-00001-of-00002.safetensors",
"layers.4.attention.out_proj.weight": "model-00001-of-00002.safetensors",
"layers.4.attention.q_proj.bias": "model-00001-of-00002.safetensors",
"layers.4.attention.q_proj.weight": "model-00001-of-00002.safetensors",
"layers.4.attention.v_proj.bias": "model-00001-of-00002.safetensors",
"layers.4.attention.v_proj.weight": "model-00001-of-00002.safetensors",
"layers.4.layer_norm1.bias": "model-00001-of-00002.safetensors",
"layers.4.layer_norm1.weight": "model-00001-of-00002.safetensors",
"layers.4.layer_norm2.bias": "model-00001-of-00002.safetensors",
"layers.4.layer_norm2.weight": "model-00001-of-00002.safetensors",
"layers.4.mlp.0.bias": "model-00001-of-00002.safetensors",
"layers.4.mlp.0.weight": "model-00001-of-00002.safetensors",
"layers.4.mlp.2.bias": "model-00001-of-00002.safetensors",
"layers.4.mlp.2.weight": "model-00001-of-00002.safetensors",
"layers.5.attention.k_proj.bias": "model-00001-of-00002.safetensors",
"layers.5.attention.k_proj.weight": "model-00001-of-00002.safetensors",
"layers.5.attention.out_proj.bias": "model-00001-of-00002.safetensors",
"layers.5.attention.out_proj.weight": "model-00001-of-00002.safetensors",
"layers.5.attention.q_proj.bias": "model-00001-of-00002.safetensors",
"layers.5.attention.q_proj.weight": "model-00001-of-00002.safetensors",
"layers.5.attention.v_proj.bias": "model-00001-of-00002.safetensors",
"layers.5.attention.v_proj.weight": "model-00001-of-00002.safetensors",
"layers.5.layer_norm1.bias": "model-00001-of-00002.safetensors",
"layers.5.layer_norm1.weight": "model-00001-of-00002.safetensors",
"layers.5.layer_norm2.bias": "model-00001-of-00002.safetensors",
"layers.5.layer_norm2.weight": "model-00001-of-00002.safetensors",
"layers.5.mlp.0.bias": "model-00001-of-00002.safetensors",
"layers.5.mlp.0.weight": "model-00001-of-00002.safetensors",
"layers.5.mlp.2.bias": "model-00001-of-00002.safetensors",
"layers.5.mlp.2.weight": "model-00001-of-00002.safetensors",
"layers.6.attention.k_proj.bias": "model-00001-of-00002.safetensors",
"layers.6.attention.k_proj.weight": "model-00001-of-00002.safetensors",
"layers.6.attention.out_proj.bias": "model-00001-of-00002.safetensors",
"layers.6.attention.out_proj.weight": "model-00001-of-00002.safetensors",
"layers.6.attention.q_proj.bias": "model-00001-of-00002.safetensors",
"layers.6.attention.q_proj.weight": "model-00001-of-00002.safetensors",
"layers.6.attention.v_proj.bias": "model-00001-of-00002.safetensors",
"layers.6.attention.v_proj.weight": "model-00001-of-00002.safetensors",
"layers.6.layer_norm1.bias": "model-00001-of-00002.safetensors",
"layers.6.layer_norm1.weight": "model-00001-of-00002.safetensors",
"layers.6.layer_norm2.bias": "model-00001-of-00002.safetensors",
"layers.6.layer_norm2.weight": "model-00001-of-00002.safetensors",
"layers.6.mlp.0.bias": "model-00001-of-00002.safetensors",
"layers.6.mlp.0.weight": "model-00001-of-00002.safetensors",
"layers.6.mlp.2.bias": "model-00001-of-00002.safetensors",
"layers.6.mlp.2.weight": "model-00001-of-00002.safetensors",
"layers.7.attention.k_proj.bias": "model-00001-of-00002.safetensors",
"layers.7.attention.k_proj.weight": "model-00001-of-00002.safetensors",
"layers.7.attention.out_proj.bias": "model-00001-of-00002.safetensors",
"layers.7.attention.out_proj.weight": "model-00001-of-00002.safetensors",
"layers.7.attention.q_proj.bias": "model-00001-of-00002.safetensors",
"layers.7.attention.q_proj.weight": "model-00001-of-00002.safetensors",
"layers.7.attention.v_proj.bias": "model-00001-of-00002.safetensors",
"layers.7.attention.v_proj.weight": "model-00001-of-00002.safetensors",
"layers.7.layer_norm1.bias": "model-00001-of-00002.safetensors",
"layers.7.layer_norm1.weight": "model-00001-of-00002.safetensors",
"layers.7.layer_norm2.bias": "model-00001-of-00002.safetensors",
"layers.7.layer_norm2.weight": "model-00001-of-00002.safetensors",
"layers.7.mlp.0.bias": "model-00001-of-00002.safetensors",
"layers.7.mlp.0.weight": "model-00001-of-00002.safetensors",
"layers.7.mlp.2.bias": "model-00001-of-00002.safetensors",
"layers.7.mlp.2.weight": "model-00001-of-00002.safetensors",
"layers.8.attention.k_proj.bias": "model-00001-of-00002.safetensors",
"layers.8.attention.k_proj.weight": "model-00001-of-00002.safetensors",
"layers.8.attention.out_proj.bias": "model-00001-of-00002.safetensors",
"layers.8.attention.out_proj.weight": "model-00001-of-00002.safetensors",
"layers.8.attention.q_proj.bias": "model-00001-of-00002.safetensors",
"layers.8.attention.q_proj.weight": "model-00001-of-00002.safetensors",
"layers.8.attention.v_proj.bias": "model-00001-of-00002.safetensors",
"layers.8.attention.v_proj.weight": "model-00001-of-00002.safetensors",
"layers.8.layer_norm1.bias": "model-00001-of-00002.safetensors",
"layers.8.layer_norm1.weight": "model-00001-of-00002.safetensors",
"layers.8.layer_norm2.bias": "model-00001-of-00002.safetensors",
"layers.8.layer_norm2.weight": "model-00001-of-00002.safetensors",
"layers.8.mlp.0.bias": "model-00001-of-00002.safetensors",
"layers.8.mlp.0.weight": "model-00001-of-00002.safetensors",
"layers.8.mlp.2.bias": "model-00001-of-00002.safetensors",
"layers.8.mlp.2.weight": "model-00001-of-00002.safetensors",
"layers.9.attention.k_proj.bias": "model-00001-of-00002.safetensors",
"layers.9.attention.k_proj.weight": "model-00001-of-00002.safetensors",
"layers.9.attention.out_proj.bias": "model-00001-of-00002.safetensors",
"layers.9.attention.out_proj.weight": "model-00001-of-00002.safetensors",
"layers.9.attention.q_proj.bias": "model-00001-of-00002.safetensors",
"layers.9.attention.q_proj.weight": "model-00001-of-00002.safetensors",
"layers.9.attention.v_proj.bias": "model-00001-of-00002.safetensors",
"layers.9.attention.v_proj.weight": "model-00001-of-00002.safetensors",
"layers.9.layer_norm1.bias": "model-00001-of-00002.safetensors",
"layers.9.layer_norm1.weight": "model-00001-of-00002.safetensors",
"layers.9.layer_norm2.bias": "model-00001-of-00002.safetensors",
"layers.9.layer_norm2.weight": "model-00001-of-00002.safetensors",
"layers.9.mlp.0.bias": "model-00001-of-00002.safetensors",
"layers.9.mlp.0.weight": "model-00001-of-00002.safetensors",
"layers.9.mlp.2.bias": "model-00001-of-00002.safetensors",
"layers.9.mlp.2.weight": "model-00001-of-00002.safetensors",
"output_head.bias": "model-00002-of-00002.safetensors",
"output_head.weight": "model-00002-of-00002.safetensors",
"position_embedding.weight": "model-00001-of-00002.safetensors"
}
}