{ "metadata": { "total_size": 10046866944 }, "weight_map": { "embedding.weight": "model-00001-of-00003.safetensors", "layer_norm.bias": "model-00002-of-00003.safetensors", "layer_norm.weight": "model-00002-of-00003.safetensors", "layers.0.attention.k_proj.bias": "model-00001-of-00003.safetensors", "layers.0.attention.k_proj.weight": "model-00001-of-00003.safetensors", "layers.0.attention.out_proj.bias": "model-00001-of-00003.safetensors", "layers.0.attention.out_proj.weight": "model-00001-of-00003.safetensors", "layers.0.attention.q_proj.bias": "model-00001-of-00003.safetensors", "layers.0.attention.q_proj.weight": "model-00001-of-00003.safetensors", "layers.0.attention.v_proj.bias": "model-00001-of-00003.safetensors", "layers.0.attention.v_proj.weight": "model-00001-of-00003.safetensors", "layers.0.layer_norm1.bias": "model-00001-of-00003.safetensors", "layers.0.layer_norm1.weight": "model-00001-of-00003.safetensors", "layers.0.layer_norm2.bias": "model-00001-of-00003.safetensors", "layers.0.layer_norm2.weight": "model-00001-of-00003.safetensors", "layers.0.mlp.0.bias": "model-00001-of-00003.safetensors", "layers.0.mlp.0.weight": "model-00001-of-00003.safetensors", "layers.0.mlp.2.bias": "model-00001-of-00003.safetensors", "layers.0.mlp.2.weight": "model-00001-of-00003.safetensors", "layers.1.attention.k_proj.bias": "model-00001-of-00003.safetensors", "layers.1.attention.k_proj.weight": "model-00001-of-00003.safetensors", "layers.1.attention.out_proj.bias": "model-00001-of-00003.safetensors", "layers.1.attention.out_proj.weight": "model-00001-of-00003.safetensors", "layers.1.attention.q_proj.bias": "model-00001-of-00003.safetensors", "layers.1.attention.q_proj.weight": "model-00001-of-00003.safetensors", "layers.1.attention.v_proj.bias": "model-00001-of-00003.safetensors", "layers.1.attention.v_proj.weight": "model-00001-of-00003.safetensors", "layers.1.layer_norm1.bias": "model-00001-of-00003.safetensors", "layers.1.layer_norm1.weight": "model-00001-of-00003.safetensors", "layers.1.layer_norm2.bias": "model-00001-of-00003.safetensors", "layers.1.layer_norm2.weight": "model-00001-of-00003.safetensors", "layers.1.mlp.0.bias": "model-00001-of-00003.safetensors", "layers.1.mlp.0.weight": "model-00001-of-00003.safetensors", "layers.1.mlp.2.bias": "model-00001-of-00003.safetensors", "layers.1.mlp.2.weight": "model-00001-of-00003.safetensors", "layers.10.attention.k_proj.bias": "model-00001-of-00003.safetensors", "layers.10.attention.k_proj.weight": "model-00001-of-00003.safetensors", "layers.10.attention.out_proj.bias": "model-00001-of-00003.safetensors", "layers.10.attention.out_proj.weight": "model-00001-of-00003.safetensors", "layers.10.attention.q_proj.bias": "model-00001-of-00003.safetensors", "layers.10.attention.q_proj.weight": "model-00001-of-00003.safetensors", "layers.10.attention.v_proj.bias": "model-00001-of-00003.safetensors", "layers.10.attention.v_proj.weight": "model-00001-of-00003.safetensors", "layers.10.layer_norm1.bias": "model-00001-of-00003.safetensors", "layers.10.layer_norm1.weight": "model-00001-of-00003.safetensors", "layers.10.layer_norm2.bias": "model-00001-of-00003.safetensors", "layers.10.layer_norm2.weight": "model-00001-of-00003.safetensors", "layers.10.mlp.0.bias": "model-00001-of-00003.safetensors", "layers.10.mlp.0.weight": "model-00001-of-00003.safetensors", "layers.10.mlp.2.bias": "model-00001-of-00003.safetensors", "layers.10.mlp.2.weight": "model-00001-of-00003.safetensors", "layers.11.attention.k_proj.bias": "model-00001-of-00003.safetensors", "layers.11.attention.k_proj.weight": "model-00001-of-00003.safetensors", "layers.11.attention.out_proj.bias": "model-00001-of-00003.safetensors", "layers.11.attention.out_proj.weight": "model-00001-of-00003.safetensors", "layers.11.attention.q_proj.bias": "model-00001-of-00003.safetensors", "layers.11.attention.q_proj.weight": "model-00001-of-00003.safetensors", "layers.11.attention.v_proj.bias": "model-00001-of-00003.safetensors", "layers.11.attention.v_proj.weight": "model-00001-of-00003.safetensors", "layers.11.layer_norm1.bias": "model-00001-of-00003.safetensors", "layers.11.layer_norm1.weight": "model-00001-of-00003.safetensors", "layers.11.layer_norm2.bias": "model-00001-of-00003.safetensors", "layers.11.layer_norm2.weight": "model-00001-of-00003.safetensors", "layers.11.mlp.0.bias": "model-00001-of-00003.safetensors", "layers.11.mlp.0.weight": "model-00001-of-00003.safetensors", "layers.11.mlp.2.bias": "model-00001-of-00003.safetensors", "layers.11.mlp.2.weight": "model-00001-of-00003.safetensors", "layers.12.attention.k_proj.bias": "model-00001-of-00003.safetensors", "layers.12.attention.k_proj.weight": "model-00001-of-00003.safetensors", "layers.12.attention.out_proj.bias": "model-00001-of-00003.safetensors", "layers.12.attention.out_proj.weight": "model-00001-of-00003.safetensors", "layers.12.attention.q_proj.bias": "model-00001-of-00003.safetensors", "layers.12.attention.q_proj.weight": "model-00001-of-00003.safetensors", "layers.12.attention.v_proj.bias": "model-00001-of-00003.safetensors", "layers.12.attention.v_proj.weight": "model-00001-of-00003.safetensors", "layers.12.layer_norm1.bias": "model-00001-of-00003.safetensors", "layers.12.layer_norm1.weight": "model-00001-of-00003.safetensors", "layers.12.layer_norm2.bias": "model-00001-of-00003.safetensors", "layers.12.layer_norm2.weight": "model-00001-of-00003.safetensors", "layers.12.mlp.0.bias": "model-00001-of-00003.safetensors", "layers.12.mlp.0.weight": "model-00001-of-00003.safetensors", "layers.12.mlp.2.bias": "model-00001-of-00003.safetensors", "layers.12.mlp.2.weight": "model-00001-of-00003.safetensors", "layers.13.attention.k_proj.bias": "model-00001-of-00003.safetensors", "layers.13.attention.k_proj.weight": "model-00001-of-00003.safetensors", "layers.13.attention.out_proj.bias": "model-00001-of-00003.safetensors", "layers.13.attention.out_proj.weight": "model-00001-of-00003.safetensors", "layers.13.attention.q_proj.bias": "model-00001-of-00003.safetensors", "layers.13.attention.q_proj.weight": "model-00001-of-00003.safetensors", "layers.13.attention.v_proj.bias": "model-00001-of-00003.safetensors", "layers.13.attention.v_proj.weight": "model-00001-of-00003.safetensors", "layers.13.layer_norm1.bias": "model-00001-of-00003.safetensors", "layers.13.layer_norm1.weight": "model-00001-of-00003.safetensors", "layers.13.layer_norm2.bias": "model-00001-of-00003.safetensors", "layers.13.layer_norm2.weight": "model-00001-of-00003.safetensors", "layers.13.mlp.0.bias": "model-00001-of-00003.safetensors", "layers.13.mlp.0.weight": "model-00001-of-00003.safetensors", "layers.13.mlp.2.bias": "model-00001-of-00003.safetensors", "layers.13.mlp.2.weight": "model-00001-of-00003.safetensors", "layers.14.attention.k_proj.bias": "model-00001-of-00003.safetensors", "layers.14.attention.k_proj.weight": "model-00001-of-00003.safetensors", "layers.14.attention.out_proj.bias": "model-00001-of-00003.safetensors", "layers.14.attention.out_proj.weight": "model-00001-of-00003.safetensors", "layers.14.attention.q_proj.bias": "model-00001-of-00003.safetensors", "layers.14.attention.q_proj.weight": "model-00001-of-00003.safetensors", "layers.14.attention.v_proj.bias": "model-00001-of-00003.safetensors", "layers.14.attention.v_proj.weight": "model-00001-of-00003.safetensors", "layers.14.layer_norm1.bias": "model-00002-of-00003.safetensors", "layers.14.layer_norm1.weight": "model-00002-of-00003.safetensors", "layers.14.layer_norm2.bias": "model-00002-of-00003.safetensors", "layers.14.layer_norm2.weight": "model-00002-of-00003.safetensors", "layers.14.mlp.0.bias": "model-00001-of-00003.safetensors", "layers.14.mlp.0.weight": "model-00001-of-00003.safetensors", "layers.14.mlp.2.bias": "model-00002-of-00003.safetensors", "layers.14.mlp.2.weight": "model-00002-of-00003.safetensors", "layers.15.attention.k_proj.bias": "model-00002-of-00003.safetensors", "layers.15.attention.k_proj.weight": "model-00002-of-00003.safetensors", "layers.15.attention.out_proj.bias": "model-00002-of-00003.safetensors", "layers.15.attention.out_proj.weight": "model-00002-of-00003.safetensors", "layers.15.attention.q_proj.bias": "model-00002-of-00003.safetensors", "layers.15.attention.q_proj.weight": "model-00002-of-00003.safetensors", "layers.15.attention.v_proj.bias": "model-00002-of-00003.safetensors", "layers.15.attention.v_proj.weight": "model-00002-of-00003.safetensors", "layers.15.layer_norm1.bias": "model-00002-of-00003.safetensors", "layers.15.layer_norm1.weight": "model-00002-of-00003.safetensors", "layers.15.layer_norm2.bias": "model-00002-of-00003.safetensors", "layers.15.layer_norm2.weight": "model-00002-of-00003.safetensors", "layers.15.mlp.0.bias": "model-00002-of-00003.safetensors", "layers.15.mlp.0.weight": "model-00002-of-00003.safetensors", "layers.15.mlp.2.bias": "model-00002-of-00003.safetensors", "layers.15.mlp.2.weight": "model-00002-of-00003.safetensors", "layers.16.attention.k_proj.bias": "model-00002-of-00003.safetensors", "layers.16.attention.k_proj.weight": "model-00002-of-00003.safetensors", "layers.16.attention.out_proj.bias": "model-00002-of-00003.safetensors", "layers.16.attention.out_proj.weight": "model-00002-of-00003.safetensors", "layers.16.attention.q_proj.bias": "model-00002-of-00003.safetensors", "layers.16.attention.q_proj.weight": "model-00002-of-00003.safetensors", "layers.16.attention.v_proj.bias": "model-00002-of-00003.safetensors", "layers.16.attention.v_proj.weight": "model-00002-of-00003.safetensors", "layers.16.layer_norm1.bias": "model-00002-of-00003.safetensors", "layers.16.layer_norm1.weight": "model-00002-of-00003.safetensors", "layers.16.layer_norm2.bias": "model-00002-of-00003.safetensors", "layers.16.layer_norm2.weight": "model-00002-of-00003.safetensors", "layers.16.mlp.0.bias": "model-00002-of-00003.safetensors", "layers.16.mlp.0.weight": "model-00002-of-00003.safetensors", "layers.16.mlp.2.bias": "model-00002-of-00003.safetensors", "layers.16.mlp.2.weight": "model-00002-of-00003.safetensors", "layers.17.attention.k_proj.bias": "model-00002-of-00003.safetensors", "layers.17.attention.k_proj.weight": "model-00002-of-00003.safetensors", "layers.17.attention.out_proj.bias": "model-00002-of-00003.safetensors", "layers.17.attention.out_proj.weight": "model-00002-of-00003.safetensors", "layers.17.attention.q_proj.bias": "model-00002-of-00003.safetensors", "layers.17.attention.q_proj.weight": "model-00002-of-00003.safetensors", "layers.17.attention.v_proj.bias": "model-00002-of-00003.safetensors", "layers.17.attention.v_proj.weight": "model-00002-of-00003.safetensors", "layers.17.layer_norm1.bias": "model-00002-of-00003.safetensors", "layers.17.layer_norm1.weight": "model-00002-of-00003.safetensors", "layers.17.layer_norm2.bias": "model-00002-of-00003.safetensors", "layers.17.layer_norm2.weight": "model-00002-of-00003.safetensors", "layers.17.mlp.0.bias": "model-00002-of-00003.safetensors", "layers.17.mlp.0.weight": "model-00002-of-00003.safetensors", "layers.17.mlp.2.bias": "model-00002-of-00003.safetensors", "layers.17.mlp.2.weight": "model-00002-of-00003.safetensors", "layers.18.attention.k_proj.bias": "model-00002-of-00003.safetensors", "layers.18.attention.k_proj.weight": "model-00002-of-00003.safetensors", "layers.18.attention.out_proj.bias": "model-00002-of-00003.safetensors", "layers.18.attention.out_proj.weight": "model-00002-of-00003.safetensors", "layers.18.attention.q_proj.bias": "model-00002-of-00003.safetensors", "layers.18.attention.q_proj.weight": "model-00002-of-00003.safetensors", "layers.18.attention.v_proj.bias": "model-00002-of-00003.safetensors", "layers.18.attention.v_proj.weight": "model-00002-of-00003.safetensors", "layers.18.layer_norm1.bias": "model-00002-of-00003.safetensors", "layers.18.layer_norm1.weight": "model-00002-of-00003.safetensors", "layers.18.layer_norm2.bias": "model-00002-of-00003.safetensors", "layers.18.layer_norm2.weight": "model-00002-of-00003.safetensors", "layers.18.mlp.0.bias": "model-00002-of-00003.safetensors", "layers.18.mlp.0.weight": "model-00002-of-00003.safetensors", "layers.18.mlp.2.bias": "model-00002-of-00003.safetensors", "layers.18.mlp.2.weight": "model-00002-of-00003.safetensors", "layers.19.attention.k_proj.bias": "model-00002-of-00003.safetensors", "layers.19.attention.k_proj.weight": "model-00002-of-00003.safetensors", "layers.19.attention.out_proj.bias": "model-00002-of-00003.safetensors", "layers.19.attention.out_proj.weight": "model-00002-of-00003.safetensors", "layers.19.attention.q_proj.bias": "model-00002-of-00003.safetensors", "layers.19.attention.q_proj.weight": "model-00002-of-00003.safetensors", "layers.19.attention.v_proj.bias": "model-00002-of-00003.safetensors", "layers.19.attention.v_proj.weight": "model-00002-of-00003.safetensors", "layers.19.layer_norm1.bias": "model-00002-of-00003.safetensors", "layers.19.layer_norm1.weight": "model-00002-of-00003.safetensors", "layers.19.layer_norm2.bias": "model-00002-of-00003.safetensors", "layers.19.layer_norm2.weight": "model-00002-of-00003.safetensors", "layers.19.mlp.0.bias": "model-00002-of-00003.safetensors", "layers.19.mlp.0.weight": "model-00002-of-00003.safetensors", "layers.19.mlp.2.bias": "model-00002-of-00003.safetensors", "layers.19.mlp.2.weight": "model-00002-of-00003.safetensors", "layers.2.attention.k_proj.bias": "model-00001-of-00003.safetensors", "layers.2.attention.k_proj.weight": "model-00001-of-00003.safetensors", "layers.2.attention.out_proj.bias": "model-00001-of-00003.safetensors", "layers.2.attention.out_proj.weight": "model-00001-of-00003.safetensors", "layers.2.attention.q_proj.bias": "model-00001-of-00003.safetensors", "layers.2.attention.q_proj.weight": "model-00001-of-00003.safetensors", "layers.2.attention.v_proj.bias": "model-00001-of-00003.safetensors", "layers.2.attention.v_proj.weight": "model-00001-of-00003.safetensors", "layers.2.layer_norm1.bias": "model-00001-of-00003.safetensors", "layers.2.layer_norm1.weight": "model-00001-of-00003.safetensors", "layers.2.layer_norm2.bias": "model-00001-of-00003.safetensors", "layers.2.layer_norm2.weight": "model-00001-of-00003.safetensors", "layers.2.mlp.0.bias": "model-00001-of-00003.safetensors", "layers.2.mlp.0.weight": "model-00001-of-00003.safetensors", "layers.2.mlp.2.bias": "model-00001-of-00003.safetensors", "layers.2.mlp.2.weight": "model-00001-of-00003.safetensors", "layers.20.attention.k_proj.bias": "model-00002-of-00003.safetensors", "layers.20.attention.k_proj.weight": "model-00002-of-00003.safetensors", "layers.20.attention.out_proj.bias": "model-00002-of-00003.safetensors", "layers.20.attention.out_proj.weight": "model-00002-of-00003.safetensors", "layers.20.attention.q_proj.bias": "model-00002-of-00003.safetensors", "layers.20.attention.q_proj.weight": "model-00002-of-00003.safetensors", "layers.20.attention.v_proj.bias": "model-00002-of-00003.safetensors", "layers.20.attention.v_proj.weight": "model-00002-of-00003.safetensors", "layers.20.layer_norm1.bias": "model-00002-of-00003.safetensors", "layers.20.layer_norm1.weight": "model-00002-of-00003.safetensors", "layers.20.layer_norm2.bias": "model-00002-of-00003.safetensors", "layers.20.layer_norm2.weight": "model-00002-of-00003.safetensors", "layers.20.mlp.0.bias": "model-00002-of-00003.safetensors", "layers.20.mlp.0.weight": "model-00002-of-00003.safetensors", "layers.20.mlp.2.bias": "model-00002-of-00003.safetensors", "layers.20.mlp.2.weight": "model-00002-of-00003.safetensors", "layers.21.attention.k_proj.bias": "model-00002-of-00003.safetensors", "layers.21.attention.k_proj.weight": "model-00002-of-00003.safetensors", "layers.21.attention.out_proj.bias": "model-00002-of-00003.safetensors", "layers.21.attention.out_proj.weight": "model-00002-of-00003.safetensors", "layers.21.attention.q_proj.bias": "model-00002-of-00003.safetensors", "layers.21.attention.q_proj.weight": "model-00002-of-00003.safetensors", "layers.21.attention.v_proj.bias": "model-00002-of-00003.safetensors", "layers.21.attention.v_proj.weight": "model-00002-of-00003.safetensors", "layers.21.layer_norm1.bias": "model-00002-of-00003.safetensors", "layers.21.layer_norm1.weight": "model-00002-of-00003.safetensors", "layers.21.layer_norm2.bias": "model-00002-of-00003.safetensors", "layers.21.layer_norm2.weight": "model-00002-of-00003.safetensors", "layers.21.mlp.0.bias": "model-00002-of-00003.safetensors", "layers.21.mlp.0.weight": "model-00002-of-00003.safetensors", "layers.21.mlp.2.bias": "model-00002-of-00003.safetensors", "layers.21.mlp.2.weight": "model-00002-of-00003.safetensors", "layers.22.attention.k_proj.bias": "model-00002-of-00003.safetensors", "layers.22.attention.k_proj.weight": "model-00002-of-00003.safetensors", "layers.22.attention.out_proj.bias": "model-00002-of-00003.safetensors", "layers.22.attention.out_proj.weight": "model-00002-of-00003.safetensors", "layers.22.attention.q_proj.bias": "model-00002-of-00003.safetensors", "layers.22.attention.q_proj.weight": "model-00002-of-00003.safetensors", "layers.22.attention.v_proj.bias": "model-00002-of-00003.safetensors", "layers.22.attention.v_proj.weight": "model-00002-of-00003.safetensors", "layers.22.layer_norm1.bias": "model-00002-of-00003.safetensors", "layers.22.layer_norm1.weight": "model-00002-of-00003.safetensors", "layers.22.layer_norm2.bias": "model-00002-of-00003.safetensors", "layers.22.layer_norm2.weight": "model-00002-of-00003.safetensors", "layers.22.mlp.0.bias": "model-00002-of-00003.safetensors", "layers.22.mlp.0.weight": "model-00002-of-00003.safetensors", "layers.22.mlp.2.bias": "model-00002-of-00003.safetensors", "layers.22.mlp.2.weight": "model-00002-of-00003.safetensors", "layers.23.attention.k_proj.bias": "model-00002-of-00003.safetensors", "layers.23.attention.k_proj.weight": "model-00002-of-00003.safetensors", "layers.23.attention.out_proj.bias": "model-00002-of-00003.safetensors", "layers.23.attention.out_proj.weight": "model-00002-of-00003.safetensors", "layers.23.attention.q_proj.bias": "model-00002-of-00003.safetensors", "layers.23.attention.q_proj.weight": "model-00002-of-00003.safetensors", "layers.23.attention.v_proj.bias": "model-00002-of-00003.safetensors", "layers.23.attention.v_proj.weight": "model-00002-of-00003.safetensors", "layers.23.layer_norm1.bias": "model-00002-of-00003.safetensors", "layers.23.layer_norm1.weight": "model-00002-of-00003.safetensors", "layers.23.layer_norm2.bias": "model-00002-of-00003.safetensors", "layers.23.layer_norm2.weight": "model-00002-of-00003.safetensors", "layers.23.mlp.0.bias": "model-00002-of-00003.safetensors", "layers.23.mlp.0.weight": "model-00002-of-00003.safetensors", "layers.23.mlp.2.bias": "model-00002-of-00003.safetensors", "layers.23.mlp.2.weight": "model-00002-of-00003.safetensors", "layers.24.attention.k_proj.bias": "model-00002-of-00003.safetensors", "layers.24.attention.k_proj.weight": "model-00002-of-00003.safetensors", "layers.24.attention.out_proj.bias": "model-00002-of-00003.safetensors", "layers.24.attention.out_proj.weight": "model-00002-of-00003.safetensors", "layers.24.attention.q_proj.bias": "model-00002-of-00003.safetensors", "layers.24.attention.q_proj.weight": "model-00002-of-00003.safetensors", "layers.24.attention.v_proj.bias": "model-00002-of-00003.safetensors", "layers.24.attention.v_proj.weight": "model-00002-of-00003.safetensors", "layers.24.layer_norm1.bias": "model-00002-of-00003.safetensors", "layers.24.layer_norm1.weight": "model-00002-of-00003.safetensors", "layers.24.layer_norm2.bias": "model-00002-of-00003.safetensors", "layers.24.layer_norm2.weight": "model-00002-of-00003.safetensors", "layers.24.mlp.0.bias": "model-00002-of-00003.safetensors", "layers.24.mlp.0.weight": "model-00002-of-00003.safetensors", "layers.24.mlp.2.bias": "model-00002-of-00003.safetensors", "layers.24.mlp.2.weight": "model-00002-of-00003.safetensors", "layers.25.attention.k_proj.bias": "model-00002-of-00003.safetensors", "layers.25.attention.k_proj.weight": "model-00002-of-00003.safetensors", "layers.25.attention.out_proj.bias": "model-00002-of-00003.safetensors", "layers.25.attention.out_proj.weight": "model-00002-of-00003.safetensors", "layers.25.attention.q_proj.bias": "model-00002-of-00003.safetensors", "layers.25.attention.q_proj.weight": "model-00002-of-00003.safetensors", "layers.25.attention.v_proj.bias": "model-00002-of-00003.safetensors", "layers.25.attention.v_proj.weight": "model-00002-of-00003.safetensors", "layers.25.layer_norm1.bias": "model-00002-of-00003.safetensors", "layers.25.layer_norm1.weight": "model-00002-of-00003.safetensors", "layers.25.layer_norm2.bias": "model-00002-of-00003.safetensors", "layers.25.layer_norm2.weight": "model-00002-of-00003.safetensors", "layers.25.mlp.0.bias": "model-00002-of-00003.safetensors", "layers.25.mlp.0.weight": "model-00002-of-00003.safetensors", "layers.25.mlp.2.bias": "model-00002-of-00003.safetensors", "layers.25.mlp.2.weight": "model-00002-of-00003.safetensors", "layers.26.attention.k_proj.bias": "model-00002-of-00003.safetensors", "layers.26.attention.k_proj.weight": "model-00002-of-00003.safetensors", "layers.26.attention.out_proj.bias": "model-00002-of-00003.safetensors", "layers.26.attention.out_proj.weight": "model-00002-of-00003.safetensors", "layers.26.attention.q_proj.bias": "model-00002-of-00003.safetensors", "layers.26.attention.q_proj.weight": "model-00002-of-00003.safetensors", "layers.26.attention.v_proj.bias": "model-00002-of-00003.safetensors", "layers.26.attention.v_proj.weight": "model-00002-of-00003.safetensors", "layers.26.layer_norm1.bias": "model-00002-of-00003.safetensors", "layers.26.layer_norm1.weight": "model-00002-of-00003.safetensors", "layers.26.layer_norm2.bias": "model-00002-of-00003.safetensors", "layers.26.layer_norm2.weight": "model-00002-of-00003.safetensors", "layers.26.mlp.0.bias": "model-00002-of-00003.safetensors", "layers.26.mlp.0.weight": "model-00002-of-00003.safetensors", "layers.26.mlp.2.bias": "model-00002-of-00003.safetensors", "layers.26.mlp.2.weight": "model-00002-of-00003.safetensors", "layers.27.attention.k_proj.bias": "model-00002-of-00003.safetensors", "layers.27.attention.k_proj.weight": "model-00002-of-00003.safetensors", "layers.27.attention.out_proj.bias": "model-00002-of-00003.safetensors", "layers.27.attention.out_proj.weight": "model-00002-of-00003.safetensors", "layers.27.attention.q_proj.bias": "model-00002-of-00003.safetensors", "layers.27.attention.q_proj.weight": "model-00002-of-00003.safetensors", "layers.27.attention.v_proj.bias": "model-00002-of-00003.safetensors", "layers.27.attention.v_proj.weight": "model-00002-of-00003.safetensors", "layers.27.layer_norm1.bias": "model-00002-of-00003.safetensors", "layers.27.layer_norm1.weight": "model-00002-of-00003.safetensors", "layers.27.layer_norm2.bias": "model-00002-of-00003.safetensors", "layers.27.layer_norm2.weight": "model-00002-of-00003.safetensors", "layers.27.mlp.0.bias": "model-00002-of-00003.safetensors", "layers.27.mlp.0.weight": "model-00002-of-00003.safetensors", "layers.27.mlp.2.bias": "model-00002-of-00003.safetensors", "layers.27.mlp.2.weight": "model-00002-of-00003.safetensors", "layers.28.attention.k_proj.bias": "model-00002-of-00003.safetensors", "layers.28.attention.k_proj.weight": "model-00002-of-00003.safetensors", "layers.28.attention.out_proj.bias": "model-00002-of-00003.safetensors", "layers.28.attention.out_proj.weight": "model-00002-of-00003.safetensors", "layers.28.attention.q_proj.bias": "model-00002-of-00003.safetensors", "layers.28.attention.q_proj.weight": "model-00002-of-00003.safetensors", "layers.28.attention.v_proj.bias": "model-00002-of-00003.safetensors", "layers.28.attention.v_proj.weight": "model-00002-of-00003.safetensors", "layers.28.layer_norm1.bias": "model-00002-of-00003.safetensors", "layers.28.layer_norm1.weight": "model-00002-of-00003.safetensors", "layers.28.layer_norm2.bias": "model-00002-of-00003.safetensors", "layers.28.layer_norm2.weight": "model-00002-of-00003.safetensors", "layers.28.mlp.0.bias": "model-00002-of-00003.safetensors", "layers.28.mlp.0.weight": "model-00002-of-00003.safetensors", "layers.28.mlp.2.bias": "model-00002-of-00003.safetensors", "layers.28.mlp.2.weight": "model-00002-of-00003.safetensors", "layers.29.attention.k_proj.bias": "model-00002-of-00003.safetensors", "layers.29.attention.k_proj.weight": "model-00002-of-00003.safetensors", "layers.29.attention.out_proj.bias": "model-00002-of-00003.safetensors", "layers.29.attention.out_proj.weight": "model-00002-of-00003.safetensors", "layers.29.attention.q_proj.bias": "model-00002-of-00003.safetensors", "layers.29.attention.q_proj.weight": "model-00002-of-00003.safetensors", "layers.29.attention.v_proj.bias": "model-00002-of-00003.safetensors", "layers.29.attention.v_proj.weight": "model-00002-of-00003.safetensors", "layers.29.layer_norm1.bias": "model-00002-of-00003.safetensors", "layers.29.layer_norm1.weight": "model-00002-of-00003.safetensors", "layers.29.layer_norm2.bias": "model-00002-of-00003.safetensors", "layers.29.layer_norm2.weight": "model-00002-of-00003.safetensors", "layers.29.mlp.0.bias": "model-00002-of-00003.safetensors", "layers.29.mlp.0.weight": "model-00002-of-00003.safetensors", "layers.29.mlp.2.bias": "model-00002-of-00003.safetensors", "layers.29.mlp.2.weight": "model-00002-of-00003.safetensors", "layers.3.attention.k_proj.bias": "model-00001-of-00003.safetensors", "layers.3.attention.k_proj.weight": "model-00001-of-00003.safetensors", "layers.3.attention.out_proj.bias": "model-00001-of-00003.safetensors", "layers.3.attention.out_proj.weight": "model-00001-of-00003.safetensors", "layers.3.attention.q_proj.bias": "model-00001-of-00003.safetensors", "layers.3.attention.q_proj.weight": "model-00001-of-00003.safetensors", "layers.3.attention.v_proj.bias": "model-00001-of-00003.safetensors", "layers.3.attention.v_proj.weight": "model-00001-of-00003.safetensors", "layers.3.layer_norm1.bias": "model-00001-of-00003.safetensors", "layers.3.layer_norm1.weight": "model-00001-of-00003.safetensors", "layers.3.layer_norm2.bias": "model-00001-of-00003.safetensors", "layers.3.layer_norm2.weight": "model-00001-of-00003.safetensors", "layers.3.mlp.0.bias": "model-00001-of-00003.safetensors", "layers.3.mlp.0.weight": "model-00001-of-00003.safetensors", "layers.3.mlp.2.bias": "model-00001-of-00003.safetensors", "layers.3.mlp.2.weight": "model-00001-of-00003.safetensors", "layers.4.attention.k_proj.bias": "model-00001-of-00003.safetensors", "layers.4.attention.k_proj.weight": "model-00001-of-00003.safetensors", "layers.4.attention.out_proj.bias": "model-00001-of-00003.safetensors", "layers.4.attention.out_proj.weight": "model-00001-of-00003.safetensors", "layers.4.attention.q_proj.bias": "model-00001-of-00003.safetensors", "layers.4.attention.q_proj.weight": "model-00001-of-00003.safetensors", "layers.4.attention.v_proj.bias": "model-00001-of-00003.safetensors", "layers.4.attention.v_proj.weight": "model-00001-of-00003.safetensors", "layers.4.layer_norm1.bias": "model-00001-of-00003.safetensors", "layers.4.layer_norm1.weight": "model-00001-of-00003.safetensors", "layers.4.layer_norm2.bias": "model-00001-of-00003.safetensors", "layers.4.layer_norm2.weight": "model-00001-of-00003.safetensors", "layers.4.mlp.0.bias": "model-00001-of-00003.safetensors", "layers.4.mlp.0.weight": "model-00001-of-00003.safetensors", "layers.4.mlp.2.bias": "model-00001-of-00003.safetensors", "layers.4.mlp.2.weight": "model-00001-of-00003.safetensors", "layers.5.attention.k_proj.bias": "model-00001-of-00003.safetensors", "layers.5.attention.k_proj.weight": "model-00001-of-00003.safetensors", "layers.5.attention.out_proj.bias": "model-00001-of-00003.safetensors", "layers.5.attention.out_proj.weight": "model-00001-of-00003.safetensors", "layers.5.attention.q_proj.bias": "model-00001-of-00003.safetensors", "layers.5.attention.q_proj.weight": "model-00001-of-00003.safetensors", "layers.5.attention.v_proj.bias": "model-00001-of-00003.safetensors", "layers.5.attention.v_proj.weight": "model-00001-of-00003.safetensors", "layers.5.layer_norm1.bias": "model-00001-of-00003.safetensors", "layers.5.layer_norm1.weight": "model-00001-of-00003.safetensors", "layers.5.layer_norm2.bias": "model-00001-of-00003.safetensors", "layers.5.layer_norm2.weight": "model-00001-of-00003.safetensors", "layers.5.mlp.0.bias": "model-00001-of-00003.safetensors", "layers.5.mlp.0.weight": "model-00001-of-00003.safetensors", "layers.5.mlp.2.bias": "model-00001-of-00003.safetensors", "layers.5.mlp.2.weight": "model-00001-of-00003.safetensors", "layers.6.attention.k_proj.bias": "model-00001-of-00003.safetensors", "layers.6.attention.k_proj.weight": "model-00001-of-00003.safetensors", "layers.6.attention.out_proj.bias": "model-00001-of-00003.safetensors", "layers.6.attention.out_proj.weight": "model-00001-of-00003.safetensors", "layers.6.attention.q_proj.bias": "model-00001-of-00003.safetensors", "layers.6.attention.q_proj.weight": "model-00001-of-00003.safetensors", "layers.6.attention.v_proj.bias": "model-00001-of-00003.safetensors", "layers.6.attention.v_proj.weight": "model-00001-of-00003.safetensors", "layers.6.layer_norm1.bias": "model-00001-of-00003.safetensors", "layers.6.layer_norm1.weight": "model-00001-of-00003.safetensors", "layers.6.layer_norm2.bias": "model-00001-of-00003.safetensors", "layers.6.layer_norm2.weight": "model-00001-of-00003.safetensors", "layers.6.mlp.0.bias": "model-00001-of-00003.safetensors", "layers.6.mlp.0.weight": "model-00001-of-00003.safetensors", "layers.6.mlp.2.bias": "model-00001-of-00003.safetensors", "layers.6.mlp.2.weight": "model-00001-of-00003.safetensors", "layers.7.attention.k_proj.bias": "model-00001-of-00003.safetensors", "layers.7.attention.k_proj.weight": "model-00001-of-00003.safetensors", "layers.7.attention.out_proj.bias": "model-00001-of-00003.safetensors", "layers.7.attention.out_proj.weight": "model-00001-of-00003.safetensors", "layers.7.attention.q_proj.bias": "model-00001-of-00003.safetensors", "layers.7.attention.q_proj.weight": "model-00001-of-00003.safetensors", "layers.7.attention.v_proj.bias": "model-00001-of-00003.safetensors", "layers.7.attention.v_proj.weight": "model-00001-of-00003.safetensors", "layers.7.layer_norm1.bias": "model-00001-of-00003.safetensors", "layers.7.layer_norm1.weight": "model-00001-of-00003.safetensors", "layers.7.layer_norm2.bias": "model-00001-of-00003.safetensors", "layers.7.layer_norm2.weight": "model-00001-of-00003.safetensors", "layers.7.mlp.0.bias": "model-00001-of-00003.safetensors", "layers.7.mlp.0.weight": "model-00001-of-00003.safetensors", "layers.7.mlp.2.bias": "model-00001-of-00003.safetensors", "layers.7.mlp.2.weight": "model-00001-of-00003.safetensors", "layers.8.attention.k_proj.bias": "model-00001-of-00003.safetensors", "layers.8.attention.k_proj.weight": "model-00001-of-00003.safetensors", "layers.8.attention.out_proj.bias": "model-00001-of-00003.safetensors", "layers.8.attention.out_proj.weight": "model-00001-of-00003.safetensors", "layers.8.attention.q_proj.bias": "model-00001-of-00003.safetensors", "layers.8.attention.q_proj.weight": "model-00001-of-00003.safetensors", "layers.8.attention.v_proj.bias": "model-00001-of-00003.safetensors", "layers.8.attention.v_proj.weight": "model-00001-of-00003.safetensors", "layers.8.layer_norm1.bias": "model-00001-of-00003.safetensors", "layers.8.layer_norm1.weight": "model-00001-of-00003.safetensors", "layers.8.layer_norm2.bias": "model-00001-of-00003.safetensors", "layers.8.layer_norm2.weight": "model-00001-of-00003.safetensors", "layers.8.mlp.0.bias": "model-00001-of-00003.safetensors", "layers.8.mlp.0.weight": "model-00001-of-00003.safetensors", "layers.8.mlp.2.bias": "model-00001-of-00003.safetensors", "layers.8.mlp.2.weight": "model-00001-of-00003.safetensors", "layers.9.attention.k_proj.bias": "model-00001-of-00003.safetensors", "layers.9.attention.k_proj.weight": "model-00001-of-00003.safetensors", "layers.9.attention.out_proj.bias": "model-00001-of-00003.safetensors", "layers.9.attention.out_proj.weight": "model-00001-of-00003.safetensors", "layers.9.attention.q_proj.bias": "model-00001-of-00003.safetensors", "layers.9.attention.q_proj.weight": "model-00001-of-00003.safetensors", "layers.9.attention.v_proj.bias": "model-00001-of-00003.safetensors", "layers.9.attention.v_proj.weight": "model-00001-of-00003.safetensors", "layers.9.layer_norm1.bias": "model-00001-of-00003.safetensors", "layers.9.layer_norm1.weight": "model-00001-of-00003.safetensors", "layers.9.layer_norm2.bias": "model-00001-of-00003.safetensors", "layers.9.layer_norm2.weight": "model-00001-of-00003.safetensors", "layers.9.mlp.0.bias": "model-00001-of-00003.safetensors", "layers.9.mlp.0.weight": "model-00001-of-00003.safetensors", "layers.9.mlp.2.bias": "model-00001-of-00003.safetensors", "layers.9.mlp.2.weight": "model-00001-of-00003.safetensors", "output_head.bias": "model-00003-of-00003.safetensors", "output_head.weight": "model-00003-of-00003.safetensors", "position_embedding.weight": "model-00001-of-00003.safetensors" } }