diff --git a/README.md b/README.md new file mode 100644 index 0000000000000000000000000000000000000000..bc5f30d6632ac0efdc7be2e9095e9e9579af2e33 --- /dev/null +++ b/README.md @@ -0,0 +1,199 @@ +--- +library_name: transformers +tags: [] +--- + +# Model Card for Model ID + + + + + +## Model Details + +### Model Description + + + +This is the model card of a 🤗 transformers model that has been pushed on the Hub. This model card has been automatically generated. + +- **Developed by:** [More Information Needed] +- **Funded by [optional]:** [More Information Needed] +- **Shared by [optional]:** [More Information Needed] +- **Model type:** [More Information Needed] +- **Language(s) (NLP):** [More Information Needed] +- **License:** [More Information Needed] +- **Finetuned from model [optional]:** [More Information Needed] + +### Model Sources [optional] + + + +- **Repository:** [More Information Needed] +- **Paper [optional]:** [More Information Needed] +- **Demo [optional]:** [More Information Needed] + +## Uses + + + +### Direct Use + + + +[More Information Needed] + +### Downstream Use [optional] + + + +[More Information Needed] + +### Out-of-Scope Use + + + +[More Information Needed] + +## Bias, Risks, and Limitations + + + +[More Information Needed] + +### Recommendations + + + +Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. + +## How to Get Started with the Model + +Use the code below to get started with the model. + +[More Information Needed] + +## Training Details + +### Training Data + + + +[More Information Needed] + +### Training Procedure + + + +#### Preprocessing [optional] + +[More Information Needed] + + +#### Training Hyperparameters + +- **Training regime:** [More Information Needed] + +#### Speeds, Sizes, Times [optional] + + + +[More Information Needed] + +## Evaluation + + + +### Testing Data, Factors & Metrics + +#### Testing Data + + + +[More Information Needed] + +#### Factors + + + +[More Information Needed] + +#### Metrics + + + +[More Information Needed] + +### Results + +[More Information Needed] + +#### Summary + + + +## Model Examination [optional] + + + +[More Information Needed] + +## Environmental Impact + + + +Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700). + +- **Hardware Type:** [More Information Needed] +- **Hours used:** [More Information Needed] +- **Cloud Provider:** [More Information Needed] +- **Compute Region:** [More Information Needed] +- **Carbon Emitted:** [More Information Needed] + +## Technical Specifications [optional] + +### Model Architecture and Objective + +[More Information Needed] + +### Compute Infrastructure + +[More Information Needed] + +#### Hardware + +[More Information Needed] + +#### Software + +[More Information Needed] + +## Citation [optional] + + + +**BibTeX:** + +[More Information Needed] + +**APA:** + +[More Information Needed] + +## Glossary [optional] + + + +[More Information Needed] + +## More Information [optional] + +[More Information Needed] + +## Model Card Authors [optional] + +[More Information Needed] + +## Model Card Contact + +[More Information Needed] \ No newline at end of file diff --git a/config.json b/config.json new file mode 100644 index 0000000000000000000000000000000000000000..99d596273185dd4b394e395f00b07c6a63d0ac30 --- /dev/null +++ b/config.json @@ -0,0 +1,34 @@ +{ + "_name_or_path": "./merged_model", + "architectures": [ + "Qwen2ForCausalLM" + ], + "attention_dropout": 0.0, + "bos_token_id": 151643, + "eos_token_id": 151645, + "hidden_act": "silu", + "hidden_size": 8192, + "initializer_range": 0.02, + "intermediate_size": 29568, + "max_position_embeddings": 32768, + "max_window_layers": 70, + "model_type": "qwen2", + "num_attention_heads": 64, + "num_hidden_layers": 80, + "num_key_value_heads": 8, + "rms_norm_eps": 1e-06, + "rope_scaling": { + "factor": 4.0, + "original_max_position_embeddings": 32768, + "rope_type": "yarn", + "type": "yarn" + }, + "rope_theta": 1000000.0, + "sliding_window": null, + "tie_word_embeddings": false, + "torch_dtype": "float32", + "transformers_version": "4.48.1", + "use_cache": false, + "use_sliding_window": false, + "vocab_size": 152064 +} diff --git a/generation_config.json b/generation_config.json new file mode 100644 index 0000000000000000000000000000000000000000..1c94a4b9390832bb6a871fffd24834b8c62bb8f5 --- /dev/null +++ b/generation_config.json @@ -0,0 +1,7 @@ +{ + "_from_model_config": true, + "bos_token_id": 151643, + "eos_token_id": 151645, + "transformers_version": "4.48.1", + "use_cache": false +} diff --git a/model-00001-of-00063.safetensors b/model-00001-of-00063.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..8ebd9bd67b2ca988e2bbb31170cb1f771fa91395 --- /dev/null +++ b/model-00001-of-00063.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:cb4385b6e07bf3f0f3609dee59fdddbd7d83b9c67a0a0e63b098bb0da2c4d4f6 +size 4982833288 diff --git a/model-00002-of-00063.safetensors b/model-00002-of-00063.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..5812eb3c4dc9c09a6c6688d5b8c4442bfef5f2d3 --- /dev/null +++ b/model-00002-of-00063.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:af218e25d6cab70afc3636a79b6949df0b89d8ce3aed106795487481cdeb24e6 +size 4114761856 diff --git a/model-00003-of-00063.safetensors b/model-00003-of-00063.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..9f6b915160ba8e7459282359b76df02ffac33fa6 --- /dev/null +++ b/model-00003-of-00063.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:a59cefea523b527e0056e0921fe5c4c6c013de4454e8ef8909457c8cc58020f5 +size 4479624664 diff --git a/model-00004-of-00063.safetensors b/model-00004-of-00063.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..53df95680a2ff9f01f32f6829e1b34082356fdbd --- /dev/null +++ b/model-00004-of-00063.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:d5bb2bbd129253c0c012ab575d48a255f9c3b25683833765b13d592df210ea93 +size 4479624664 diff --git a/model-00005-of-00063.safetensors b/model-00005-of-00063.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..b8943dea4a4b54d494467941b7feff5e8a2ef14f --- /dev/null +++ b/model-00005-of-00063.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:3f0969517477c9ba9439bd10000cf3b2032a09cd6714865731d366d8eb9c1815 +size 4815276384 diff --git a/model-00006-of-00063.safetensors b/model-00006-of-00063.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..452c9b102dcb2c9b117de98dc4d28b573a585218 --- /dev/null +++ b/model-00006-of-00063.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:ab0702de6dba4a54ce10ba443ace602e0e17db8c4b8d2f6b2e62583768ad05d8 +size 4748060240 diff --git a/model-00007-of-00063.safetensors b/model-00007-of-00063.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..be8a92a83c09a5168db28fd6a7d71711e7750ed8 --- /dev/null +++ b/model-00007-of-00063.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:eecd2e4898821b4e1a0bff0517ef89ba6f5c5a797f29a8327ab8b8b0cc7a7d40 +size 4479624664 diff --git a/model-00008-of-00063.safetensors b/model-00008-of-00063.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..a73d79d29b420d5a6820135c9097029d9c7ad5db --- /dev/null +++ b/model-00008-of-00063.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:15b405c35a34fb2ef247a5f5e1f214a404d5c97cdcf7befcb3b81eac43bcd24a +size 4815276384 diff --git a/model-00009-of-00063.safetensors b/model-00009-of-00063.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..dfdf37028ba0e64c324ee835ce55f9e28ba96bdf --- /dev/null +++ b/model-00009-of-00063.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:6a30ae5943526c31e47d68e1bffbbbf2144466bae7451e6578debb3578b5604f +size 4748060256 diff --git a/model-00010-of-00063.safetensors b/model-00010-of-00063.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..9df71f076ecb03893feb8a92b4c166f723667143 --- /dev/null +++ b/model-00010-of-00063.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:fd74838f24fc4d114c081ef610e42ab263a18cbc64fb2237eebfa25459b6ca59 +size 4479624672 diff --git a/model-00011-of-00063.safetensors b/model-00011-of-00063.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..8f4690a967e9a2131eb06de4669c182c04e1292c --- /dev/null +++ b/model-00011-of-00063.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:22d5c34d3b11d8d8aec81ea736be5c37ffeb5d865293f0f2f82847c18e5129fb +size 4815276400 diff --git a/model-00012-of-00063.safetensors b/model-00012-of-00063.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..343056a5bfbd06afe8894034ee1fe09d2b02b4ab --- /dev/null +++ b/model-00012-of-00063.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:34a87767044adcf0a38981c9075c2b44740cfbfb96037247e966ef19f463568d +size 4748060248 diff --git a/model-00013-of-00063.safetensors b/model-00013-of-00063.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..027ab507e0aca50aee923c9b75f8a97dda565102 --- /dev/null +++ b/model-00013-of-00063.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:2c4f6b93d3e622738a4376dd6cb2cf77485e817a5ab8fbd59f04b66288006ec2 +size 4479624672 diff --git a/model-00014-of-00063.safetensors b/model-00014-of-00063.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..144017e48ce14210084c52a2f1ee1a1cd04f795e --- /dev/null +++ b/model-00014-of-00063.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:0c8668efcdeb5da6954515a781e7c6ec0a0d7aef1e44980c1fc5babfa1dfc95a +size 4815276400 diff --git a/model-00015-of-00063.safetensors b/model-00015-of-00063.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..ad82f6d6907219a9638a8acac183a38e6bebdbd6 --- /dev/null +++ b/model-00015-of-00063.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:e447316a005ea4ef44177bfe53bbdc43b89a3637274e3fab06e5e6b97ef0204a +size 4748060248 diff --git a/model-00016-of-00063.safetensors b/model-00016-of-00063.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..bfb6fcd6de38aaf3bf9dbbe0e9f21d52b40be126 --- /dev/null +++ b/model-00016-of-00063.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:fab4b3e03a923dcac3b11160d3e80e79bb1008816f44dc17c48631c35787ecd8 +size 4479624672 diff --git a/model-00017-of-00063.safetensors b/model-00017-of-00063.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..a0fd92f09ee335d41f610245fff1f56d47db2659 --- /dev/null +++ b/model-00017-of-00063.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:4e0d837ff239ef71090bddb3418c57c663d65c03d055ca969dddfa4416231a81 +size 4815276400 diff --git a/model-00018-of-00063.safetensors b/model-00018-of-00063.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..0e71e28ef151c926291ac5fb990e987e07288952 --- /dev/null +++ b/model-00018-of-00063.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:d56229ff56167528bea201411bff5d8f74ff79da8fd01719ca6594be37f5fb45 +size 4748060248 diff --git a/model-00019-of-00063.safetensors b/model-00019-of-00063.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..a12d2ce8688021747dc0a940762278581cb9d7d8 --- /dev/null +++ b/model-00019-of-00063.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:e372c7f31a32ff8762f591718d390b0223ad1225aee04ca82051ddca615ed7bf +size 4479624672 diff --git a/model-00020-of-00063.safetensors b/model-00020-of-00063.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..3d47f6ee9fd1fc8a4fd3470c2f3ca73b601fccac --- /dev/null +++ b/model-00020-of-00063.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:4249c53a2a6be82852e727ca93badc329ad2981b2213daacd01aeb0739f93175 +size 4815276400 diff --git a/model-00021-of-00063.safetensors b/model-00021-of-00063.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..3080a883d0823af3d24c13d48f5be9cd5353efa1 --- /dev/null +++ b/model-00021-of-00063.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:e35ae8efde96d209e480b2456537c4777870d9025034e753c8a9f0712d358029 +size 4748060248 diff --git a/model-00022-of-00063.safetensors b/model-00022-of-00063.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..1a5199b24ee24a40200cb35088baeff1b231d54c --- /dev/null +++ b/model-00022-of-00063.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:e39d39faf6cc4d6ca9d087a28a67edd5e58891c30f27d3c9f6f6c4906a347a8c +size 4479624672 diff --git a/model-00023-of-00063.safetensors b/model-00023-of-00063.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..2dbcba8d5c402530d21855a15b2994eff44916a3 --- /dev/null +++ b/model-00023-of-00063.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:e73a97a704a0558c48d139611a6799ea2c419426039621514bfd85e120930120 +size 4815276400 diff --git a/model-00024-of-00063.safetensors b/model-00024-of-00063.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..14a63b5758acd8b6f5da43ce970699bbb97d50a6 --- /dev/null +++ b/model-00024-of-00063.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:d2315fb96eccf8d14732816cff63a4065b39b36d8b217343593420a11b2a1690 +size 4748060248 diff --git a/model-00025-of-00063.safetensors b/model-00025-of-00063.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..daf5fe43e89827d303ff025bab0ad96ae34a8fe0 --- /dev/null +++ b/model-00025-of-00063.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:8040b38a70cb8f1b353e2a049c2a5c1808ddc04951f01d2efee123e0aa86f7cc +size 4479624672 diff --git a/model-00026-of-00063.safetensors b/model-00026-of-00063.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..551cea23b019771f9a251ef06e82c1e3a2ee00b2 --- /dev/null +++ b/model-00026-of-00063.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:fd983718bcce29768a6513c25176b7d8ac08c65514758725c4c474967bcf192f +size 4815276400 diff --git a/model-00027-of-00063.safetensors b/model-00027-of-00063.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..ef805da586c4a6edef3d27f877b3f94437996452 --- /dev/null +++ b/model-00027-of-00063.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:c064275ec884ef047bb8ee472a9c8e077e1a7c3676981ae415d1c3bf5fec2856 +size 4748060248 diff --git a/model-00028-of-00063.safetensors b/model-00028-of-00063.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..65219bda0bb710fb40d7c6cbb262fa4f2290af07 --- /dev/null +++ b/model-00028-of-00063.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:02666817477d00ec2ecb55e4b3ceb696d18dd809d50429901d639f1c5f2e63d1 +size 4479624672 diff --git a/model-00029-of-00063.safetensors b/model-00029-of-00063.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..568963b7a66b8e2cc6dc464faa2b0ebc15411529 --- /dev/null +++ b/model-00029-of-00063.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:dded3cc7dda8f90d1b89f8474677ca958cc77995370f2c7b2c923113eff3c129 +size 4815276400 diff --git a/model-00030-of-00063.safetensors b/model-00030-of-00063.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..73948e6386567853395046976af62817e11d94ed --- /dev/null +++ b/model-00030-of-00063.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:bf9211abbf9b02bb0ae7756fcf6933e606aa0516305ea15f6318af68a489fdf8 +size 4748060248 diff --git a/model-00031-of-00063.safetensors b/model-00031-of-00063.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..5fceed342ba203073d767e2cab889752843ca6cc --- /dev/null +++ b/model-00031-of-00063.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:43dc3be4d04e54bb35a6bc6f645635a7041ba24aca034935c3fe0e3e065aeef8 +size 4479624672 diff --git a/model-00032-of-00063.safetensors b/model-00032-of-00063.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..09f7ea6319782b49d7f4b0a0770eb8031ba17c94 --- /dev/null +++ b/model-00032-of-00063.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:4db0f1094b60584274269bb948fb7f1b381ce9f13fe1b00e6d21274e483479fa +size 4815276400 diff --git a/model-00033-of-00063.safetensors b/model-00033-of-00063.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..2f2af30ec3c2b1fe500a50eaafe14057b4d67dcc --- /dev/null +++ b/model-00033-of-00063.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:c43b7232e3e62343ae2554b02c3c7750c3bd23e4917e019591c846e2235ed014 +size 4748060248 diff --git a/model-00034-of-00063.safetensors b/model-00034-of-00063.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..a7b05fba0e4543ef9413dc7a91d31999e1d7055f --- /dev/null +++ b/model-00034-of-00063.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:a25411349e6f3f5a59aade3c2a29f32e4b5325799e6e5691d940bb2eca68696c +size 4479624672 diff --git a/model-00035-of-00063.safetensors b/model-00035-of-00063.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..0c7979d0ef5ef9e02d51e2e81aca8b755fd893dd --- /dev/null +++ b/model-00035-of-00063.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:8a1997db5ef049fadcd7b438ec8291d012be120c8582f88139673951d942eb22 +size 4815276400 diff --git a/model-00036-of-00063.safetensors b/model-00036-of-00063.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..9c29836ab33a719f89993cbd14f30ad296957458 --- /dev/null +++ b/model-00036-of-00063.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:5d68e540e7d5f66b61567a9bc8b3ecd72eab73f6c436d71357dbe264416c93e3 +size 4748060248 diff --git a/model-00037-of-00063.safetensors b/model-00037-of-00063.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..613ba71328c476293066a3b51878378e8b6c9c2d --- /dev/null +++ b/model-00037-of-00063.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:1701e0c70d0614ebb7dd6aa61de01eb56182f0d43f68970e63e3fe50e486cb68 +size 4479624672 diff --git a/model-00038-of-00063.safetensors b/model-00038-of-00063.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..368047adad3ec2f96e32a9ec3139f7eaf6f6d1b9 --- /dev/null +++ b/model-00038-of-00063.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:ddff5a5b6cbcf655c195b8223d48225c66b28499b98de4ce740b49fcffef1f69 +size 4815276400 diff --git a/model-00039-of-00063.safetensors b/model-00039-of-00063.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..f32478efb7543cd3c0acef94c51d34f58cdcf789 --- /dev/null +++ b/model-00039-of-00063.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:c2c0e10d913a6c1e36d21fc66819473223597e39ce2e5be02ded7399691b9d3f +size 4748060248 diff --git a/model-00040-of-00063.safetensors b/model-00040-of-00063.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..e48c8eccac8b5d670d045496b54dbf2a13ecc1fe --- /dev/null +++ b/model-00040-of-00063.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:bea720beacc2af19c55392c1061f57e30466a94cd442866bc0278aa2642f1e87 +size 4479624672 diff --git a/model-00041-of-00063.safetensors b/model-00041-of-00063.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..179715beaea387c345d2e184c9aa687f69a1ac9e --- /dev/null +++ b/model-00041-of-00063.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:a833a8c2125c6c96c755a61e22cb8b8f8a2b6235e2f681c7fddb523cb2ef436b +size 4815276400 diff --git a/model-00042-of-00063.safetensors b/model-00042-of-00063.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..dead6221931c2375a33d961f6bc8b575f83f63a8 --- /dev/null +++ b/model-00042-of-00063.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:67f47beb58ccdd5393b89e77e8d7c901afd8972a300341dd10237bf67f11bdca +size 4748060248 diff --git a/model-00043-of-00063.safetensors b/model-00043-of-00063.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..4ab5d5d5f905cf14baf549f04dab886bc7547815 --- /dev/null +++ b/model-00043-of-00063.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:39a53c9055ab896fc366c7b201ef7d41ce890d76d7e58d8eafa074d1f7ed687d +size 4479624672 diff --git a/model-00044-of-00063.safetensors b/model-00044-of-00063.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..29e331c92970b8c1b07b1fabc49d59b1ce048bc7 --- /dev/null +++ b/model-00044-of-00063.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:b870188beef3b9c83ea9f11163135b1f508c0f43a84835f59a3afc99292e594b +size 4815276400 diff --git a/model-00045-of-00063.safetensors b/model-00045-of-00063.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..440c8d75f6622164c7898535d358caf41e5349bc --- /dev/null +++ b/model-00045-of-00063.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:bc1f094faf5545d07b83ffcdbba8d2d0c7e035f6e3b17a0d33fcc1e96e47d792 +size 4748060248 diff --git a/model-00046-of-00063.safetensors b/model-00046-of-00063.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..55bbb2baab238c3aad81e7ed938c30f7735045d6 --- /dev/null +++ b/model-00046-of-00063.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:6611f6ba06801b26337d03ba1536ffc47cf0d76374dcb18190065ddec96c8edf +size 4479624672 diff --git a/model-00047-of-00063.safetensors b/model-00047-of-00063.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..e32025beddf9114c1983995bc17a084e7ec79e3b --- /dev/null +++ b/model-00047-of-00063.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:6ad395d9a5aa31f10edb46e9456ddf21608bb5a8352072777ac13246e3bb04d2 +size 4815276400 diff --git a/model-00048-of-00063.safetensors b/model-00048-of-00063.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..5c4d02c92464bf2ad7fe705081879cfa462e2c8a --- /dev/null +++ b/model-00048-of-00063.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:7e70353dfe65bf186fafca8e6fc72eb46e586bf06f182146db3271060da1ca5e +size 4748060248 diff --git a/model-00049-of-00063.safetensors b/model-00049-of-00063.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..260fcc2b42440d3bc6828f6e972f4acdca4e51f2 --- /dev/null +++ b/model-00049-of-00063.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:b95e7023041a89c0ce4df5edb5d1f30dc5bf253ce2835b906de9732a3b797e08 +size 4479624672 diff --git a/model-00050-of-00063.safetensors b/model-00050-of-00063.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..6bcdc4d02c9ae3ea7d01c6a31f514ab7dad99c56 --- /dev/null +++ b/model-00050-of-00063.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:779d89a77f3560eca7287b968da2e0795ae35f2508518c2ad944613ad402a0be +size 4815276400 diff --git a/model-00051-of-00063.safetensors b/model-00051-of-00063.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..44a12aad944f01ea290f99d76b09e4e9b65c945f --- /dev/null +++ b/model-00051-of-00063.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:da22559b7fbad72ecee01f86c8d8feeb657eb37b157f706b6b99df973ada593b +size 4748060248 diff --git a/model-00052-of-00063.safetensors b/model-00052-of-00063.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..60088d2f512310fefbe4b3f1f3c8401bbe7424eb --- /dev/null +++ b/model-00052-of-00063.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:3799ac9e2040dbc9fb5775c230face5fb5ce78ca661fbcfde54debc95f869839 +size 4479624672 diff --git a/model-00053-of-00063.safetensors b/model-00053-of-00063.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..82fda1a0806fde62fdc402fc21ef90f7c06e5a2b --- /dev/null +++ b/model-00053-of-00063.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:c9f6750e388ccca5d0a5d46bf6b7946d885c9c07acfbd44e29ab596d89a8cbb5 +size 4815276400 diff --git a/model-00054-of-00063.safetensors b/model-00054-of-00063.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..6a504b88800f773a33591f1d548e2882e5d877eb --- /dev/null +++ b/model-00054-of-00063.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:aa82bb35200c0d036987dfa557830dffd0b2f3b9a7e1f6546d67e5fee8a1521e +size 4748060248 diff --git a/model-00055-of-00063.safetensors b/model-00055-of-00063.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..8603264a8b8b902596fd6f61c4358e5be0a83a01 --- /dev/null +++ b/model-00055-of-00063.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:07890d768a80f1b8bc3dd9735fc0c196ed33ac04d02541ae389e964207c32c99 +size 4479624672 diff --git a/model-00056-of-00063.safetensors b/model-00056-of-00063.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..f685c65a4d9427d908e2da29fbaca84ecb41a347 --- /dev/null +++ b/model-00056-of-00063.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:c8a736e2dbbcf71c1a18305541ea60e1231b0f8ba6cb32a37eb9620a5ff1c0ee +size 4815276400 diff --git a/model-00057-of-00063.safetensors b/model-00057-of-00063.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..9b7c095cd1db4e2e9e3723c6e4a95e3550043c7f --- /dev/null +++ b/model-00057-of-00063.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:73af43a4723e7a65f666790bcb04abd02ebf221ef0b6448b11b8e4611016e09c +size 4748060248 diff --git a/model-00058-of-00063.safetensors b/model-00058-of-00063.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..61acbaa7f030544ff3632d96acd058597b6f34b4 --- /dev/null +++ b/model-00058-of-00063.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:1213e70e667ebbba83ac592beb2672e7b66b6057f38032e9313142aa719de698 +size 4479624672 diff --git a/model-00059-of-00063.safetensors b/model-00059-of-00063.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..07b1fa9beaf0ab8da5e833c4d019072f93aa6f9d --- /dev/null +++ b/model-00059-of-00063.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:ffdaa541c724f7047cf92968e07ce797381e58658fa20138bc0af387210d244c +size 4815276400 diff --git a/model-00060-of-00063.safetensors b/model-00060-of-00063.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..42af8fd82fc1aed6be67e864bc341f9a778b36e8 --- /dev/null +++ b/model-00060-of-00063.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:418bbf330430c0fbc15f951e5715b1410d30f1832467a36c2207d0dacb4ba813 +size 4748060248 diff --git a/model-00061-of-00063.safetensors b/model-00061-of-00063.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..1471dd443cd0ac336ffb7f713f398ac8bc47a2c0 --- /dev/null +++ b/model-00061-of-00063.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:97248ec9207f94f8bb64f869ec7ec07aab45dce1620521568ca1c72bc14b88ef +size 4479624672 diff --git a/model-00062-of-00063.safetensors b/model-00062-of-00063.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..0e3a58711e7dfa01426617b39f32bf16e2ee84f1 --- /dev/null +++ b/model-00062-of-00063.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:d46612934dd7fa2eb3ae2cd39dcee4bcd5cc55d9e407062b7ad03fe983dadc3f +size 968982984 diff --git a/model-00063-of-00063.safetensors b/model-00063-of-00063.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..6512b6538af46d1570ea3667120b6eadcaf9a6d6 --- /dev/null +++ b/model-00063-of-00063.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:64a8ddb793032121464f68bef25b8fade0c53bbb2eb81c9e59f25770aa4bd013 +size 4982833280 diff --git a/model.safetensors.index.json b/model.safetensors.index.json new file mode 100644 index 0000000000000000000000000000000000000000..0d94ccf36cd25fbeb196bf9da00cb9a0debc6ef3 --- /dev/null +++ b/model.safetensors.index.json @@ -0,0 +1,970 @@ +{ + "metadata": { + "total_size": 290824814592 + }, + "weight_map": { + "lm_head.weight": "model-00063-of-00063.safetensors", + "model.embed_tokens.weight": "model-00001-of-00063.safetensors", + "model.layers.0.input_layernorm.weight": "model-00002-of-00063.safetensors", + "model.layers.0.mlp.down_proj.weight": "model-00002-of-00063.safetensors", + "model.layers.0.mlp.gate_proj.weight": "model-00002-of-00063.safetensors", + "model.layers.0.mlp.up_proj.weight": "model-00002-of-00063.safetensors", + "model.layers.0.post_attention_layernorm.weight": "model-00002-of-00063.safetensors", + "model.layers.0.self_attn.k_proj.bias": "model-00002-of-00063.safetensors", + "model.layers.0.self_attn.k_proj.weight": "model-00002-of-00063.safetensors", + "model.layers.0.self_attn.o_proj.weight": "model-00002-of-00063.safetensors", + "model.layers.0.self_attn.q_proj.bias": "model-00002-of-00063.safetensors", + "model.layers.0.self_attn.q_proj.weight": "model-00002-of-00063.safetensors", + "model.layers.0.self_attn.v_proj.bias": "model-00002-of-00063.safetensors", + "model.layers.0.self_attn.v_proj.weight": "model-00002-of-00063.safetensors", + "model.layers.1.input_layernorm.weight": "model-00003-of-00063.safetensors", + "model.layers.1.mlp.down_proj.weight": "model-00003-of-00063.safetensors", + "model.layers.1.mlp.gate_proj.weight": "model-00003-of-00063.safetensors", + "model.layers.1.mlp.up_proj.weight": "model-00003-of-00063.safetensors", + "model.layers.1.post_attention_layernorm.weight": "model-00003-of-00063.safetensors", + "model.layers.1.self_attn.k_proj.bias": "model-00002-of-00063.safetensors", + "model.layers.1.self_attn.k_proj.weight": "model-00002-of-00063.safetensors", + "model.layers.1.self_attn.o_proj.weight": "model-00002-of-00063.safetensors", + "model.layers.1.self_attn.q_proj.bias": "model-00002-of-00063.safetensors", + "model.layers.1.self_attn.q_proj.weight": "model-00002-of-00063.safetensors", + "model.layers.1.self_attn.v_proj.bias": "model-00002-of-00063.safetensors", + "model.layers.1.self_attn.v_proj.weight": "model-00002-of-00063.safetensors", + "model.layers.10.input_layernorm.weight": "model-00010-of-00063.safetensors", + "model.layers.10.mlp.down_proj.weight": "model-00010-of-00063.safetensors", + "model.layers.10.mlp.gate_proj.weight": "model-00009-of-00063.safetensors", + "model.layers.10.mlp.up_proj.weight": "model-00010-of-00063.safetensors", + "model.layers.10.post_attention_layernorm.weight": "model-00010-of-00063.safetensors", + "model.layers.10.self_attn.k_proj.bias": "model-00009-of-00063.safetensors", + "model.layers.10.self_attn.k_proj.weight": "model-00009-of-00063.safetensors", + "model.layers.10.self_attn.o_proj.weight": "model-00009-of-00063.safetensors", + "model.layers.10.self_attn.q_proj.bias": "model-00009-of-00063.safetensors", + "model.layers.10.self_attn.q_proj.weight": "model-00009-of-00063.safetensors", + "model.layers.10.self_attn.v_proj.bias": "model-00009-of-00063.safetensors", + "model.layers.10.self_attn.v_proj.weight": "model-00009-of-00063.safetensors", + "model.layers.11.input_layernorm.weight": "model-00011-of-00063.safetensors", + "model.layers.11.mlp.down_proj.weight": "model-00011-of-00063.safetensors", + "model.layers.11.mlp.gate_proj.weight": "model-00010-of-00063.safetensors", + "model.layers.11.mlp.up_proj.weight": "model-00010-of-00063.safetensors", + "model.layers.11.post_attention_layernorm.weight": "model-00011-of-00063.safetensors", + "model.layers.11.self_attn.k_proj.bias": "model-00010-of-00063.safetensors", + "model.layers.11.self_attn.k_proj.weight": "model-00010-of-00063.safetensors", + "model.layers.11.self_attn.o_proj.weight": "model-00010-of-00063.safetensors", + "model.layers.11.self_attn.q_proj.bias": "model-00010-of-00063.safetensors", + "model.layers.11.self_attn.q_proj.weight": "model-00010-of-00063.safetensors", + "model.layers.11.self_attn.v_proj.bias": "model-00010-of-00063.safetensors", + "model.layers.11.self_attn.v_proj.weight": "model-00010-of-00063.safetensors", + "model.layers.12.input_layernorm.weight": "model-00011-of-00063.safetensors", + "model.layers.12.mlp.down_proj.weight": "model-00011-of-00063.safetensors", + "model.layers.12.mlp.gate_proj.weight": "model-00011-of-00063.safetensors", + "model.layers.12.mlp.up_proj.weight": "model-00011-of-00063.safetensors", + "model.layers.12.post_attention_layernorm.weight": "model-00011-of-00063.safetensors", + "model.layers.12.self_attn.k_proj.bias": "model-00011-of-00063.safetensors", + "model.layers.12.self_attn.k_proj.weight": "model-00011-of-00063.safetensors", + "model.layers.12.self_attn.o_proj.weight": "model-00011-of-00063.safetensors", + "model.layers.12.self_attn.q_proj.bias": "model-00011-of-00063.safetensors", + "model.layers.12.self_attn.q_proj.weight": "model-00011-of-00063.safetensors", + "model.layers.12.self_attn.v_proj.bias": "model-00011-of-00063.safetensors", + "model.layers.12.self_attn.v_proj.weight": "model-00011-of-00063.safetensors", + "model.layers.13.input_layernorm.weight": "model-00012-of-00063.safetensors", + "model.layers.13.mlp.down_proj.weight": "model-00012-of-00063.safetensors", + "model.layers.13.mlp.gate_proj.weight": "model-00012-of-00063.safetensors", + "model.layers.13.mlp.up_proj.weight": "model-00012-of-00063.safetensors", + "model.layers.13.post_attention_layernorm.weight": "model-00012-of-00063.safetensors", + "model.layers.13.self_attn.k_proj.bias": "model-00011-of-00063.safetensors", + "model.layers.13.self_attn.k_proj.weight": "model-00011-of-00063.safetensors", + "model.layers.13.self_attn.o_proj.weight": "model-00012-of-00063.safetensors", + "model.layers.13.self_attn.q_proj.bias": "model-00011-of-00063.safetensors", + "model.layers.13.self_attn.q_proj.weight": "model-00011-of-00063.safetensors", + "model.layers.13.self_attn.v_proj.bias": "model-00011-of-00063.safetensors", + "model.layers.13.self_attn.v_proj.weight": "model-00011-of-00063.safetensors", + "model.layers.14.input_layernorm.weight": "model-00013-of-00063.safetensors", + "model.layers.14.mlp.down_proj.weight": "model-00013-of-00063.safetensors", + "model.layers.14.mlp.gate_proj.weight": "model-00012-of-00063.safetensors", + "model.layers.14.mlp.up_proj.weight": "model-00013-of-00063.safetensors", + "model.layers.14.post_attention_layernorm.weight": "model-00013-of-00063.safetensors", + "model.layers.14.self_attn.k_proj.bias": "model-00012-of-00063.safetensors", + "model.layers.14.self_attn.k_proj.weight": "model-00012-of-00063.safetensors", + "model.layers.14.self_attn.o_proj.weight": "model-00012-of-00063.safetensors", + "model.layers.14.self_attn.q_proj.bias": "model-00012-of-00063.safetensors", + "model.layers.14.self_attn.q_proj.weight": "model-00012-of-00063.safetensors", + "model.layers.14.self_attn.v_proj.bias": "model-00012-of-00063.safetensors", + "model.layers.14.self_attn.v_proj.weight": "model-00012-of-00063.safetensors", + "model.layers.15.input_layernorm.weight": "model-00014-of-00063.safetensors", + "model.layers.15.mlp.down_proj.weight": "model-00014-of-00063.safetensors", + "model.layers.15.mlp.gate_proj.weight": "model-00013-of-00063.safetensors", + "model.layers.15.mlp.up_proj.weight": "model-00013-of-00063.safetensors", + "model.layers.15.post_attention_layernorm.weight": "model-00014-of-00063.safetensors", + "model.layers.15.self_attn.k_proj.bias": "model-00013-of-00063.safetensors", + "model.layers.15.self_attn.k_proj.weight": "model-00013-of-00063.safetensors", + "model.layers.15.self_attn.o_proj.weight": "model-00013-of-00063.safetensors", + "model.layers.15.self_attn.q_proj.bias": "model-00013-of-00063.safetensors", + "model.layers.15.self_attn.q_proj.weight": "model-00013-of-00063.safetensors", + "model.layers.15.self_attn.v_proj.bias": "model-00013-of-00063.safetensors", + "model.layers.15.self_attn.v_proj.weight": "model-00013-of-00063.safetensors", + "model.layers.16.input_layernorm.weight": "model-00014-of-00063.safetensors", + "model.layers.16.mlp.down_proj.weight": "model-00014-of-00063.safetensors", + "model.layers.16.mlp.gate_proj.weight": "model-00014-of-00063.safetensors", + "model.layers.16.mlp.up_proj.weight": "model-00014-of-00063.safetensors", + "model.layers.16.post_attention_layernorm.weight": "model-00014-of-00063.safetensors", + "model.layers.16.self_attn.k_proj.bias": "model-00014-of-00063.safetensors", + "model.layers.16.self_attn.k_proj.weight": "model-00014-of-00063.safetensors", + "model.layers.16.self_attn.o_proj.weight": "model-00014-of-00063.safetensors", + "model.layers.16.self_attn.q_proj.bias": "model-00014-of-00063.safetensors", + "model.layers.16.self_attn.q_proj.weight": "model-00014-of-00063.safetensors", + "model.layers.16.self_attn.v_proj.bias": "model-00014-of-00063.safetensors", + "model.layers.16.self_attn.v_proj.weight": "model-00014-of-00063.safetensors", + "model.layers.17.input_layernorm.weight": "model-00015-of-00063.safetensors", + "model.layers.17.mlp.down_proj.weight": "model-00015-of-00063.safetensors", + "model.layers.17.mlp.gate_proj.weight": "model-00015-of-00063.safetensors", + "model.layers.17.mlp.up_proj.weight": "model-00015-of-00063.safetensors", + "model.layers.17.post_attention_layernorm.weight": "model-00015-of-00063.safetensors", + "model.layers.17.self_attn.k_proj.bias": "model-00014-of-00063.safetensors", + "model.layers.17.self_attn.k_proj.weight": "model-00014-of-00063.safetensors", + "model.layers.17.self_attn.o_proj.weight": "model-00015-of-00063.safetensors", + "model.layers.17.self_attn.q_proj.bias": "model-00014-of-00063.safetensors", + "model.layers.17.self_attn.q_proj.weight": "model-00014-of-00063.safetensors", + "model.layers.17.self_attn.v_proj.bias": "model-00014-of-00063.safetensors", + "model.layers.17.self_attn.v_proj.weight": "model-00014-of-00063.safetensors", + "model.layers.18.input_layernorm.weight": "model-00016-of-00063.safetensors", + "model.layers.18.mlp.down_proj.weight": "model-00016-of-00063.safetensors", + "model.layers.18.mlp.gate_proj.weight": "model-00015-of-00063.safetensors", + "model.layers.18.mlp.up_proj.weight": "model-00016-of-00063.safetensors", + "model.layers.18.post_attention_layernorm.weight": "model-00016-of-00063.safetensors", + "model.layers.18.self_attn.k_proj.bias": "model-00015-of-00063.safetensors", + "model.layers.18.self_attn.k_proj.weight": "model-00015-of-00063.safetensors", + "model.layers.18.self_attn.o_proj.weight": "model-00015-of-00063.safetensors", + "model.layers.18.self_attn.q_proj.bias": "model-00015-of-00063.safetensors", + "model.layers.18.self_attn.q_proj.weight": "model-00015-of-00063.safetensors", + "model.layers.18.self_attn.v_proj.bias": "model-00015-of-00063.safetensors", + "model.layers.18.self_attn.v_proj.weight": "model-00015-of-00063.safetensors", + "model.layers.19.input_layernorm.weight": "model-00017-of-00063.safetensors", + "model.layers.19.mlp.down_proj.weight": "model-00017-of-00063.safetensors", + "model.layers.19.mlp.gate_proj.weight": "model-00016-of-00063.safetensors", + "model.layers.19.mlp.up_proj.weight": "model-00016-of-00063.safetensors", + "model.layers.19.post_attention_layernorm.weight": "model-00017-of-00063.safetensors", + "model.layers.19.self_attn.k_proj.bias": "model-00016-of-00063.safetensors", + "model.layers.19.self_attn.k_proj.weight": "model-00016-of-00063.safetensors", + "model.layers.19.self_attn.o_proj.weight": "model-00016-of-00063.safetensors", + "model.layers.19.self_attn.q_proj.bias": "model-00016-of-00063.safetensors", + "model.layers.19.self_attn.q_proj.weight": "model-00016-of-00063.safetensors", + "model.layers.19.self_attn.v_proj.bias": "model-00016-of-00063.safetensors", + "model.layers.19.self_attn.v_proj.weight": "model-00016-of-00063.safetensors", + "model.layers.2.input_layernorm.weight": "model-00004-of-00063.safetensors", + "model.layers.2.mlp.down_proj.weight": "model-00004-of-00063.safetensors", + "model.layers.2.mlp.gate_proj.weight": "model-00003-of-00063.safetensors", + "model.layers.2.mlp.up_proj.weight": "model-00004-of-00063.safetensors", + "model.layers.2.post_attention_layernorm.weight": "model-00004-of-00063.safetensors", + "model.layers.2.self_attn.k_proj.bias": "model-00003-of-00063.safetensors", + "model.layers.2.self_attn.k_proj.weight": "model-00003-of-00063.safetensors", + "model.layers.2.self_attn.o_proj.weight": "model-00003-of-00063.safetensors", + "model.layers.2.self_attn.q_proj.bias": "model-00003-of-00063.safetensors", + "model.layers.2.self_attn.q_proj.weight": "model-00003-of-00063.safetensors", + "model.layers.2.self_attn.v_proj.bias": "model-00003-of-00063.safetensors", + "model.layers.2.self_attn.v_proj.weight": "model-00003-of-00063.safetensors", + "model.layers.20.input_layernorm.weight": "model-00017-of-00063.safetensors", + "model.layers.20.mlp.down_proj.weight": "model-00017-of-00063.safetensors", + "model.layers.20.mlp.gate_proj.weight": "model-00017-of-00063.safetensors", + "model.layers.20.mlp.up_proj.weight": "model-00017-of-00063.safetensors", + "model.layers.20.post_attention_layernorm.weight": "model-00017-of-00063.safetensors", + "model.layers.20.self_attn.k_proj.bias": "model-00017-of-00063.safetensors", + "model.layers.20.self_attn.k_proj.weight": "model-00017-of-00063.safetensors", + "model.layers.20.self_attn.o_proj.weight": "model-00017-of-00063.safetensors", + "model.layers.20.self_attn.q_proj.bias": "model-00017-of-00063.safetensors", + "model.layers.20.self_attn.q_proj.weight": "model-00017-of-00063.safetensors", + "model.layers.20.self_attn.v_proj.bias": "model-00017-of-00063.safetensors", + "model.layers.20.self_attn.v_proj.weight": "model-00017-of-00063.safetensors", + "model.layers.21.input_layernorm.weight": "model-00018-of-00063.safetensors", + "model.layers.21.mlp.down_proj.weight": "model-00018-of-00063.safetensors", + "model.layers.21.mlp.gate_proj.weight": "model-00018-of-00063.safetensors", + "model.layers.21.mlp.up_proj.weight": "model-00018-of-00063.safetensors", + "model.layers.21.post_attention_layernorm.weight": "model-00018-of-00063.safetensors", + "model.layers.21.self_attn.k_proj.bias": "model-00017-of-00063.safetensors", + "model.layers.21.self_attn.k_proj.weight": "model-00017-of-00063.safetensors", + "model.layers.21.self_attn.o_proj.weight": "model-00018-of-00063.safetensors", + "model.layers.21.self_attn.q_proj.bias": "model-00017-of-00063.safetensors", + "model.layers.21.self_attn.q_proj.weight": "model-00017-of-00063.safetensors", + "model.layers.21.self_attn.v_proj.bias": "model-00017-of-00063.safetensors", + "model.layers.21.self_attn.v_proj.weight": "model-00017-of-00063.safetensors", + "model.layers.22.input_layernorm.weight": "model-00019-of-00063.safetensors", + "model.layers.22.mlp.down_proj.weight": "model-00019-of-00063.safetensors", + "model.layers.22.mlp.gate_proj.weight": "model-00018-of-00063.safetensors", + "model.layers.22.mlp.up_proj.weight": "model-00019-of-00063.safetensors", + "model.layers.22.post_attention_layernorm.weight": "model-00019-of-00063.safetensors", + "model.layers.22.self_attn.k_proj.bias": "model-00018-of-00063.safetensors", + "model.layers.22.self_attn.k_proj.weight": "model-00018-of-00063.safetensors", + "model.layers.22.self_attn.o_proj.weight": "model-00018-of-00063.safetensors", + "model.layers.22.self_attn.q_proj.bias": "model-00018-of-00063.safetensors", + "model.layers.22.self_attn.q_proj.weight": "model-00018-of-00063.safetensors", + "model.layers.22.self_attn.v_proj.bias": "model-00018-of-00063.safetensors", + "model.layers.22.self_attn.v_proj.weight": "model-00018-of-00063.safetensors", + "model.layers.23.input_layernorm.weight": "model-00020-of-00063.safetensors", + "model.layers.23.mlp.down_proj.weight": "model-00020-of-00063.safetensors", + "model.layers.23.mlp.gate_proj.weight": "model-00019-of-00063.safetensors", + "model.layers.23.mlp.up_proj.weight": "model-00019-of-00063.safetensors", + "model.layers.23.post_attention_layernorm.weight": "model-00020-of-00063.safetensors", + "model.layers.23.self_attn.k_proj.bias": "model-00019-of-00063.safetensors", + "model.layers.23.self_attn.k_proj.weight": "model-00019-of-00063.safetensors", + "model.layers.23.self_attn.o_proj.weight": "model-00019-of-00063.safetensors", + "model.layers.23.self_attn.q_proj.bias": "model-00019-of-00063.safetensors", + "model.layers.23.self_attn.q_proj.weight": "model-00019-of-00063.safetensors", + "model.layers.23.self_attn.v_proj.bias": "model-00019-of-00063.safetensors", + "model.layers.23.self_attn.v_proj.weight": "model-00019-of-00063.safetensors", + "model.layers.24.input_layernorm.weight": "model-00020-of-00063.safetensors", + "model.layers.24.mlp.down_proj.weight": "model-00020-of-00063.safetensors", + "model.layers.24.mlp.gate_proj.weight": "model-00020-of-00063.safetensors", + "model.layers.24.mlp.up_proj.weight": "model-00020-of-00063.safetensors", + "model.layers.24.post_attention_layernorm.weight": "model-00020-of-00063.safetensors", + "model.layers.24.self_attn.k_proj.bias": "model-00020-of-00063.safetensors", + "model.layers.24.self_attn.k_proj.weight": "model-00020-of-00063.safetensors", + "model.layers.24.self_attn.o_proj.weight": "model-00020-of-00063.safetensors", + "model.layers.24.self_attn.q_proj.bias": "model-00020-of-00063.safetensors", + "model.layers.24.self_attn.q_proj.weight": "model-00020-of-00063.safetensors", + "model.layers.24.self_attn.v_proj.bias": "model-00020-of-00063.safetensors", + "model.layers.24.self_attn.v_proj.weight": "model-00020-of-00063.safetensors", + "model.layers.25.input_layernorm.weight": "model-00021-of-00063.safetensors", + "model.layers.25.mlp.down_proj.weight": "model-00021-of-00063.safetensors", + "model.layers.25.mlp.gate_proj.weight": "model-00021-of-00063.safetensors", + "model.layers.25.mlp.up_proj.weight": "model-00021-of-00063.safetensors", + "model.layers.25.post_attention_layernorm.weight": "model-00021-of-00063.safetensors", + "model.layers.25.self_attn.k_proj.bias": "model-00020-of-00063.safetensors", + "model.layers.25.self_attn.k_proj.weight": "model-00020-of-00063.safetensors", + "model.layers.25.self_attn.o_proj.weight": "model-00021-of-00063.safetensors", + "model.layers.25.self_attn.q_proj.bias": "model-00020-of-00063.safetensors", + "model.layers.25.self_attn.q_proj.weight": "model-00020-of-00063.safetensors", + "model.layers.25.self_attn.v_proj.bias": "model-00020-of-00063.safetensors", + "model.layers.25.self_attn.v_proj.weight": "model-00020-of-00063.safetensors", + "model.layers.26.input_layernorm.weight": "model-00022-of-00063.safetensors", + "model.layers.26.mlp.down_proj.weight": "model-00022-of-00063.safetensors", + "model.layers.26.mlp.gate_proj.weight": "model-00021-of-00063.safetensors", + "model.layers.26.mlp.up_proj.weight": "model-00022-of-00063.safetensors", + "model.layers.26.post_attention_layernorm.weight": "model-00022-of-00063.safetensors", + "model.layers.26.self_attn.k_proj.bias": "model-00021-of-00063.safetensors", + "model.layers.26.self_attn.k_proj.weight": "model-00021-of-00063.safetensors", + "model.layers.26.self_attn.o_proj.weight": "model-00021-of-00063.safetensors", + "model.layers.26.self_attn.q_proj.bias": "model-00021-of-00063.safetensors", + "model.layers.26.self_attn.q_proj.weight": "model-00021-of-00063.safetensors", + "model.layers.26.self_attn.v_proj.bias": "model-00021-of-00063.safetensors", + "model.layers.26.self_attn.v_proj.weight": "model-00021-of-00063.safetensors", + "model.layers.27.input_layernorm.weight": "model-00023-of-00063.safetensors", + "model.layers.27.mlp.down_proj.weight": "model-00023-of-00063.safetensors", + "model.layers.27.mlp.gate_proj.weight": "model-00022-of-00063.safetensors", + "model.layers.27.mlp.up_proj.weight": "model-00022-of-00063.safetensors", + "model.layers.27.post_attention_layernorm.weight": "model-00023-of-00063.safetensors", + "model.layers.27.self_attn.k_proj.bias": "model-00022-of-00063.safetensors", + "model.layers.27.self_attn.k_proj.weight": "model-00022-of-00063.safetensors", + "model.layers.27.self_attn.o_proj.weight": "model-00022-of-00063.safetensors", + "model.layers.27.self_attn.q_proj.bias": "model-00022-of-00063.safetensors", + "model.layers.27.self_attn.q_proj.weight": "model-00022-of-00063.safetensors", + "model.layers.27.self_attn.v_proj.bias": "model-00022-of-00063.safetensors", + "model.layers.27.self_attn.v_proj.weight": "model-00022-of-00063.safetensors", + "model.layers.28.input_layernorm.weight": "model-00023-of-00063.safetensors", + "model.layers.28.mlp.down_proj.weight": "model-00023-of-00063.safetensors", + "model.layers.28.mlp.gate_proj.weight": "model-00023-of-00063.safetensors", + "model.layers.28.mlp.up_proj.weight": "model-00023-of-00063.safetensors", + "model.layers.28.post_attention_layernorm.weight": "model-00023-of-00063.safetensors", + "model.layers.28.self_attn.k_proj.bias": "model-00023-of-00063.safetensors", + "model.layers.28.self_attn.k_proj.weight": "model-00023-of-00063.safetensors", + "model.layers.28.self_attn.o_proj.weight": "model-00023-of-00063.safetensors", + "model.layers.28.self_attn.q_proj.bias": "model-00023-of-00063.safetensors", + "model.layers.28.self_attn.q_proj.weight": "model-00023-of-00063.safetensors", + "model.layers.28.self_attn.v_proj.bias": "model-00023-of-00063.safetensors", + "model.layers.28.self_attn.v_proj.weight": "model-00023-of-00063.safetensors", + "model.layers.29.input_layernorm.weight": "model-00024-of-00063.safetensors", + "model.layers.29.mlp.down_proj.weight": "model-00024-of-00063.safetensors", + "model.layers.29.mlp.gate_proj.weight": "model-00024-of-00063.safetensors", + "model.layers.29.mlp.up_proj.weight": "model-00024-of-00063.safetensors", + "model.layers.29.post_attention_layernorm.weight": "model-00024-of-00063.safetensors", + "model.layers.29.self_attn.k_proj.bias": "model-00023-of-00063.safetensors", + "model.layers.29.self_attn.k_proj.weight": "model-00023-of-00063.safetensors", + "model.layers.29.self_attn.o_proj.weight": "model-00024-of-00063.safetensors", + "model.layers.29.self_attn.q_proj.bias": "model-00023-of-00063.safetensors", + "model.layers.29.self_attn.q_proj.weight": "model-00023-of-00063.safetensors", + "model.layers.29.self_attn.v_proj.bias": "model-00023-of-00063.safetensors", + "model.layers.29.self_attn.v_proj.weight": "model-00023-of-00063.safetensors", + "model.layers.3.input_layernorm.weight": "model-00005-of-00063.safetensors", + "model.layers.3.mlp.down_proj.weight": "model-00005-of-00063.safetensors", + "model.layers.3.mlp.gate_proj.weight": "model-00004-of-00063.safetensors", + "model.layers.3.mlp.up_proj.weight": "model-00004-of-00063.safetensors", + "model.layers.3.post_attention_layernorm.weight": "model-00005-of-00063.safetensors", + "model.layers.3.self_attn.k_proj.bias": "model-00004-of-00063.safetensors", + "model.layers.3.self_attn.k_proj.weight": "model-00004-of-00063.safetensors", + "model.layers.3.self_attn.o_proj.weight": "model-00004-of-00063.safetensors", + "model.layers.3.self_attn.q_proj.bias": "model-00004-of-00063.safetensors", + "model.layers.3.self_attn.q_proj.weight": "model-00004-of-00063.safetensors", + "model.layers.3.self_attn.v_proj.bias": "model-00004-of-00063.safetensors", + "model.layers.3.self_attn.v_proj.weight": "model-00004-of-00063.safetensors", + "model.layers.30.input_layernorm.weight": "model-00025-of-00063.safetensors", + "model.layers.30.mlp.down_proj.weight": "model-00025-of-00063.safetensors", + "model.layers.30.mlp.gate_proj.weight": "model-00024-of-00063.safetensors", + "model.layers.30.mlp.up_proj.weight": "model-00025-of-00063.safetensors", + "model.layers.30.post_attention_layernorm.weight": "model-00025-of-00063.safetensors", + "model.layers.30.self_attn.k_proj.bias": "model-00024-of-00063.safetensors", + "model.layers.30.self_attn.k_proj.weight": "model-00024-of-00063.safetensors", + "model.layers.30.self_attn.o_proj.weight": "model-00024-of-00063.safetensors", + "model.layers.30.self_attn.q_proj.bias": "model-00024-of-00063.safetensors", + "model.layers.30.self_attn.q_proj.weight": "model-00024-of-00063.safetensors", + "model.layers.30.self_attn.v_proj.bias": "model-00024-of-00063.safetensors", + "model.layers.30.self_attn.v_proj.weight": "model-00024-of-00063.safetensors", + "model.layers.31.input_layernorm.weight": "model-00026-of-00063.safetensors", + "model.layers.31.mlp.down_proj.weight": "model-00026-of-00063.safetensors", + "model.layers.31.mlp.gate_proj.weight": "model-00025-of-00063.safetensors", + "model.layers.31.mlp.up_proj.weight": "model-00025-of-00063.safetensors", + "model.layers.31.post_attention_layernorm.weight": "model-00026-of-00063.safetensors", + "model.layers.31.self_attn.k_proj.bias": "model-00025-of-00063.safetensors", + "model.layers.31.self_attn.k_proj.weight": "model-00025-of-00063.safetensors", + "model.layers.31.self_attn.o_proj.weight": "model-00025-of-00063.safetensors", + "model.layers.31.self_attn.q_proj.bias": "model-00025-of-00063.safetensors", + "model.layers.31.self_attn.q_proj.weight": "model-00025-of-00063.safetensors", + "model.layers.31.self_attn.v_proj.bias": "model-00025-of-00063.safetensors", + "model.layers.31.self_attn.v_proj.weight": "model-00025-of-00063.safetensors", + "model.layers.32.input_layernorm.weight": "model-00026-of-00063.safetensors", + "model.layers.32.mlp.down_proj.weight": "model-00026-of-00063.safetensors", + "model.layers.32.mlp.gate_proj.weight": "model-00026-of-00063.safetensors", + "model.layers.32.mlp.up_proj.weight": "model-00026-of-00063.safetensors", + "model.layers.32.post_attention_layernorm.weight": "model-00026-of-00063.safetensors", + "model.layers.32.self_attn.k_proj.bias": "model-00026-of-00063.safetensors", + "model.layers.32.self_attn.k_proj.weight": "model-00026-of-00063.safetensors", + "model.layers.32.self_attn.o_proj.weight": "model-00026-of-00063.safetensors", + "model.layers.32.self_attn.q_proj.bias": "model-00026-of-00063.safetensors", + "model.layers.32.self_attn.q_proj.weight": "model-00026-of-00063.safetensors", + "model.layers.32.self_attn.v_proj.bias": "model-00026-of-00063.safetensors", + "model.layers.32.self_attn.v_proj.weight": "model-00026-of-00063.safetensors", + "model.layers.33.input_layernorm.weight": "model-00027-of-00063.safetensors", + "model.layers.33.mlp.down_proj.weight": "model-00027-of-00063.safetensors", + "model.layers.33.mlp.gate_proj.weight": "model-00027-of-00063.safetensors", + "model.layers.33.mlp.up_proj.weight": "model-00027-of-00063.safetensors", + "model.layers.33.post_attention_layernorm.weight": "model-00027-of-00063.safetensors", + "model.layers.33.self_attn.k_proj.bias": "model-00026-of-00063.safetensors", + "model.layers.33.self_attn.k_proj.weight": "model-00026-of-00063.safetensors", + "model.layers.33.self_attn.o_proj.weight": "model-00027-of-00063.safetensors", + "model.layers.33.self_attn.q_proj.bias": "model-00026-of-00063.safetensors", + "model.layers.33.self_attn.q_proj.weight": "model-00026-of-00063.safetensors", + "model.layers.33.self_attn.v_proj.bias": "model-00026-of-00063.safetensors", + "model.layers.33.self_attn.v_proj.weight": "model-00026-of-00063.safetensors", + "model.layers.34.input_layernorm.weight": "model-00028-of-00063.safetensors", + "model.layers.34.mlp.down_proj.weight": "model-00028-of-00063.safetensors", + "model.layers.34.mlp.gate_proj.weight": "model-00027-of-00063.safetensors", + "model.layers.34.mlp.up_proj.weight": "model-00028-of-00063.safetensors", + "model.layers.34.post_attention_layernorm.weight": "model-00028-of-00063.safetensors", + "model.layers.34.self_attn.k_proj.bias": "model-00027-of-00063.safetensors", + "model.layers.34.self_attn.k_proj.weight": "model-00027-of-00063.safetensors", + "model.layers.34.self_attn.o_proj.weight": "model-00027-of-00063.safetensors", + "model.layers.34.self_attn.q_proj.bias": "model-00027-of-00063.safetensors", + "model.layers.34.self_attn.q_proj.weight": "model-00027-of-00063.safetensors", + "model.layers.34.self_attn.v_proj.bias": "model-00027-of-00063.safetensors", + "model.layers.34.self_attn.v_proj.weight": "model-00027-of-00063.safetensors", + "model.layers.35.input_layernorm.weight": "model-00029-of-00063.safetensors", + "model.layers.35.mlp.down_proj.weight": "model-00029-of-00063.safetensors", + "model.layers.35.mlp.gate_proj.weight": "model-00028-of-00063.safetensors", + "model.layers.35.mlp.up_proj.weight": "model-00028-of-00063.safetensors", + "model.layers.35.post_attention_layernorm.weight": "model-00029-of-00063.safetensors", + "model.layers.35.self_attn.k_proj.bias": "model-00028-of-00063.safetensors", + "model.layers.35.self_attn.k_proj.weight": "model-00028-of-00063.safetensors", + "model.layers.35.self_attn.o_proj.weight": "model-00028-of-00063.safetensors", + "model.layers.35.self_attn.q_proj.bias": "model-00028-of-00063.safetensors", + "model.layers.35.self_attn.q_proj.weight": "model-00028-of-00063.safetensors", + "model.layers.35.self_attn.v_proj.bias": "model-00028-of-00063.safetensors", + "model.layers.35.self_attn.v_proj.weight": "model-00028-of-00063.safetensors", + "model.layers.36.input_layernorm.weight": "model-00029-of-00063.safetensors", + "model.layers.36.mlp.down_proj.weight": "model-00029-of-00063.safetensors", + "model.layers.36.mlp.gate_proj.weight": "model-00029-of-00063.safetensors", + "model.layers.36.mlp.up_proj.weight": "model-00029-of-00063.safetensors", + "model.layers.36.post_attention_layernorm.weight": "model-00029-of-00063.safetensors", + "model.layers.36.self_attn.k_proj.bias": "model-00029-of-00063.safetensors", + "model.layers.36.self_attn.k_proj.weight": "model-00029-of-00063.safetensors", + "model.layers.36.self_attn.o_proj.weight": "model-00029-of-00063.safetensors", + "model.layers.36.self_attn.q_proj.bias": "model-00029-of-00063.safetensors", + "model.layers.36.self_attn.q_proj.weight": "model-00029-of-00063.safetensors", + "model.layers.36.self_attn.v_proj.bias": "model-00029-of-00063.safetensors", + "model.layers.36.self_attn.v_proj.weight": "model-00029-of-00063.safetensors", + "model.layers.37.input_layernorm.weight": "model-00030-of-00063.safetensors", + "model.layers.37.mlp.down_proj.weight": "model-00030-of-00063.safetensors", + "model.layers.37.mlp.gate_proj.weight": "model-00030-of-00063.safetensors", + "model.layers.37.mlp.up_proj.weight": "model-00030-of-00063.safetensors", + "model.layers.37.post_attention_layernorm.weight": "model-00030-of-00063.safetensors", + "model.layers.37.self_attn.k_proj.bias": "model-00029-of-00063.safetensors", + "model.layers.37.self_attn.k_proj.weight": "model-00029-of-00063.safetensors", + "model.layers.37.self_attn.o_proj.weight": "model-00030-of-00063.safetensors", + "model.layers.37.self_attn.q_proj.bias": "model-00029-of-00063.safetensors", + "model.layers.37.self_attn.q_proj.weight": "model-00029-of-00063.safetensors", + "model.layers.37.self_attn.v_proj.bias": "model-00029-of-00063.safetensors", + "model.layers.37.self_attn.v_proj.weight": "model-00029-of-00063.safetensors", + "model.layers.38.input_layernorm.weight": "model-00031-of-00063.safetensors", + "model.layers.38.mlp.down_proj.weight": "model-00031-of-00063.safetensors", + "model.layers.38.mlp.gate_proj.weight": "model-00030-of-00063.safetensors", + "model.layers.38.mlp.up_proj.weight": "model-00031-of-00063.safetensors", + "model.layers.38.post_attention_layernorm.weight": "model-00031-of-00063.safetensors", + "model.layers.38.self_attn.k_proj.bias": "model-00030-of-00063.safetensors", + "model.layers.38.self_attn.k_proj.weight": "model-00030-of-00063.safetensors", + "model.layers.38.self_attn.o_proj.weight": "model-00030-of-00063.safetensors", + "model.layers.38.self_attn.q_proj.bias": "model-00030-of-00063.safetensors", + "model.layers.38.self_attn.q_proj.weight": "model-00030-of-00063.safetensors", + "model.layers.38.self_attn.v_proj.bias": "model-00030-of-00063.safetensors", + "model.layers.38.self_attn.v_proj.weight": "model-00030-of-00063.safetensors", + "model.layers.39.input_layernorm.weight": "model-00032-of-00063.safetensors", + "model.layers.39.mlp.down_proj.weight": "model-00032-of-00063.safetensors", + "model.layers.39.mlp.gate_proj.weight": "model-00031-of-00063.safetensors", + "model.layers.39.mlp.up_proj.weight": "model-00031-of-00063.safetensors", + "model.layers.39.post_attention_layernorm.weight": "model-00032-of-00063.safetensors", + "model.layers.39.self_attn.k_proj.bias": "model-00031-of-00063.safetensors", + "model.layers.39.self_attn.k_proj.weight": "model-00031-of-00063.safetensors", + "model.layers.39.self_attn.o_proj.weight": "model-00031-of-00063.safetensors", + "model.layers.39.self_attn.q_proj.bias": "model-00031-of-00063.safetensors", + "model.layers.39.self_attn.q_proj.weight": "model-00031-of-00063.safetensors", + "model.layers.39.self_attn.v_proj.bias": "model-00031-of-00063.safetensors", + "model.layers.39.self_attn.v_proj.weight": "model-00031-of-00063.safetensors", + "model.layers.4.input_layernorm.weight": "model-00005-of-00063.safetensors", + "model.layers.4.mlp.down_proj.weight": "model-00005-of-00063.safetensors", + "model.layers.4.mlp.gate_proj.weight": "model-00005-of-00063.safetensors", + "model.layers.4.mlp.up_proj.weight": "model-00005-of-00063.safetensors", + "model.layers.4.post_attention_layernorm.weight": "model-00005-of-00063.safetensors", + "model.layers.4.self_attn.k_proj.bias": "model-00005-of-00063.safetensors", + "model.layers.4.self_attn.k_proj.weight": "model-00005-of-00063.safetensors", + "model.layers.4.self_attn.o_proj.weight": "model-00005-of-00063.safetensors", + "model.layers.4.self_attn.q_proj.bias": "model-00005-of-00063.safetensors", + "model.layers.4.self_attn.q_proj.weight": "model-00005-of-00063.safetensors", + "model.layers.4.self_attn.v_proj.bias": "model-00005-of-00063.safetensors", + "model.layers.4.self_attn.v_proj.weight": "model-00005-of-00063.safetensors", + "model.layers.40.input_layernorm.weight": "model-00032-of-00063.safetensors", + "model.layers.40.mlp.down_proj.weight": "model-00032-of-00063.safetensors", + "model.layers.40.mlp.gate_proj.weight": "model-00032-of-00063.safetensors", + "model.layers.40.mlp.up_proj.weight": "model-00032-of-00063.safetensors", + "model.layers.40.post_attention_layernorm.weight": "model-00032-of-00063.safetensors", + "model.layers.40.self_attn.k_proj.bias": "model-00032-of-00063.safetensors", + "model.layers.40.self_attn.k_proj.weight": "model-00032-of-00063.safetensors", + "model.layers.40.self_attn.o_proj.weight": "model-00032-of-00063.safetensors", + "model.layers.40.self_attn.q_proj.bias": "model-00032-of-00063.safetensors", + "model.layers.40.self_attn.q_proj.weight": "model-00032-of-00063.safetensors", + "model.layers.40.self_attn.v_proj.bias": "model-00032-of-00063.safetensors", + "model.layers.40.self_attn.v_proj.weight": "model-00032-of-00063.safetensors", + "model.layers.41.input_layernorm.weight": "model-00033-of-00063.safetensors", + "model.layers.41.mlp.down_proj.weight": "model-00033-of-00063.safetensors", + "model.layers.41.mlp.gate_proj.weight": "model-00033-of-00063.safetensors", + "model.layers.41.mlp.up_proj.weight": "model-00033-of-00063.safetensors", + "model.layers.41.post_attention_layernorm.weight": "model-00033-of-00063.safetensors", + "model.layers.41.self_attn.k_proj.bias": "model-00032-of-00063.safetensors", + "model.layers.41.self_attn.k_proj.weight": "model-00032-of-00063.safetensors", + "model.layers.41.self_attn.o_proj.weight": "model-00033-of-00063.safetensors", + "model.layers.41.self_attn.q_proj.bias": "model-00032-of-00063.safetensors", + "model.layers.41.self_attn.q_proj.weight": "model-00032-of-00063.safetensors", + "model.layers.41.self_attn.v_proj.bias": "model-00032-of-00063.safetensors", + "model.layers.41.self_attn.v_proj.weight": "model-00032-of-00063.safetensors", + "model.layers.42.input_layernorm.weight": "model-00034-of-00063.safetensors", + "model.layers.42.mlp.down_proj.weight": "model-00034-of-00063.safetensors", + "model.layers.42.mlp.gate_proj.weight": "model-00033-of-00063.safetensors", + "model.layers.42.mlp.up_proj.weight": "model-00034-of-00063.safetensors", + "model.layers.42.post_attention_layernorm.weight": "model-00034-of-00063.safetensors", + "model.layers.42.self_attn.k_proj.bias": "model-00033-of-00063.safetensors", + "model.layers.42.self_attn.k_proj.weight": "model-00033-of-00063.safetensors", + "model.layers.42.self_attn.o_proj.weight": "model-00033-of-00063.safetensors", + "model.layers.42.self_attn.q_proj.bias": "model-00033-of-00063.safetensors", + "model.layers.42.self_attn.q_proj.weight": "model-00033-of-00063.safetensors", + "model.layers.42.self_attn.v_proj.bias": "model-00033-of-00063.safetensors", + "model.layers.42.self_attn.v_proj.weight": "model-00033-of-00063.safetensors", + "model.layers.43.input_layernorm.weight": "model-00035-of-00063.safetensors", + "model.layers.43.mlp.down_proj.weight": "model-00035-of-00063.safetensors", + "model.layers.43.mlp.gate_proj.weight": "model-00034-of-00063.safetensors", + "model.layers.43.mlp.up_proj.weight": "model-00034-of-00063.safetensors", + "model.layers.43.post_attention_layernorm.weight": "model-00035-of-00063.safetensors", + "model.layers.43.self_attn.k_proj.bias": "model-00034-of-00063.safetensors", + "model.layers.43.self_attn.k_proj.weight": "model-00034-of-00063.safetensors", + "model.layers.43.self_attn.o_proj.weight": "model-00034-of-00063.safetensors", + "model.layers.43.self_attn.q_proj.bias": "model-00034-of-00063.safetensors", + "model.layers.43.self_attn.q_proj.weight": "model-00034-of-00063.safetensors", + "model.layers.43.self_attn.v_proj.bias": "model-00034-of-00063.safetensors", + "model.layers.43.self_attn.v_proj.weight": "model-00034-of-00063.safetensors", + "model.layers.44.input_layernorm.weight": "model-00035-of-00063.safetensors", + "model.layers.44.mlp.down_proj.weight": "model-00035-of-00063.safetensors", + "model.layers.44.mlp.gate_proj.weight": "model-00035-of-00063.safetensors", + "model.layers.44.mlp.up_proj.weight": "model-00035-of-00063.safetensors", + "model.layers.44.post_attention_layernorm.weight": "model-00035-of-00063.safetensors", + "model.layers.44.self_attn.k_proj.bias": "model-00035-of-00063.safetensors", + "model.layers.44.self_attn.k_proj.weight": "model-00035-of-00063.safetensors", + "model.layers.44.self_attn.o_proj.weight": "model-00035-of-00063.safetensors", + "model.layers.44.self_attn.q_proj.bias": "model-00035-of-00063.safetensors", + "model.layers.44.self_attn.q_proj.weight": "model-00035-of-00063.safetensors", + "model.layers.44.self_attn.v_proj.bias": "model-00035-of-00063.safetensors", + "model.layers.44.self_attn.v_proj.weight": "model-00035-of-00063.safetensors", + "model.layers.45.input_layernorm.weight": "model-00036-of-00063.safetensors", + "model.layers.45.mlp.down_proj.weight": "model-00036-of-00063.safetensors", + "model.layers.45.mlp.gate_proj.weight": "model-00036-of-00063.safetensors", + "model.layers.45.mlp.up_proj.weight": "model-00036-of-00063.safetensors", + "model.layers.45.post_attention_layernorm.weight": "model-00036-of-00063.safetensors", + "model.layers.45.self_attn.k_proj.bias": "model-00035-of-00063.safetensors", + "model.layers.45.self_attn.k_proj.weight": "model-00035-of-00063.safetensors", + "model.layers.45.self_attn.o_proj.weight": "model-00036-of-00063.safetensors", + "model.layers.45.self_attn.q_proj.bias": "model-00035-of-00063.safetensors", + "model.layers.45.self_attn.q_proj.weight": "model-00035-of-00063.safetensors", + "model.layers.45.self_attn.v_proj.bias": "model-00035-of-00063.safetensors", + "model.layers.45.self_attn.v_proj.weight": "model-00035-of-00063.safetensors", + "model.layers.46.input_layernorm.weight": "model-00037-of-00063.safetensors", + "model.layers.46.mlp.down_proj.weight": "model-00037-of-00063.safetensors", + "model.layers.46.mlp.gate_proj.weight": "model-00036-of-00063.safetensors", + "model.layers.46.mlp.up_proj.weight": "model-00037-of-00063.safetensors", + "model.layers.46.post_attention_layernorm.weight": "model-00037-of-00063.safetensors", + "model.layers.46.self_attn.k_proj.bias": "model-00036-of-00063.safetensors", + "model.layers.46.self_attn.k_proj.weight": "model-00036-of-00063.safetensors", + "model.layers.46.self_attn.o_proj.weight": "model-00036-of-00063.safetensors", + "model.layers.46.self_attn.q_proj.bias": "model-00036-of-00063.safetensors", + "model.layers.46.self_attn.q_proj.weight": "model-00036-of-00063.safetensors", + "model.layers.46.self_attn.v_proj.bias": "model-00036-of-00063.safetensors", + "model.layers.46.self_attn.v_proj.weight": "model-00036-of-00063.safetensors", + "model.layers.47.input_layernorm.weight": "model-00038-of-00063.safetensors", + "model.layers.47.mlp.down_proj.weight": "model-00038-of-00063.safetensors", + "model.layers.47.mlp.gate_proj.weight": "model-00037-of-00063.safetensors", + "model.layers.47.mlp.up_proj.weight": "model-00037-of-00063.safetensors", + "model.layers.47.post_attention_layernorm.weight": "model-00038-of-00063.safetensors", + "model.layers.47.self_attn.k_proj.bias": "model-00037-of-00063.safetensors", + "model.layers.47.self_attn.k_proj.weight": "model-00037-of-00063.safetensors", + "model.layers.47.self_attn.o_proj.weight": "model-00037-of-00063.safetensors", + "model.layers.47.self_attn.q_proj.bias": "model-00037-of-00063.safetensors", + "model.layers.47.self_attn.q_proj.weight": "model-00037-of-00063.safetensors", + "model.layers.47.self_attn.v_proj.bias": "model-00037-of-00063.safetensors", + "model.layers.47.self_attn.v_proj.weight": "model-00037-of-00063.safetensors", + "model.layers.48.input_layernorm.weight": "model-00038-of-00063.safetensors", + "model.layers.48.mlp.down_proj.weight": "model-00038-of-00063.safetensors", + "model.layers.48.mlp.gate_proj.weight": "model-00038-of-00063.safetensors", + "model.layers.48.mlp.up_proj.weight": "model-00038-of-00063.safetensors", + "model.layers.48.post_attention_layernorm.weight": "model-00038-of-00063.safetensors", + "model.layers.48.self_attn.k_proj.bias": "model-00038-of-00063.safetensors", + "model.layers.48.self_attn.k_proj.weight": "model-00038-of-00063.safetensors", + "model.layers.48.self_attn.o_proj.weight": "model-00038-of-00063.safetensors", + "model.layers.48.self_attn.q_proj.bias": "model-00038-of-00063.safetensors", + "model.layers.48.self_attn.q_proj.weight": "model-00038-of-00063.safetensors", + "model.layers.48.self_attn.v_proj.bias": "model-00038-of-00063.safetensors", + "model.layers.48.self_attn.v_proj.weight": "model-00038-of-00063.safetensors", + "model.layers.49.input_layernorm.weight": "model-00039-of-00063.safetensors", + "model.layers.49.mlp.down_proj.weight": "model-00039-of-00063.safetensors", + "model.layers.49.mlp.gate_proj.weight": "model-00039-of-00063.safetensors", + "model.layers.49.mlp.up_proj.weight": "model-00039-of-00063.safetensors", + "model.layers.49.post_attention_layernorm.weight": "model-00039-of-00063.safetensors", + "model.layers.49.self_attn.k_proj.bias": "model-00038-of-00063.safetensors", + "model.layers.49.self_attn.k_proj.weight": "model-00038-of-00063.safetensors", + "model.layers.49.self_attn.o_proj.weight": "model-00039-of-00063.safetensors", + "model.layers.49.self_attn.q_proj.bias": "model-00038-of-00063.safetensors", + "model.layers.49.self_attn.q_proj.weight": "model-00038-of-00063.safetensors", + "model.layers.49.self_attn.v_proj.bias": "model-00038-of-00063.safetensors", + "model.layers.49.self_attn.v_proj.weight": "model-00038-of-00063.safetensors", + "model.layers.5.input_layernorm.weight": "model-00006-of-00063.safetensors", + "model.layers.5.mlp.down_proj.weight": "model-00006-of-00063.safetensors", + "model.layers.5.mlp.gate_proj.weight": "model-00006-of-00063.safetensors", + "model.layers.5.mlp.up_proj.weight": "model-00006-of-00063.safetensors", + "model.layers.5.post_attention_layernorm.weight": "model-00006-of-00063.safetensors", + "model.layers.5.self_attn.k_proj.bias": "model-00005-of-00063.safetensors", + "model.layers.5.self_attn.k_proj.weight": "model-00005-of-00063.safetensors", + "model.layers.5.self_attn.o_proj.weight": "model-00006-of-00063.safetensors", + "model.layers.5.self_attn.q_proj.bias": "model-00005-of-00063.safetensors", + "model.layers.5.self_attn.q_proj.weight": "model-00005-of-00063.safetensors", + "model.layers.5.self_attn.v_proj.bias": "model-00005-of-00063.safetensors", + "model.layers.5.self_attn.v_proj.weight": "model-00005-of-00063.safetensors", + "model.layers.50.input_layernorm.weight": "model-00040-of-00063.safetensors", + "model.layers.50.mlp.down_proj.weight": "model-00040-of-00063.safetensors", + "model.layers.50.mlp.gate_proj.weight": "model-00039-of-00063.safetensors", + "model.layers.50.mlp.up_proj.weight": "model-00040-of-00063.safetensors", + "model.layers.50.post_attention_layernorm.weight": "model-00040-of-00063.safetensors", + "model.layers.50.self_attn.k_proj.bias": "model-00039-of-00063.safetensors", + "model.layers.50.self_attn.k_proj.weight": "model-00039-of-00063.safetensors", + "model.layers.50.self_attn.o_proj.weight": "model-00039-of-00063.safetensors", + "model.layers.50.self_attn.q_proj.bias": "model-00039-of-00063.safetensors", + "model.layers.50.self_attn.q_proj.weight": "model-00039-of-00063.safetensors", + "model.layers.50.self_attn.v_proj.bias": "model-00039-of-00063.safetensors", + "model.layers.50.self_attn.v_proj.weight": "model-00039-of-00063.safetensors", + "model.layers.51.input_layernorm.weight": "model-00041-of-00063.safetensors", + "model.layers.51.mlp.down_proj.weight": "model-00041-of-00063.safetensors", + "model.layers.51.mlp.gate_proj.weight": "model-00040-of-00063.safetensors", + "model.layers.51.mlp.up_proj.weight": "model-00040-of-00063.safetensors", + "model.layers.51.post_attention_layernorm.weight": "model-00041-of-00063.safetensors", + "model.layers.51.self_attn.k_proj.bias": "model-00040-of-00063.safetensors", + "model.layers.51.self_attn.k_proj.weight": "model-00040-of-00063.safetensors", + "model.layers.51.self_attn.o_proj.weight": "model-00040-of-00063.safetensors", + "model.layers.51.self_attn.q_proj.bias": "model-00040-of-00063.safetensors", + "model.layers.51.self_attn.q_proj.weight": "model-00040-of-00063.safetensors", + "model.layers.51.self_attn.v_proj.bias": "model-00040-of-00063.safetensors", + "model.layers.51.self_attn.v_proj.weight": "model-00040-of-00063.safetensors", + "model.layers.52.input_layernorm.weight": "model-00041-of-00063.safetensors", + "model.layers.52.mlp.down_proj.weight": "model-00041-of-00063.safetensors", + "model.layers.52.mlp.gate_proj.weight": "model-00041-of-00063.safetensors", + "model.layers.52.mlp.up_proj.weight": "model-00041-of-00063.safetensors", + "model.layers.52.post_attention_layernorm.weight": "model-00041-of-00063.safetensors", + "model.layers.52.self_attn.k_proj.bias": "model-00041-of-00063.safetensors", + "model.layers.52.self_attn.k_proj.weight": "model-00041-of-00063.safetensors", + "model.layers.52.self_attn.o_proj.weight": "model-00041-of-00063.safetensors", + "model.layers.52.self_attn.q_proj.bias": "model-00041-of-00063.safetensors", + "model.layers.52.self_attn.q_proj.weight": "model-00041-of-00063.safetensors", + "model.layers.52.self_attn.v_proj.bias": "model-00041-of-00063.safetensors", + "model.layers.52.self_attn.v_proj.weight": "model-00041-of-00063.safetensors", + "model.layers.53.input_layernorm.weight": "model-00042-of-00063.safetensors", + "model.layers.53.mlp.down_proj.weight": "model-00042-of-00063.safetensors", + "model.layers.53.mlp.gate_proj.weight": "model-00042-of-00063.safetensors", + "model.layers.53.mlp.up_proj.weight": "model-00042-of-00063.safetensors", + "model.layers.53.post_attention_layernorm.weight": "model-00042-of-00063.safetensors", + "model.layers.53.self_attn.k_proj.bias": "model-00041-of-00063.safetensors", + "model.layers.53.self_attn.k_proj.weight": "model-00041-of-00063.safetensors", + "model.layers.53.self_attn.o_proj.weight": "model-00042-of-00063.safetensors", + "model.layers.53.self_attn.q_proj.bias": "model-00041-of-00063.safetensors", + "model.layers.53.self_attn.q_proj.weight": "model-00041-of-00063.safetensors", + "model.layers.53.self_attn.v_proj.bias": "model-00041-of-00063.safetensors", + "model.layers.53.self_attn.v_proj.weight": "model-00041-of-00063.safetensors", + "model.layers.54.input_layernorm.weight": "model-00043-of-00063.safetensors", + "model.layers.54.mlp.down_proj.weight": "model-00043-of-00063.safetensors", + "model.layers.54.mlp.gate_proj.weight": "model-00042-of-00063.safetensors", + "model.layers.54.mlp.up_proj.weight": "model-00043-of-00063.safetensors", + "model.layers.54.post_attention_layernorm.weight": "model-00043-of-00063.safetensors", + "model.layers.54.self_attn.k_proj.bias": "model-00042-of-00063.safetensors", + "model.layers.54.self_attn.k_proj.weight": "model-00042-of-00063.safetensors", + "model.layers.54.self_attn.o_proj.weight": "model-00042-of-00063.safetensors", + "model.layers.54.self_attn.q_proj.bias": "model-00042-of-00063.safetensors", + "model.layers.54.self_attn.q_proj.weight": "model-00042-of-00063.safetensors", + "model.layers.54.self_attn.v_proj.bias": "model-00042-of-00063.safetensors", + "model.layers.54.self_attn.v_proj.weight": "model-00042-of-00063.safetensors", + "model.layers.55.input_layernorm.weight": "model-00044-of-00063.safetensors", + "model.layers.55.mlp.down_proj.weight": "model-00044-of-00063.safetensors", + "model.layers.55.mlp.gate_proj.weight": "model-00043-of-00063.safetensors", + "model.layers.55.mlp.up_proj.weight": "model-00043-of-00063.safetensors", + "model.layers.55.post_attention_layernorm.weight": "model-00044-of-00063.safetensors", + "model.layers.55.self_attn.k_proj.bias": "model-00043-of-00063.safetensors", + "model.layers.55.self_attn.k_proj.weight": "model-00043-of-00063.safetensors", + "model.layers.55.self_attn.o_proj.weight": "model-00043-of-00063.safetensors", + "model.layers.55.self_attn.q_proj.bias": "model-00043-of-00063.safetensors", + "model.layers.55.self_attn.q_proj.weight": "model-00043-of-00063.safetensors", + "model.layers.55.self_attn.v_proj.bias": "model-00043-of-00063.safetensors", + "model.layers.55.self_attn.v_proj.weight": "model-00043-of-00063.safetensors", + "model.layers.56.input_layernorm.weight": "model-00044-of-00063.safetensors", + "model.layers.56.mlp.down_proj.weight": "model-00044-of-00063.safetensors", + "model.layers.56.mlp.gate_proj.weight": "model-00044-of-00063.safetensors", + "model.layers.56.mlp.up_proj.weight": "model-00044-of-00063.safetensors", + "model.layers.56.post_attention_layernorm.weight": "model-00044-of-00063.safetensors", + "model.layers.56.self_attn.k_proj.bias": "model-00044-of-00063.safetensors", + "model.layers.56.self_attn.k_proj.weight": "model-00044-of-00063.safetensors", + "model.layers.56.self_attn.o_proj.weight": "model-00044-of-00063.safetensors", + "model.layers.56.self_attn.q_proj.bias": "model-00044-of-00063.safetensors", + "model.layers.56.self_attn.q_proj.weight": "model-00044-of-00063.safetensors", + "model.layers.56.self_attn.v_proj.bias": "model-00044-of-00063.safetensors", + "model.layers.56.self_attn.v_proj.weight": "model-00044-of-00063.safetensors", + "model.layers.57.input_layernorm.weight": "model-00045-of-00063.safetensors", + "model.layers.57.mlp.down_proj.weight": "model-00045-of-00063.safetensors", + "model.layers.57.mlp.gate_proj.weight": "model-00045-of-00063.safetensors", + "model.layers.57.mlp.up_proj.weight": "model-00045-of-00063.safetensors", + "model.layers.57.post_attention_layernorm.weight": "model-00045-of-00063.safetensors", + "model.layers.57.self_attn.k_proj.bias": "model-00044-of-00063.safetensors", + "model.layers.57.self_attn.k_proj.weight": "model-00044-of-00063.safetensors", + "model.layers.57.self_attn.o_proj.weight": "model-00045-of-00063.safetensors", + "model.layers.57.self_attn.q_proj.bias": "model-00044-of-00063.safetensors", + "model.layers.57.self_attn.q_proj.weight": "model-00044-of-00063.safetensors", + "model.layers.57.self_attn.v_proj.bias": "model-00044-of-00063.safetensors", + "model.layers.57.self_attn.v_proj.weight": "model-00044-of-00063.safetensors", + "model.layers.58.input_layernorm.weight": "model-00046-of-00063.safetensors", + "model.layers.58.mlp.down_proj.weight": "model-00046-of-00063.safetensors", + "model.layers.58.mlp.gate_proj.weight": "model-00045-of-00063.safetensors", + "model.layers.58.mlp.up_proj.weight": "model-00046-of-00063.safetensors", + "model.layers.58.post_attention_layernorm.weight": "model-00046-of-00063.safetensors", + "model.layers.58.self_attn.k_proj.bias": "model-00045-of-00063.safetensors", + "model.layers.58.self_attn.k_proj.weight": "model-00045-of-00063.safetensors", + "model.layers.58.self_attn.o_proj.weight": "model-00045-of-00063.safetensors", + "model.layers.58.self_attn.q_proj.bias": "model-00045-of-00063.safetensors", + "model.layers.58.self_attn.q_proj.weight": "model-00045-of-00063.safetensors", + "model.layers.58.self_attn.v_proj.bias": "model-00045-of-00063.safetensors", + "model.layers.58.self_attn.v_proj.weight": "model-00045-of-00063.safetensors", + "model.layers.59.input_layernorm.weight": "model-00047-of-00063.safetensors", + "model.layers.59.mlp.down_proj.weight": "model-00047-of-00063.safetensors", + "model.layers.59.mlp.gate_proj.weight": "model-00046-of-00063.safetensors", + "model.layers.59.mlp.up_proj.weight": "model-00046-of-00063.safetensors", + "model.layers.59.post_attention_layernorm.weight": "model-00047-of-00063.safetensors", + "model.layers.59.self_attn.k_proj.bias": "model-00046-of-00063.safetensors", + "model.layers.59.self_attn.k_proj.weight": "model-00046-of-00063.safetensors", + "model.layers.59.self_attn.o_proj.weight": "model-00046-of-00063.safetensors", + "model.layers.59.self_attn.q_proj.bias": "model-00046-of-00063.safetensors", + "model.layers.59.self_attn.q_proj.weight": "model-00046-of-00063.safetensors", + "model.layers.59.self_attn.v_proj.bias": "model-00046-of-00063.safetensors", + "model.layers.59.self_attn.v_proj.weight": "model-00046-of-00063.safetensors", + "model.layers.6.input_layernorm.weight": "model-00007-of-00063.safetensors", + "model.layers.6.mlp.down_proj.weight": "model-00007-of-00063.safetensors", + "model.layers.6.mlp.gate_proj.weight": "model-00006-of-00063.safetensors", + "model.layers.6.mlp.up_proj.weight": "model-00007-of-00063.safetensors", + "model.layers.6.post_attention_layernorm.weight": "model-00007-of-00063.safetensors", + "model.layers.6.self_attn.k_proj.bias": "model-00006-of-00063.safetensors", + "model.layers.6.self_attn.k_proj.weight": "model-00006-of-00063.safetensors", + "model.layers.6.self_attn.o_proj.weight": "model-00006-of-00063.safetensors", + "model.layers.6.self_attn.q_proj.bias": "model-00006-of-00063.safetensors", + "model.layers.6.self_attn.q_proj.weight": "model-00006-of-00063.safetensors", + "model.layers.6.self_attn.v_proj.bias": "model-00006-of-00063.safetensors", + "model.layers.6.self_attn.v_proj.weight": "model-00006-of-00063.safetensors", + "model.layers.60.input_layernorm.weight": "model-00047-of-00063.safetensors", + "model.layers.60.mlp.down_proj.weight": "model-00047-of-00063.safetensors", + "model.layers.60.mlp.gate_proj.weight": "model-00047-of-00063.safetensors", + "model.layers.60.mlp.up_proj.weight": "model-00047-of-00063.safetensors", + "model.layers.60.post_attention_layernorm.weight": "model-00047-of-00063.safetensors", + "model.layers.60.self_attn.k_proj.bias": "model-00047-of-00063.safetensors", + "model.layers.60.self_attn.k_proj.weight": "model-00047-of-00063.safetensors", + "model.layers.60.self_attn.o_proj.weight": "model-00047-of-00063.safetensors", + "model.layers.60.self_attn.q_proj.bias": "model-00047-of-00063.safetensors", + "model.layers.60.self_attn.q_proj.weight": "model-00047-of-00063.safetensors", + "model.layers.60.self_attn.v_proj.bias": "model-00047-of-00063.safetensors", + "model.layers.60.self_attn.v_proj.weight": "model-00047-of-00063.safetensors", + "model.layers.61.input_layernorm.weight": "model-00048-of-00063.safetensors", + "model.layers.61.mlp.down_proj.weight": "model-00048-of-00063.safetensors", + "model.layers.61.mlp.gate_proj.weight": "model-00048-of-00063.safetensors", + "model.layers.61.mlp.up_proj.weight": "model-00048-of-00063.safetensors", + "model.layers.61.post_attention_layernorm.weight": "model-00048-of-00063.safetensors", + "model.layers.61.self_attn.k_proj.bias": "model-00047-of-00063.safetensors", + "model.layers.61.self_attn.k_proj.weight": "model-00047-of-00063.safetensors", + "model.layers.61.self_attn.o_proj.weight": "model-00048-of-00063.safetensors", + "model.layers.61.self_attn.q_proj.bias": "model-00047-of-00063.safetensors", + "model.layers.61.self_attn.q_proj.weight": "model-00047-of-00063.safetensors", + "model.layers.61.self_attn.v_proj.bias": "model-00047-of-00063.safetensors", + "model.layers.61.self_attn.v_proj.weight": "model-00047-of-00063.safetensors", + "model.layers.62.input_layernorm.weight": "model-00049-of-00063.safetensors", + "model.layers.62.mlp.down_proj.weight": "model-00049-of-00063.safetensors", + "model.layers.62.mlp.gate_proj.weight": "model-00048-of-00063.safetensors", + "model.layers.62.mlp.up_proj.weight": "model-00049-of-00063.safetensors", + "model.layers.62.post_attention_layernorm.weight": "model-00049-of-00063.safetensors", + "model.layers.62.self_attn.k_proj.bias": "model-00048-of-00063.safetensors", + "model.layers.62.self_attn.k_proj.weight": "model-00048-of-00063.safetensors", + "model.layers.62.self_attn.o_proj.weight": "model-00048-of-00063.safetensors", + "model.layers.62.self_attn.q_proj.bias": "model-00048-of-00063.safetensors", + "model.layers.62.self_attn.q_proj.weight": "model-00048-of-00063.safetensors", + "model.layers.62.self_attn.v_proj.bias": "model-00048-of-00063.safetensors", + "model.layers.62.self_attn.v_proj.weight": "model-00048-of-00063.safetensors", + "model.layers.63.input_layernorm.weight": "model-00050-of-00063.safetensors", + "model.layers.63.mlp.down_proj.weight": "model-00050-of-00063.safetensors", + "model.layers.63.mlp.gate_proj.weight": "model-00049-of-00063.safetensors", + "model.layers.63.mlp.up_proj.weight": "model-00049-of-00063.safetensors", + "model.layers.63.post_attention_layernorm.weight": "model-00050-of-00063.safetensors", + "model.layers.63.self_attn.k_proj.bias": "model-00049-of-00063.safetensors", + "model.layers.63.self_attn.k_proj.weight": "model-00049-of-00063.safetensors", + "model.layers.63.self_attn.o_proj.weight": "model-00049-of-00063.safetensors", + "model.layers.63.self_attn.q_proj.bias": "model-00049-of-00063.safetensors", + "model.layers.63.self_attn.q_proj.weight": "model-00049-of-00063.safetensors", + "model.layers.63.self_attn.v_proj.bias": "model-00049-of-00063.safetensors", + "model.layers.63.self_attn.v_proj.weight": "model-00049-of-00063.safetensors", + "model.layers.64.input_layernorm.weight": "model-00050-of-00063.safetensors", + "model.layers.64.mlp.down_proj.weight": "model-00050-of-00063.safetensors", + "model.layers.64.mlp.gate_proj.weight": "model-00050-of-00063.safetensors", + "model.layers.64.mlp.up_proj.weight": "model-00050-of-00063.safetensors", + "model.layers.64.post_attention_layernorm.weight": "model-00050-of-00063.safetensors", + "model.layers.64.self_attn.k_proj.bias": "model-00050-of-00063.safetensors", + "model.layers.64.self_attn.k_proj.weight": "model-00050-of-00063.safetensors", + "model.layers.64.self_attn.o_proj.weight": "model-00050-of-00063.safetensors", + "model.layers.64.self_attn.q_proj.bias": "model-00050-of-00063.safetensors", + "model.layers.64.self_attn.q_proj.weight": "model-00050-of-00063.safetensors", + "model.layers.64.self_attn.v_proj.bias": "model-00050-of-00063.safetensors", + "model.layers.64.self_attn.v_proj.weight": "model-00050-of-00063.safetensors", + "model.layers.65.input_layernorm.weight": "model-00051-of-00063.safetensors", + "model.layers.65.mlp.down_proj.weight": "model-00051-of-00063.safetensors", + "model.layers.65.mlp.gate_proj.weight": "model-00051-of-00063.safetensors", + "model.layers.65.mlp.up_proj.weight": "model-00051-of-00063.safetensors", + "model.layers.65.post_attention_layernorm.weight": "model-00051-of-00063.safetensors", + "model.layers.65.self_attn.k_proj.bias": "model-00050-of-00063.safetensors", + "model.layers.65.self_attn.k_proj.weight": "model-00050-of-00063.safetensors", + "model.layers.65.self_attn.o_proj.weight": "model-00051-of-00063.safetensors", + "model.layers.65.self_attn.q_proj.bias": "model-00050-of-00063.safetensors", + "model.layers.65.self_attn.q_proj.weight": "model-00050-of-00063.safetensors", + "model.layers.65.self_attn.v_proj.bias": "model-00050-of-00063.safetensors", + "model.layers.65.self_attn.v_proj.weight": "model-00050-of-00063.safetensors", + "model.layers.66.input_layernorm.weight": "model-00052-of-00063.safetensors", + "model.layers.66.mlp.down_proj.weight": "model-00052-of-00063.safetensors", + "model.layers.66.mlp.gate_proj.weight": "model-00051-of-00063.safetensors", + "model.layers.66.mlp.up_proj.weight": "model-00052-of-00063.safetensors", + "model.layers.66.post_attention_layernorm.weight": "model-00052-of-00063.safetensors", + "model.layers.66.self_attn.k_proj.bias": "model-00051-of-00063.safetensors", + "model.layers.66.self_attn.k_proj.weight": "model-00051-of-00063.safetensors", + "model.layers.66.self_attn.o_proj.weight": "model-00051-of-00063.safetensors", + "model.layers.66.self_attn.q_proj.bias": "model-00051-of-00063.safetensors", + "model.layers.66.self_attn.q_proj.weight": "model-00051-of-00063.safetensors", + "model.layers.66.self_attn.v_proj.bias": "model-00051-of-00063.safetensors", + "model.layers.66.self_attn.v_proj.weight": "model-00051-of-00063.safetensors", + "model.layers.67.input_layernorm.weight": "model-00053-of-00063.safetensors", + "model.layers.67.mlp.down_proj.weight": "model-00053-of-00063.safetensors", + "model.layers.67.mlp.gate_proj.weight": "model-00052-of-00063.safetensors", + "model.layers.67.mlp.up_proj.weight": "model-00052-of-00063.safetensors", + "model.layers.67.post_attention_layernorm.weight": "model-00053-of-00063.safetensors", + "model.layers.67.self_attn.k_proj.bias": "model-00052-of-00063.safetensors", + "model.layers.67.self_attn.k_proj.weight": "model-00052-of-00063.safetensors", + "model.layers.67.self_attn.o_proj.weight": "model-00052-of-00063.safetensors", + "model.layers.67.self_attn.q_proj.bias": "model-00052-of-00063.safetensors", + "model.layers.67.self_attn.q_proj.weight": "model-00052-of-00063.safetensors", + "model.layers.67.self_attn.v_proj.bias": "model-00052-of-00063.safetensors", + "model.layers.67.self_attn.v_proj.weight": "model-00052-of-00063.safetensors", + "model.layers.68.input_layernorm.weight": "model-00053-of-00063.safetensors", + "model.layers.68.mlp.down_proj.weight": "model-00053-of-00063.safetensors", + "model.layers.68.mlp.gate_proj.weight": "model-00053-of-00063.safetensors", + "model.layers.68.mlp.up_proj.weight": "model-00053-of-00063.safetensors", + "model.layers.68.post_attention_layernorm.weight": "model-00053-of-00063.safetensors", + "model.layers.68.self_attn.k_proj.bias": "model-00053-of-00063.safetensors", + "model.layers.68.self_attn.k_proj.weight": "model-00053-of-00063.safetensors", + "model.layers.68.self_attn.o_proj.weight": "model-00053-of-00063.safetensors", + "model.layers.68.self_attn.q_proj.bias": "model-00053-of-00063.safetensors", + "model.layers.68.self_attn.q_proj.weight": "model-00053-of-00063.safetensors", + "model.layers.68.self_attn.v_proj.bias": "model-00053-of-00063.safetensors", + "model.layers.68.self_attn.v_proj.weight": "model-00053-of-00063.safetensors", + "model.layers.69.input_layernorm.weight": "model-00054-of-00063.safetensors", + "model.layers.69.mlp.down_proj.weight": "model-00054-of-00063.safetensors", + "model.layers.69.mlp.gate_proj.weight": "model-00054-of-00063.safetensors", + "model.layers.69.mlp.up_proj.weight": "model-00054-of-00063.safetensors", + "model.layers.69.post_attention_layernorm.weight": "model-00054-of-00063.safetensors", + "model.layers.69.self_attn.k_proj.bias": "model-00053-of-00063.safetensors", + "model.layers.69.self_attn.k_proj.weight": "model-00053-of-00063.safetensors", + "model.layers.69.self_attn.o_proj.weight": "model-00054-of-00063.safetensors", + "model.layers.69.self_attn.q_proj.bias": "model-00053-of-00063.safetensors", + "model.layers.69.self_attn.q_proj.weight": "model-00053-of-00063.safetensors", + "model.layers.69.self_attn.v_proj.bias": "model-00053-of-00063.safetensors", + "model.layers.69.self_attn.v_proj.weight": "model-00053-of-00063.safetensors", + "model.layers.7.input_layernorm.weight": "model-00008-of-00063.safetensors", + "model.layers.7.mlp.down_proj.weight": "model-00008-of-00063.safetensors", + "model.layers.7.mlp.gate_proj.weight": "model-00007-of-00063.safetensors", + "model.layers.7.mlp.up_proj.weight": "model-00007-of-00063.safetensors", + "model.layers.7.post_attention_layernorm.weight": "model-00008-of-00063.safetensors", + "model.layers.7.self_attn.k_proj.bias": "model-00007-of-00063.safetensors", + "model.layers.7.self_attn.k_proj.weight": "model-00007-of-00063.safetensors", + "model.layers.7.self_attn.o_proj.weight": "model-00007-of-00063.safetensors", + "model.layers.7.self_attn.q_proj.bias": "model-00007-of-00063.safetensors", + "model.layers.7.self_attn.q_proj.weight": "model-00007-of-00063.safetensors", + "model.layers.7.self_attn.v_proj.bias": "model-00007-of-00063.safetensors", + "model.layers.7.self_attn.v_proj.weight": "model-00007-of-00063.safetensors", + "model.layers.70.input_layernorm.weight": "model-00055-of-00063.safetensors", + "model.layers.70.mlp.down_proj.weight": "model-00055-of-00063.safetensors", + "model.layers.70.mlp.gate_proj.weight": "model-00054-of-00063.safetensors", + "model.layers.70.mlp.up_proj.weight": "model-00055-of-00063.safetensors", + "model.layers.70.post_attention_layernorm.weight": "model-00055-of-00063.safetensors", + "model.layers.70.self_attn.k_proj.bias": "model-00054-of-00063.safetensors", + "model.layers.70.self_attn.k_proj.weight": "model-00054-of-00063.safetensors", + "model.layers.70.self_attn.o_proj.weight": "model-00054-of-00063.safetensors", + "model.layers.70.self_attn.q_proj.bias": "model-00054-of-00063.safetensors", + "model.layers.70.self_attn.q_proj.weight": "model-00054-of-00063.safetensors", + "model.layers.70.self_attn.v_proj.bias": "model-00054-of-00063.safetensors", + "model.layers.70.self_attn.v_proj.weight": "model-00054-of-00063.safetensors", + "model.layers.71.input_layernorm.weight": "model-00056-of-00063.safetensors", + "model.layers.71.mlp.down_proj.weight": "model-00056-of-00063.safetensors", + "model.layers.71.mlp.gate_proj.weight": "model-00055-of-00063.safetensors", + "model.layers.71.mlp.up_proj.weight": "model-00055-of-00063.safetensors", + "model.layers.71.post_attention_layernorm.weight": "model-00056-of-00063.safetensors", + "model.layers.71.self_attn.k_proj.bias": "model-00055-of-00063.safetensors", + "model.layers.71.self_attn.k_proj.weight": "model-00055-of-00063.safetensors", + "model.layers.71.self_attn.o_proj.weight": "model-00055-of-00063.safetensors", + "model.layers.71.self_attn.q_proj.bias": "model-00055-of-00063.safetensors", + "model.layers.71.self_attn.q_proj.weight": "model-00055-of-00063.safetensors", + "model.layers.71.self_attn.v_proj.bias": "model-00055-of-00063.safetensors", + "model.layers.71.self_attn.v_proj.weight": "model-00055-of-00063.safetensors", + "model.layers.72.input_layernorm.weight": "model-00056-of-00063.safetensors", + "model.layers.72.mlp.down_proj.weight": "model-00056-of-00063.safetensors", + "model.layers.72.mlp.gate_proj.weight": "model-00056-of-00063.safetensors", + "model.layers.72.mlp.up_proj.weight": "model-00056-of-00063.safetensors", + "model.layers.72.post_attention_layernorm.weight": "model-00056-of-00063.safetensors", + "model.layers.72.self_attn.k_proj.bias": "model-00056-of-00063.safetensors", + "model.layers.72.self_attn.k_proj.weight": "model-00056-of-00063.safetensors", + "model.layers.72.self_attn.o_proj.weight": "model-00056-of-00063.safetensors", + "model.layers.72.self_attn.q_proj.bias": "model-00056-of-00063.safetensors", + "model.layers.72.self_attn.q_proj.weight": "model-00056-of-00063.safetensors", + "model.layers.72.self_attn.v_proj.bias": "model-00056-of-00063.safetensors", + "model.layers.72.self_attn.v_proj.weight": "model-00056-of-00063.safetensors", + "model.layers.73.input_layernorm.weight": "model-00057-of-00063.safetensors", + "model.layers.73.mlp.down_proj.weight": "model-00057-of-00063.safetensors", + "model.layers.73.mlp.gate_proj.weight": "model-00057-of-00063.safetensors", + "model.layers.73.mlp.up_proj.weight": "model-00057-of-00063.safetensors", + "model.layers.73.post_attention_layernorm.weight": "model-00057-of-00063.safetensors", + "model.layers.73.self_attn.k_proj.bias": "model-00056-of-00063.safetensors", + "model.layers.73.self_attn.k_proj.weight": "model-00056-of-00063.safetensors", + "model.layers.73.self_attn.o_proj.weight": "model-00057-of-00063.safetensors", + "model.layers.73.self_attn.q_proj.bias": "model-00056-of-00063.safetensors", + "model.layers.73.self_attn.q_proj.weight": "model-00056-of-00063.safetensors", + "model.layers.73.self_attn.v_proj.bias": "model-00056-of-00063.safetensors", + "model.layers.73.self_attn.v_proj.weight": "model-00056-of-00063.safetensors", + "model.layers.74.input_layernorm.weight": "model-00058-of-00063.safetensors", + "model.layers.74.mlp.down_proj.weight": "model-00058-of-00063.safetensors", + "model.layers.74.mlp.gate_proj.weight": "model-00057-of-00063.safetensors", + "model.layers.74.mlp.up_proj.weight": "model-00058-of-00063.safetensors", + "model.layers.74.post_attention_layernorm.weight": "model-00058-of-00063.safetensors", + "model.layers.74.self_attn.k_proj.bias": "model-00057-of-00063.safetensors", + "model.layers.74.self_attn.k_proj.weight": "model-00057-of-00063.safetensors", + "model.layers.74.self_attn.o_proj.weight": "model-00057-of-00063.safetensors", + "model.layers.74.self_attn.q_proj.bias": "model-00057-of-00063.safetensors", + "model.layers.74.self_attn.q_proj.weight": "model-00057-of-00063.safetensors", + "model.layers.74.self_attn.v_proj.bias": "model-00057-of-00063.safetensors", + "model.layers.74.self_attn.v_proj.weight": "model-00057-of-00063.safetensors", + "model.layers.75.input_layernorm.weight": "model-00059-of-00063.safetensors", + "model.layers.75.mlp.down_proj.weight": "model-00059-of-00063.safetensors", + "model.layers.75.mlp.gate_proj.weight": "model-00058-of-00063.safetensors", + "model.layers.75.mlp.up_proj.weight": "model-00058-of-00063.safetensors", + "model.layers.75.post_attention_layernorm.weight": "model-00059-of-00063.safetensors", + "model.layers.75.self_attn.k_proj.bias": "model-00058-of-00063.safetensors", + "model.layers.75.self_attn.k_proj.weight": "model-00058-of-00063.safetensors", + "model.layers.75.self_attn.o_proj.weight": "model-00058-of-00063.safetensors", + "model.layers.75.self_attn.q_proj.bias": "model-00058-of-00063.safetensors", + "model.layers.75.self_attn.q_proj.weight": "model-00058-of-00063.safetensors", + "model.layers.75.self_attn.v_proj.bias": "model-00058-of-00063.safetensors", + "model.layers.75.self_attn.v_proj.weight": "model-00058-of-00063.safetensors", + "model.layers.76.input_layernorm.weight": "model-00059-of-00063.safetensors", + "model.layers.76.mlp.down_proj.weight": "model-00059-of-00063.safetensors", + "model.layers.76.mlp.gate_proj.weight": "model-00059-of-00063.safetensors", + "model.layers.76.mlp.up_proj.weight": "model-00059-of-00063.safetensors", + "model.layers.76.post_attention_layernorm.weight": "model-00059-of-00063.safetensors", + "model.layers.76.self_attn.k_proj.bias": "model-00059-of-00063.safetensors", + "model.layers.76.self_attn.k_proj.weight": "model-00059-of-00063.safetensors", + "model.layers.76.self_attn.o_proj.weight": "model-00059-of-00063.safetensors", + "model.layers.76.self_attn.q_proj.bias": "model-00059-of-00063.safetensors", + "model.layers.76.self_attn.q_proj.weight": "model-00059-of-00063.safetensors", + "model.layers.76.self_attn.v_proj.bias": "model-00059-of-00063.safetensors", + "model.layers.76.self_attn.v_proj.weight": "model-00059-of-00063.safetensors", + "model.layers.77.input_layernorm.weight": "model-00060-of-00063.safetensors", + "model.layers.77.mlp.down_proj.weight": "model-00060-of-00063.safetensors", + "model.layers.77.mlp.gate_proj.weight": "model-00060-of-00063.safetensors", + "model.layers.77.mlp.up_proj.weight": "model-00060-of-00063.safetensors", + "model.layers.77.post_attention_layernorm.weight": "model-00060-of-00063.safetensors", + "model.layers.77.self_attn.k_proj.bias": "model-00059-of-00063.safetensors", + "model.layers.77.self_attn.k_proj.weight": "model-00059-of-00063.safetensors", + "model.layers.77.self_attn.o_proj.weight": "model-00060-of-00063.safetensors", + "model.layers.77.self_attn.q_proj.bias": "model-00059-of-00063.safetensors", + "model.layers.77.self_attn.q_proj.weight": "model-00059-of-00063.safetensors", + "model.layers.77.self_attn.v_proj.bias": "model-00059-of-00063.safetensors", + "model.layers.77.self_attn.v_proj.weight": "model-00059-of-00063.safetensors", + "model.layers.78.input_layernorm.weight": "model-00061-of-00063.safetensors", + "model.layers.78.mlp.down_proj.weight": "model-00061-of-00063.safetensors", + "model.layers.78.mlp.gate_proj.weight": "model-00060-of-00063.safetensors", + "model.layers.78.mlp.up_proj.weight": "model-00061-of-00063.safetensors", + "model.layers.78.post_attention_layernorm.weight": "model-00061-of-00063.safetensors", + "model.layers.78.self_attn.k_proj.bias": "model-00060-of-00063.safetensors", + "model.layers.78.self_attn.k_proj.weight": "model-00060-of-00063.safetensors", + "model.layers.78.self_attn.o_proj.weight": "model-00060-of-00063.safetensors", + "model.layers.78.self_attn.q_proj.bias": "model-00060-of-00063.safetensors", + "model.layers.78.self_attn.q_proj.weight": "model-00060-of-00063.safetensors", + "model.layers.78.self_attn.v_proj.bias": "model-00060-of-00063.safetensors", + "model.layers.78.self_attn.v_proj.weight": "model-00060-of-00063.safetensors", + "model.layers.79.input_layernorm.weight": "model-00062-of-00063.safetensors", + "model.layers.79.mlp.down_proj.weight": "model-00062-of-00063.safetensors", + "model.layers.79.mlp.gate_proj.weight": "model-00061-of-00063.safetensors", + "model.layers.79.mlp.up_proj.weight": "model-00061-of-00063.safetensors", + "model.layers.79.post_attention_layernorm.weight": "model-00062-of-00063.safetensors", + "model.layers.79.self_attn.k_proj.bias": "model-00061-of-00063.safetensors", + "model.layers.79.self_attn.k_proj.weight": "model-00061-of-00063.safetensors", + "model.layers.79.self_attn.o_proj.weight": "model-00061-of-00063.safetensors", + "model.layers.79.self_attn.q_proj.bias": "model-00061-of-00063.safetensors", + "model.layers.79.self_attn.q_proj.weight": "model-00061-of-00063.safetensors", + "model.layers.79.self_attn.v_proj.bias": "model-00061-of-00063.safetensors", + "model.layers.79.self_attn.v_proj.weight": "model-00061-of-00063.safetensors", + "model.layers.8.input_layernorm.weight": "model-00008-of-00063.safetensors", + "model.layers.8.mlp.down_proj.weight": "model-00008-of-00063.safetensors", + "model.layers.8.mlp.gate_proj.weight": "model-00008-of-00063.safetensors", + "model.layers.8.mlp.up_proj.weight": "model-00008-of-00063.safetensors", + "model.layers.8.post_attention_layernorm.weight": "model-00008-of-00063.safetensors", + "model.layers.8.self_attn.k_proj.bias": "model-00008-of-00063.safetensors", + "model.layers.8.self_attn.k_proj.weight": "model-00008-of-00063.safetensors", + "model.layers.8.self_attn.o_proj.weight": "model-00008-of-00063.safetensors", + "model.layers.8.self_attn.q_proj.bias": "model-00008-of-00063.safetensors", + "model.layers.8.self_attn.q_proj.weight": "model-00008-of-00063.safetensors", + "model.layers.8.self_attn.v_proj.bias": "model-00008-of-00063.safetensors", + "model.layers.8.self_attn.v_proj.weight": "model-00008-of-00063.safetensors", + "model.layers.9.input_layernorm.weight": "model-00009-of-00063.safetensors", + "model.layers.9.mlp.down_proj.weight": "model-00009-of-00063.safetensors", + "model.layers.9.mlp.gate_proj.weight": "model-00009-of-00063.safetensors", + "model.layers.9.mlp.up_proj.weight": "model-00009-of-00063.safetensors", + "model.layers.9.post_attention_layernorm.weight": "model-00009-of-00063.safetensors", + "model.layers.9.self_attn.k_proj.bias": "model-00008-of-00063.safetensors", + "model.layers.9.self_attn.k_proj.weight": "model-00008-of-00063.safetensors", + "model.layers.9.self_attn.o_proj.weight": "model-00009-of-00063.safetensors", + "model.layers.9.self_attn.q_proj.bias": "model-00008-of-00063.safetensors", + "model.layers.9.self_attn.q_proj.weight": "model-00008-of-00063.safetensors", + "model.layers.9.self_attn.v_proj.bias": "model-00008-of-00063.safetensors", + "model.layers.9.self_attn.v_proj.weight": "model-00008-of-00063.safetensors", + "model.norm.weight": "model-00062-of-00063.safetensors" + } +}