diff --git a/config.json b/config.json new file mode 100644 index 0000000000000000000000000000000000000000..88fdf43417673a5cf45b3195465d3d9d84d0ad63 --- /dev/null +++ b/config.json @@ -0,0 +1,29 @@ +{ + "_name_or_path": "./merged", + "architectures": [ + "LlamaForCausalLM" + ], + "attention_bias": false, + "attention_dropout": 0.0, + "bos_token_id": 1, + "eos_token_id": 2, + "hidden_act": "silu", + "hidden_size": 8192, + "initializer_range": 0.02, + "intermediate_size": 28672, + "max_position_embeddings": 4096, + "model_type": "llama", + "num_attention_heads": 64, + "num_hidden_layers": 140, + "num_key_value_heads": 8, + "pad_token_id": 0, + "pretraining_tp": 1, + "rms_norm_eps": 1e-05, + "rope_scaling": null, + "rope_theta": 10000.0, + "tie_word_embeddings": false, + "torch_dtype": "float32", + "transformers_version": "4.36.2", + "use_cache": false, + "vocab_size": 32000 +} diff --git a/generation_config.json b/generation_config.json new file mode 100644 index 0000000000000000000000000000000000000000..5786877b25c8295c74590b8a5aace96539c44dd2 --- /dev/null +++ b/generation_config.json @@ -0,0 +1,8 @@ +{ + "_from_model_config": true, + "bos_token_id": 1, + "eos_token_id": 2, + "pad_token_id": 0, + "transformers_version": "4.36.2", + "use_cache": false +} diff --git a/model-00001-of-00106.safetensors b/model-00001-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..6c9223c208cdb7c7d1d3a834efcd8c20debe84e4 --- /dev/null +++ b/model-00001-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:13668de8b7511121fd9a34ba2c6d3e3101343cf6b282d6e47271849c402fda52 +size 4806739440 diff --git a/model-00002-of-00106.safetensors b/model-00002-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..4f418325400cd58ba8ad4d159e08fc54c62f1a05 --- /dev/null +++ b/model-00002-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:c21a2e1da864355a405464c8580ca31350e8b3f04dba6bb25f702a04e4a5227a +size 4630578440 diff --git a/model-00003-of-00106.safetensors b/model-00003-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..da4dc76edd86ecde269f0b439933933e798910a4 --- /dev/null +++ b/model-00003-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:ebf3e929c255eb158c8e1dd522710aae90785aa93770981288546b5751f045a5 +size 4362142864 diff --git a/model-00004-of-00106.safetensors b/model-00004-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..2e07cde14d0bc9ca28237a3b57e52df7b51279f2 --- /dev/null +++ b/model-00004-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:3de29266d83040dece977b9fc8cdd60b882502945fce6e8f86575ecadeb301bd +size 4966188864 diff --git a/model-00005-of-00106.safetensors b/model-00005-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..ee920621bac0fcb1176d0c112dfb7d2665cd12f2 --- /dev/null +++ b/model-00005-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:206a73f4c7be2908fc5b8b6f09763c8f67d9066cf01961f7017e2f921ba23b5c +size 4362142864 diff --git a/model-00006-of-00106.safetensors b/model-00006-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..a3092cd84b93e2e9511a8a526f71fa2a73eebd38 --- /dev/null +++ b/model-00006-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:81b730dc11d5b5f14b14168c6fd1ecfab34027c473c585f8ec74bd6d61112b82 +size 4362142864 diff --git a/model-00007-of-00106.safetensors b/model-00007-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..d42bd5269cd3085f6ab1d7ceff3d628b58dda01d --- /dev/null +++ b/model-00007-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:b9168a425c9a8f5b02a7472be062a36e84aa40c45c4ddcdd00c5744126b46664 +size 4966188864 diff --git a/model-00008-of-00106.safetensors b/model-00008-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..68ae95a49c4837b1e7f9236b92081ee8b23f0b94 --- /dev/null +++ b/model-00008-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:2101b1e48da7bec31a4ec96764c6d002e18c20a0daae6d48cbf86a79b37b9b23 +size 4362142880 diff --git a/model-00009-of-00106.safetensors b/model-00009-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..994fea7a51b38dc7ed64fb9fa70b601a00258573 --- /dev/null +++ b/model-00009-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:d9a8d3fd8fdbd3d693dd7900d3dd3d6fd7177b44592ab0d8689637f692ace111 +size 4362142872 diff --git a/model-00010-of-00106.safetensors b/model-00010-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..01bb38099c8d374cdf60b7ddc9f5de59c2c89b3d --- /dev/null +++ b/model-00010-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:bd7fb3096beadbe22ff4fd56d909762b3b875a8c25ad03369a3b120c69fc0371 +size 4966188880 diff --git a/model-00011-of-00106.safetensors b/model-00011-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..8155226d7098bda4e3c245c145c3d276bfd30383 --- /dev/null +++ b/model-00011-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:41b9252dae15650603e4be24d306a37ac22c36613bf94455929c4aa7dd20ba16 +size 4362142872 diff --git a/model-00012-of-00106.safetensors b/model-00012-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..56b308ce674bb8850768c6ef45a5d914adbf2211 --- /dev/null +++ b/model-00012-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:2d14275127545947a1f7ee84df7cf0bf867966e343d6596bb7e7191d7f629c06 +size 4362142872 diff --git a/model-00013-of-00106.safetensors b/model-00013-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..9aafa6c8f036bb94c4f24b67da02970b050566f4 --- /dev/null +++ b/model-00013-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:b14292bf5efa09495e45dd487391325ec94b94509f31497c05eb3aa252c4403f +size 4966188880 diff --git a/model-00014-of-00106.safetensors b/model-00014-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..8db7a5a236b1f6a367c99127bcb6b249a8264d89 --- /dev/null +++ b/model-00014-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:8579526668ef755dfa3bf8aac385d3379affb76f23ae5cf5093e301e98f7e027 +size 4362142872 diff --git a/model-00015-of-00106.safetensors b/model-00015-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..61d5f142435e3e15903a021ba764f2bb793c0711 --- /dev/null +++ b/model-00015-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:bcf1ada9ba2d25f419e9e2f91c20914e4be1377f650fa49d2dfb9e5ce51ab398 +size 4362142872 diff --git a/model-00016-of-00106.safetensors b/model-00016-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..c22fb6ef7ef480fd63748367bf7157e0a091480b --- /dev/null +++ b/model-00016-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:dfb3f6a729f36bc7619a5fc837802a74812df8c69b7f22dda260db9651c86031 +size 4966188880 diff --git a/model-00017-of-00106.safetensors b/model-00017-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..e37a51ff4f73db115f074c644c3d056668d35afd --- /dev/null +++ b/model-00017-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:453462f72541378743e47ecf4e07ac5102c217c364494b1de0b4dca46ef75e9d +size 4362142872 diff --git a/model-00018-of-00106.safetensors b/model-00018-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..135f2fd7934d1a77464f7bad3b12743040b46c09 --- /dev/null +++ b/model-00018-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:e53b08f4b761076b6aa413e43e5b988b765cf0d46d45c49371e110f13f5f037b +size 4362142872 diff --git a/model-00019-of-00106.safetensors b/model-00019-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..08531a0efdeedd6a20a956ec6c585dab28d066ee --- /dev/null +++ b/model-00019-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:65096c7783b957e40a03daeeb079a642f4c7ab62e3df92d07fca6a96bdf4861d +size 4966188880 diff --git a/model-00020-of-00106.safetensors b/model-00020-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..1feb13d56bb25fa5ec67c8d0a9cf6c72f66b3fa6 --- /dev/null +++ b/model-00020-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:806a9eb3da6a1882f5a43b86a2eab21a4592c987dbbe99df64e09f522925fbe2 +size 4362142872 diff --git a/model-00021-of-00106.safetensors b/model-00021-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..b9aa7b2c0e8b3391930d1d4c66e6bddc75aa26df --- /dev/null +++ b/model-00021-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:8ee3c8546a50a9cecf187451590d5b1aca7009a5bd9364cfa87a5b8ddfdb4022 +size 4362142872 diff --git a/model-00022-of-00106.safetensors b/model-00022-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..da34b9246f2b7196602bb23401de519b8b7e057f --- /dev/null +++ b/model-00022-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:b7c148533588907213221a276b4d673af8a97b94f5786a6fb4c038cbc39f701c +size 4966188880 diff --git a/model-00023-of-00106.safetensors b/model-00023-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..bc59b51244996178401af7c3f2deae01e5a5146c --- /dev/null +++ b/model-00023-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:2d322bb8653aa4929901b9119dfa0fdcd6633b05d0741b48213c6612e57b76b7 +size 4362142872 diff --git a/model-00024-of-00106.safetensors b/model-00024-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..7d0f9a1ea11399252e0a2771db3d3297d92998bf --- /dev/null +++ b/model-00024-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:3fb7fee679e4b646fce7df478858f7cb7570357ac81ad946564158b2833db3b1 +size 4362142872 diff --git a/model-00025-of-00106.safetensors b/model-00025-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..567826a022ea5a852f880ceb51d6e6f9cc6fdf8f --- /dev/null +++ b/model-00025-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:e2ad5ee15cee34e13106d7addc028405e72ce167032f88efbea995877017118c +size 4966188880 diff --git a/model-00026-of-00106.safetensors b/model-00026-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..c836b46a2a1ba853096f3dd4c6df9a77f10ba26d --- /dev/null +++ b/model-00026-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:7f5c5ad57fb921e16d18be91deb9a01f9ed89cc515af0be9dcd13f3c0946e7c2 +size 4362142872 diff --git a/model-00027-of-00106.safetensors b/model-00027-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..34884225cf0259e079841a48a312c827742828e4 --- /dev/null +++ b/model-00027-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:1861787259cdb9cf0dbb1cbc2dd907567655ab7dd22e146306ce1fde51ef09a1 +size 4362142872 diff --git a/model-00028-of-00106.safetensors b/model-00028-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..583cfb0d196ac88e4dc08fb4d07474d72b5cba0a --- /dev/null +++ b/model-00028-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:919ac271a48742cd63d8e334d4639a2b6affaa8abf1230af44d0ae98928baf7c +size 4966188880 diff --git a/model-00029-of-00106.safetensors b/model-00029-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..5ddba782a8381ab9fb5f0bd55e16f6e613b2aa8a --- /dev/null +++ b/model-00029-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:68f62e4e0b70a2b6c0c8dc0821b9640e2058eadd2518061b70e31af35e2c677d +size 4362142872 diff --git a/model-00030-of-00106.safetensors b/model-00030-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..c0a00ccb1627b60cb12d7bf3ffd7c904cb6b9418 --- /dev/null +++ b/model-00030-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:65fa1210cf65d43c47df0d41188e276190fb3b8f1656654fc01225c9e68a54f1 +size 4362142872 diff --git a/model-00031-of-00106.safetensors b/model-00031-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..580c3e705ae91397c7928c3ba2b63ce9d12bfdea --- /dev/null +++ b/model-00031-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:1477296b0c43aa6f94b6419cf648d905695f3b02cca451f6760e88f63cd92f7b +size 4966188880 diff --git a/model-00032-of-00106.safetensors b/model-00032-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..f68e5923a739bdd863ad4f4011a1fe65bc72d4d1 --- /dev/null +++ b/model-00032-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:d527dce5cddc274ac42d91f883b894624d01fa682792bf6901648b8d2d973e59 +size 4362142872 diff --git a/model-00033-of-00106.safetensors b/model-00033-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..412e2af011279c7ab8106fb22b118fccf8560f85 --- /dev/null +++ b/model-00033-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:4e54efdcf7bec565248080ccbda1629aeb0d2e4085985d996fd622e4fefb83e5 +size 4362142872 diff --git a/model-00034-of-00106.safetensors b/model-00034-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..ca7ff8346a09e099234b8c3cb4fe8d5c2518fbbe --- /dev/null +++ b/model-00034-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:12eff68e1ab5553881a1129f5921dc53c1131bf2ae3a335b989ccddb2ab032bd +size 4966188880 diff --git a/model-00035-of-00106.safetensors b/model-00035-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..f9ae69a50326650c9c8e779dcadbb1620439a73e --- /dev/null +++ b/model-00035-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:22c7ebabe895a7ca100fd97d6f339b7ab5ca67c4fed03a3c69659bd6a31b0747 +size 4362142872 diff --git a/model-00036-of-00106.safetensors b/model-00036-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..023a760849dd4637e80e4438137d6e61f9ab10bd --- /dev/null +++ b/model-00036-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:3963ca58c8343fd096240ab13c9c6a044d1f8f63518ae00eaacd4b49d35cfb66 +size 4362142872 diff --git a/model-00037-of-00106.safetensors b/model-00037-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..5f877a20589b59033e3e4514b105f3102d163c6d --- /dev/null +++ b/model-00037-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:cf995ee0ccd594e56e321f6527c35d6022b6613fd0244d46299d8484296cd358 +size 4966188880 diff --git a/model-00038-of-00106.safetensors b/model-00038-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..21ffe666adecb24cf065112fe296b3fb15962eff --- /dev/null +++ b/model-00038-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:ab2f4957ba21721580beab6aad3d05f8a113198943fac833e66bf38c555265bb +size 4362142872 diff --git a/model-00039-of-00106.safetensors b/model-00039-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..13321e5ffc224816aa036aa06bb203ef4707dd2e --- /dev/null +++ b/model-00039-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:82fb192e9cafc7e970dc44f6f3b639414113068b7172e3981708a7027537db34 +size 4362142872 diff --git a/model-00040-of-00106.safetensors b/model-00040-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..dea64b06fc726cfa96588a1a9d2f3b7624c75633 --- /dev/null +++ b/model-00040-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:ff283f4e5f9c8da8a8c43023cce34503f6fad46e7a4e38b53bf92c5be96819da +size 4966188880 diff --git a/model-00041-of-00106.safetensors b/model-00041-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..af5790006b206c89d32d35a37a630a07c35d6fda --- /dev/null +++ b/model-00041-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:3234a503a9f6bf7e1383f2aab2ea7657ba6da313ec27412c8de6f209fc6314fa +size 4362142872 diff --git a/model-00042-of-00106.safetensors b/model-00042-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..baf922c251102ea217660b3758116984588d88c5 --- /dev/null +++ b/model-00042-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:a4634579c66f12c0eadf748242e5fd02d5668f3ff0cb757b377d10eecafa63c5 +size 4362142872 diff --git a/model-00043-of-00106.safetensors b/model-00043-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..860f80f50f251f755c1c0899a3442101fdc2fb84 --- /dev/null +++ b/model-00043-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:dd898142c17a32a31ac8c1af864937596bba7f085452e3205243d208f04258fb +size 4966188880 diff --git a/model-00044-of-00106.safetensors b/model-00044-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..adab94c31675497ae5ab621bdbb6a585725d6f7c --- /dev/null +++ b/model-00044-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:f4c8e07789b06e9648062dcf04f7d3202818692cb042c35e8f3f3306cf913276 +size 4362142872 diff --git a/model-00045-of-00106.safetensors b/model-00045-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..c2fe8778c740cae41fdd0897a8edce907213027a --- /dev/null +++ b/model-00045-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:ec51be1731eaa80ed1dfb08746e5c1ee7c4f7f51a1ab155333d85a676018db1f +size 4362142872 diff --git a/model-00046-of-00106.safetensors b/model-00046-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..87d00abc7c9ac5cac03332b9e2f0da5e8806e980 --- /dev/null +++ b/model-00046-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:c4cfefff0e39aad76ef7ff7c787d237acec863100fe0e91838b393ba910dc2b3 +size 4966188880 diff --git a/model-00047-of-00106.safetensors b/model-00047-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..f689ca826f2708d2fe9177015fc6be07ea8a2c26 --- /dev/null +++ b/model-00047-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:2e2339e33130ea1fedf92b14a9af0f81122d0964ae5e3afa65507135723701d2 +size 4362142872 diff --git a/model-00048-of-00106.safetensors b/model-00048-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..05489e140949dde4164cafca882b964824598a12 --- /dev/null +++ b/model-00048-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:8c1086da9328bf942506c9d0ebcbcbca5f22597bc59257599035997afe73ea2e +size 4362142872 diff --git a/model-00049-of-00106.safetensors b/model-00049-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..980baa2ceb3a354dc64f2581b78378f7daa7f1bf --- /dev/null +++ b/model-00049-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:978f01d07d1ff70d114aece2931ff62d0480d3f1fdf7db24fe0368312499940a +size 4966188880 diff --git a/model-00050-of-00106.safetensors b/model-00050-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..bd7eb776b4489dfe442786e1200545d7bd490261 --- /dev/null +++ b/model-00050-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:e7c970049708693af842da27cacfccc51f225a9087ea25bb90b8d84e8cce5df2 +size 4362142872 diff --git a/model-00051-of-00106.safetensors b/model-00051-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..405cfb5c1b1155c4c456fdfd52fcad340464cd5a --- /dev/null +++ b/model-00051-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:38b2476f0410a4e44a6db9953a77e20d1eeae3967cec9a506c0a78bf4379d259 +size 4362142872 diff --git a/model-00052-of-00106.safetensors b/model-00052-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..06130572d59427146daee03d6048504351ef3c08 --- /dev/null +++ b/model-00052-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:dd02d19b55ccc01a407b99b787acac20f960608029077dfb94792a31b81f9298 +size 4966188880 diff --git a/model-00053-of-00106.safetensors b/model-00053-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..261b975cd9e8ca41915bd95cf1f201db9b175b69 --- /dev/null +++ b/model-00053-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:b1e3ff90ade347fda4f209184f9790d091fb9753397de7df78d8a2b1e806b8e5 +size 4362142872 diff --git a/model-00054-of-00106.safetensors b/model-00054-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..3962244bd9b188cbf1a7dad25f38b6fa17b4f41a --- /dev/null +++ b/model-00054-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:be466284e319acc7a82d2d0da41c89daed386cb55457a8b0da0991dcd4397ab6 +size 4362142872 diff --git a/model-00055-of-00106.safetensors b/model-00055-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..d0903f5b7b9ea38b3522d195d7115b7ca60d8e35 --- /dev/null +++ b/model-00055-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:6c1ea292b24206ecd970059e07441b9cf03b0ed76f0b4e0cf86e077f40f1e9d3 +size 4966188880 diff --git a/model-00056-of-00106.safetensors b/model-00056-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..3e2102bd72aad4470cf9a250624e33b622485007 --- /dev/null +++ b/model-00056-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:f7f8066f67902e71af6cc120ffd3fc2c388bb7c677e79a1fe064345f1dd94a50 +size 4362142872 diff --git a/model-00057-of-00106.safetensors b/model-00057-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..4093d80e7016ccdbbd58659993414ec85686d644 --- /dev/null +++ b/model-00057-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:90edb4112b0e8d52bc8744c283474df33b02804b64ae73ab353cee57e86248aa +size 4362142872 diff --git a/model-00058-of-00106.safetensors b/model-00058-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..0f7b9bbb383f0b990d98ff959de4b7551f19c480 --- /dev/null +++ b/model-00058-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:292275e21e65ce9b817caed53f68af49ee7bd82517f75cc77c23fb59b6c90a90 +size 4966188880 diff --git a/model-00059-of-00106.safetensors b/model-00059-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..446e80a384912ce3e7b53e7c3dad442680b4de90 --- /dev/null +++ b/model-00059-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:d40e68cbf2cc55b482324645bb86fd258c99efa15b6be5b604c5d7c69601acdd +size 4362142872 diff --git a/model-00060-of-00106.safetensors b/model-00060-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..38215b0f50c6874398f697babf3edc8392ee6f63 --- /dev/null +++ b/model-00060-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:6335e6740e648c9daf9abd7dafda9c24b1458cd4f739135b970d36e3015b088c +size 4362142872 diff --git a/model-00061-of-00106.safetensors b/model-00061-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..5c562f719cb886eaba77f90287a5f2433b2805ec --- /dev/null +++ b/model-00061-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:56a3a892b83b1328812850dfdc58ab97028840d8965014fc85ee8eef0978005a +size 4966188880 diff --git a/model-00062-of-00106.safetensors b/model-00062-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..0a8a2caea67a34d262bb29db3f3c862374750b68 --- /dev/null +++ b/model-00062-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:bd454647be722f287c1e9699c028a5a4f72beeba9c4e71052a756da43874349a +size 4362142872 diff --git a/model-00063-of-00106.safetensors b/model-00063-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..784849ff4a55269728b52fa152ff4b47bfc2bf83 --- /dev/null +++ b/model-00063-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:a990dab8294b458eefd6093af87f88e20b8fad4103d5e6d07d5ab4ff61dc65cb +size 4362142872 diff --git a/model-00064-of-00106.safetensors b/model-00064-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..1c0f5b0cbeefb66633385e12abb756426c2b049b --- /dev/null +++ b/model-00064-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:71957d14c8d02c3a9ff7e56bcd6f4dcae6f816d90fe74bb81611bde06d6a3171 +size 4966188880 diff --git a/model-00065-of-00106.safetensors b/model-00065-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..fc82cbe2032a8a9d95ad820adb6ff50d96d18fe5 --- /dev/null +++ b/model-00065-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:fb4b2bb78758e4a9b48fe645343c9fb154463edb49358f4d5917da2c04757b00 +size 4362142872 diff --git a/model-00066-of-00106.safetensors b/model-00066-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..eb6c7561734e6c008aac2b13a33299e886b9638e --- /dev/null +++ b/model-00066-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:211c7eae19e5edc340d139ef58003b55ab6db5dc426411c962553a7608bdcfa2 +size 4362142872 diff --git a/model-00067-of-00106.safetensors b/model-00067-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..72aabb1b37dd12fff0921fab54b0dcd0433badb9 --- /dev/null +++ b/model-00067-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:627450074e9646600bcd911c8c6c103a375a8bfb3a34260af31ea6bc2461b535 +size 4966188880 diff --git a/model-00068-of-00106.safetensors b/model-00068-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..d21ef958d3a628df4cb595e55dc43f26e95b62ab --- /dev/null +++ b/model-00068-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:b80127eb68600af927d27790e497162f35096a4bce28d3460ce80292d6e01e36 +size 4362142872 diff --git a/model-00069-of-00106.safetensors b/model-00069-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..dd5907c97cfc75fc7661adf2f282589af08c8eb4 --- /dev/null +++ b/model-00069-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:ae8312c964e8984826106530e38dc62a71bff0b2f2fb5c1c4e3ce1a2b84c3925 +size 4362142872 diff --git a/model-00070-of-00106.safetensors b/model-00070-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..e4e581bc9c0befe275685d8747d817dd8e7016d9 --- /dev/null +++ b/model-00070-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:aa322e0f8bb9f978e7c9084c26105207332380c03a63dcebaa5ae8017816eb01 +size 4966188880 diff --git a/model-00071-of-00106.safetensors b/model-00071-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..a86f54dd6fd974e1c35ee1977edb8a0895441ac1 --- /dev/null +++ b/model-00071-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:62b7a76792f5d96ef21abc2774dadc85c9d3d4d6b59527b8ebfa02c7528d4905 +size 4362142872 diff --git a/model-00072-of-00106.safetensors b/model-00072-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..ec6adac409eb66d423de0891df71b10399cf37cd --- /dev/null +++ b/model-00072-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:2f00a3f92639a08360941e63de3da9a3d196b060e70ccd3723c1b20f506cc280 +size 4362142872 diff --git a/model-00073-of-00106.safetensors b/model-00073-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..ae7add7222f7d4f30f17af3a45c2990578110e30 --- /dev/null +++ b/model-00073-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:7f770a0dfb5ba9e1e99607a4812a31383c9c2f809da4091b569e29d4aa60182c +size 4966188880 diff --git a/model-00074-of-00106.safetensors b/model-00074-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..aa3257848696973bbf505a435288988f5f8cdc71 --- /dev/null +++ b/model-00074-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:81686f6c4ccc80614be07b3f915ca8fd40c4e0439f4d5d345c3830327ac0c3c4 +size 4362142872 diff --git a/model-00075-of-00106.safetensors b/model-00075-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..803185bbf9b34e97d3fc44f5c82b26b1bc382a64 --- /dev/null +++ b/model-00075-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:e0902b9574c63bf5afb7ce591b7132535fdcdcc6f20ecbfcff9201b9780982e2 +size 4362142872 diff --git a/model-00076-of-00106.safetensors b/model-00076-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..279993386072c3a05e467182f76f7bdc3b043187 --- /dev/null +++ b/model-00076-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:061b312e46d6aeed24b547568e4d1d89f9173db921280edabe914a34afe8a1df +size 4966188896 diff --git a/model-00077-of-00106.safetensors b/model-00077-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..8c2e6f2c47024cf0313f3ad20f78d25132b1e0ff --- /dev/null +++ b/model-00077-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:a33046501c9870e8645dfa37a6e388d089b57ab1c3da932b2c4c8ee6ddaec417 +size 4362142888 diff --git a/model-00078-of-00106.safetensors b/model-00078-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..91d37a58dc7c80a53a04ee237f97f810e0f48570 --- /dev/null +++ b/model-00078-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:f0fda336214e44826b269656918833baca3c03e92779335559778351fb1c7bb8 +size 4362142880 diff --git a/model-00079-of-00106.safetensors b/model-00079-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..b1077e09c7beb927eaece2ce9d2fd3f229245a16 --- /dev/null +++ b/model-00079-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:02f30386188ff713d1ca1c1eaedc85443d3087c4ac95ca7d475cdce8a685b450 +size 4966188896 diff --git a/model-00080-of-00106.safetensors b/model-00080-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..539b45fe759e661ca18a2413f36071b5439e5736 --- /dev/null +++ b/model-00080-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:8a88a65de579ce76e86b06e0bdeea4f8915ae2636fb9ef037f1dd94bf29222b4 +size 4362142888 diff --git a/model-00081-of-00106.safetensors b/model-00081-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..94dd893bb6d4ff556afb38f31a8712528a254f01 --- /dev/null +++ b/model-00081-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:f2b9f15263a1d40a5c2c295e431d6ca25642cf1c114d22b3c5e10f1963e2cddd +size 4362142880 diff --git a/model-00082-of-00106.safetensors b/model-00082-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..5bc354d5e0a55bcf6e2a67db7bffcc4143ffab31 --- /dev/null +++ b/model-00082-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:e52dbb2444b69189e04728eb47557abb0a5740ace35e4a446cede53f20f35e02 +size 4966188896 diff --git a/model-00083-of-00106.safetensors b/model-00083-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..0e2185f7e2fdda12f703052186a2cf475b8626c0 --- /dev/null +++ b/model-00083-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:33656ae2d9e051514934bceb615f3ac3c7ff5d3ebf39f4df5a2e722a99a05445 +size 4362142888 diff --git a/model-00084-of-00106.safetensors b/model-00084-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..4c1ea0adfdfc845a9cac628de14caa226ed74490 --- /dev/null +++ b/model-00084-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:b6403ac5039ac43bf56a32304847b72af15bcd1c42a0cf3972230bf2007b9d8d +size 4362142880 diff --git a/model-00085-of-00106.safetensors b/model-00085-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..0bdc0b68dbff6db18e574746c9c8d9428ec52a47 --- /dev/null +++ b/model-00085-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:f475e542e067d9d3e61711d6512015bc516e8f1b67429d90fc340b7ca92c5b64 +size 4966188896 diff --git a/model-00086-of-00106.safetensors b/model-00086-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..d8ec0ef67cefbaebc3d7e1c0bfb7633223128117 --- /dev/null +++ b/model-00086-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:f8a10bc335a0e3e00d46cd2fc2207218f57a62631f71eb05de69615b964e7850 +size 4362142888 diff --git a/model-00087-of-00106.safetensors b/model-00087-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..db8d640d9fa8e6f64d105c9763a0d4a319c9d0c1 --- /dev/null +++ b/model-00087-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:619d69b4a1b26adfc3fc49dfa2c44cfc27f23a10a3634dda64df0e710aeacb1b +size 4362142880 diff --git a/model-00088-of-00106.safetensors b/model-00088-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..08c655a650f80b4863426fa91f32e38ad28dc369 --- /dev/null +++ b/model-00088-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:1aad4f238d5c41aa75c01dcb2afd7adae86a1534ee9b8f33da7c6e09539e605c +size 4966188896 diff --git a/model-00089-of-00106.safetensors b/model-00089-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..5df57a6ea92de1b5d181b382bbaaa17cc10e6d00 --- /dev/null +++ b/model-00089-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:f362fe30b9436594e81cb871f3280feec3fc8d387d166d862ea07f13ea0f7c6c +size 4362142888 diff --git a/model-00090-of-00106.safetensors b/model-00090-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..06ad4c7c8fe4d0b09153196e56ba7fd5a3a35465 --- /dev/null +++ b/model-00090-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:3e5ac353465d98d4fbd13eeaf07021c10e09e17fa964580dc5faf6f37ca1d541 +size 4362142880 diff --git a/model-00091-of-00106.safetensors b/model-00091-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..955dcf47d9d1508877f24ec2d33531a37a6f4a7f --- /dev/null +++ b/model-00091-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:d3ba29f8bc93f2061dc027646b3fe38eb6bfc6b0d89c5556e5630ed91bca0686 +size 4966188896 diff --git a/model-00092-of-00106.safetensors b/model-00092-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..ffc1900c2ac57d87d0f49b8a97a0853572ced1de --- /dev/null +++ b/model-00092-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:d19d866f746e89f5eaaa2e9808ca8365cfede0eabe38fbaa85b6ff03c9e62068 +size 4362142888 diff --git a/model-00093-of-00106.safetensors b/model-00093-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..78be3a008f9c89364caa17bd22d2b7f0caadfb12 --- /dev/null +++ b/model-00093-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:7b76288dfca56c10928708a7ea5e6b9714e90af17a3398712d6de56553bfb3cf +size 4362142880 diff --git a/model-00094-of-00106.safetensors b/model-00094-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..45b59363b44c2e60fc90bdb255c7edca142cb6ff --- /dev/null +++ b/model-00094-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:259b7fd884bd2af86adf0105a7b359e0133319a6048bfabf4283c6eb84074e4f +size 4966188896 diff --git a/model-00095-of-00106.safetensors b/model-00095-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..de763fd7a50887c142a4100f5fa742f774b2b096 --- /dev/null +++ b/model-00095-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:c121375461f85483feb83edcb4feea9ea3fe401441f09d62184eaab077972002 +size 4362142888 diff --git a/model-00096-of-00106.safetensors b/model-00096-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..d2fa38f884ecb9007d5fd32828db9ba6d5a6d7d7 --- /dev/null +++ b/model-00096-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:071bf7013bed8300bf90e4ad7b35fc64f3ca6e511d61d47c918c8ff9626b9e40 +size 4362142880 diff --git a/model-00097-of-00106.safetensors b/model-00097-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..c77635f3c65b0e9cd64afb363d688e393348dba3 --- /dev/null +++ b/model-00097-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:e785f352b0a84e9c3813b75b2b9d25e0dea886b3e32144c500102f46f43399bf +size 4966188896 diff --git a/model-00098-of-00106.safetensors b/model-00098-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..8fa509004c7275207bf9354b345a160f815ee23c --- /dev/null +++ b/model-00098-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:6618d04851e9b57ca0e763bdd22a18ee3f4c648013d143af06493f86fb711c85 +size 4362142888 diff --git a/model-00099-of-00106.safetensors b/model-00099-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..0b3f22ec561bac15a6a5cc1cc2e87bc0677d9414 --- /dev/null +++ b/model-00099-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:40fbc154ea9c76ea69ef47733680e7fd513b52308c7d1476b527141bb5afd802 +size 4362142880 diff --git a/model-00100-of-00106.safetensors b/model-00100-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..66d6534e84398e0bb130e1da1bb22457a82dd02d --- /dev/null +++ b/model-00100-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:38613948191fa636ea2d4355039b4da227fc774045a30e19de3de5d4f0adde78 +size 4966188896 diff --git a/model-00101-of-00106.safetensors b/model-00101-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..3eca31190805e69ecbf743d5a6d7394859e3cfb9 --- /dev/null +++ b/model-00101-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:6f19ec01901a83f67c0566fb0bf9f18b285b9e870be5978ccd51daf87c1dcbee +size 4362142888 diff --git a/model-00102-of-00106.safetensors b/model-00102-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..a31570d5425cc672211d29e5ff325a1fdd4125bb --- /dev/null +++ b/model-00102-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:5296b8d72f65f20b7329127094cd6a877269b5f6b6093b7e9a33ffbb359ca66b +size 4362142880 diff --git a/model-00103-of-00106.safetensors b/model-00103-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..b5ad96704e19b8d1c0c88ba00ffff0afc4a16345 --- /dev/null +++ b/model-00103-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:19e77e0709c38438955963a4fa24971c5998fed01de70d9fb5688807f2be0acf +size 4966188896 diff --git a/model-00104-of-00106.safetensors b/model-00104-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..cd1173b9e1f10eda5f48216e624078a9c82b850d --- /dev/null +++ b/model-00104-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:22b7768c709d0658e5b33930ba9ebf2c9eb2d91815e9375f75108582dfe879bd +size 4362142888 diff --git a/model-00105-of-00106.safetensors b/model-00105-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..5577ad520136ba41349da594466997155e3ce8ea --- /dev/null +++ b/model-00105-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:f7167570648b646407b34907d27e7ca19654f664a3c4e95e5d5e854bbb8bb624 +size 4362142880 diff --git a/model-00106-of-00106.safetensors b/model-00106-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..53e98cdea10876523b66e972dac9cea1b402e636 --- /dev/null +++ b/model-00106-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:0b3bbf6a33ad269241006dffb23030b74fd8d5c05b99adb79652d266e624ceaa +size 1988198968 diff --git a/model.safetensors.index.json b/model.safetensors.index.json new file mode 100644 index 0000000000000000000000000000000000000000..4205262bd427465e4a13b0284bd583c3db9c93cb --- /dev/null +++ b/model.safetensors.index.json @@ -0,0 +1,1270 @@ +{ + "metadata": { + "total_size": 481263648768 + }, + "weight_map": { + "lm_head.weight": "model-00106-of-00106.safetensors", + "model.embed_tokens.weight": "model-00001-of-00106.safetensors", + "model.layers.0.input_layernorm.weight": "model-00001-of-00106.safetensors", + "model.layers.0.mlp.down_proj.weight": "model-00001-of-00106.safetensors", + "model.layers.0.mlp.gate_proj.weight": "model-00001-of-00106.safetensors", + "model.layers.0.mlp.up_proj.weight": "model-00001-of-00106.safetensors", + "model.layers.0.post_attention_layernorm.weight": "model-00001-of-00106.safetensors", + "model.layers.0.self_attn.k_proj.weight": "model-00001-of-00106.safetensors", + "model.layers.0.self_attn.o_proj.weight": "model-00001-of-00106.safetensors", + "model.layers.0.self_attn.q_proj.weight": "model-00001-of-00106.safetensors", + "model.layers.0.self_attn.v_proj.weight": "model-00001-of-00106.safetensors", + "model.layers.1.input_layernorm.weight": "model-00002-of-00106.safetensors", + "model.layers.1.mlp.down_proj.weight": "model-00002-of-00106.safetensors", + "model.layers.1.mlp.gate_proj.weight": "model-00002-of-00106.safetensors", + "model.layers.1.mlp.up_proj.weight": "model-00002-of-00106.safetensors", + "model.layers.1.post_attention_layernorm.weight": "model-00002-of-00106.safetensors", + "model.layers.1.self_attn.k_proj.weight": "model-00001-of-00106.safetensors", + "model.layers.1.self_attn.o_proj.weight": "model-00002-of-00106.safetensors", + "model.layers.1.self_attn.q_proj.weight": "model-00001-of-00106.safetensors", + "model.layers.1.self_attn.v_proj.weight": "model-00001-of-00106.safetensors", + "model.layers.10.input_layernorm.weight": "model-00009-of-00106.safetensors", + "model.layers.10.mlp.down_proj.weight": "model-00009-of-00106.safetensors", + "model.layers.10.mlp.gate_proj.weight": "model-00008-of-00106.safetensors", + "model.layers.10.mlp.up_proj.weight": "model-00009-of-00106.safetensors", + "model.layers.10.post_attention_layernorm.weight": "model-00009-of-00106.safetensors", + "model.layers.10.self_attn.k_proj.weight": "model-00008-of-00106.safetensors", + "model.layers.10.self_attn.o_proj.weight": "model-00008-of-00106.safetensors", + "model.layers.10.self_attn.q_proj.weight": "model-00008-of-00106.safetensors", + "model.layers.10.self_attn.v_proj.weight": "model-00008-of-00106.safetensors", + "model.layers.100.input_layernorm.weight": "model-00076-of-00106.safetensors", + "model.layers.100.mlp.down_proj.weight": "model-00076-of-00106.safetensors", + "model.layers.100.mlp.gate_proj.weight": "model-00076-of-00106.safetensors", + "model.layers.100.mlp.up_proj.weight": "model-00076-of-00106.safetensors", + "model.layers.100.post_attention_layernorm.weight": "model-00076-of-00106.safetensors", + "model.layers.100.self_attn.k_proj.weight": "model-00076-of-00106.safetensors", + "model.layers.100.self_attn.o_proj.weight": "model-00076-of-00106.safetensors", + "model.layers.100.self_attn.q_proj.weight": "model-00076-of-00106.safetensors", + "model.layers.100.self_attn.v_proj.weight": "model-00076-of-00106.safetensors", + "model.layers.101.input_layernorm.weight": "model-00077-of-00106.safetensors", + "model.layers.101.mlp.down_proj.weight": "model-00077-of-00106.safetensors", + "model.layers.101.mlp.gate_proj.weight": "model-00077-of-00106.safetensors", + "model.layers.101.mlp.up_proj.weight": "model-00077-of-00106.safetensors", + "model.layers.101.post_attention_layernorm.weight": "model-00077-of-00106.safetensors", + "model.layers.101.self_attn.k_proj.weight": "model-00076-of-00106.safetensors", + "model.layers.101.self_attn.o_proj.weight": "model-00076-of-00106.safetensors", + "model.layers.101.self_attn.q_proj.weight": "model-00076-of-00106.safetensors", + "model.layers.101.self_attn.v_proj.weight": "model-00076-of-00106.safetensors", + "model.layers.102.input_layernorm.weight": "model-00078-of-00106.safetensors", + "model.layers.102.mlp.down_proj.weight": "model-00078-of-00106.safetensors", + "model.layers.102.mlp.gate_proj.weight": "model-00077-of-00106.safetensors", + "model.layers.102.mlp.up_proj.weight": "model-00078-of-00106.safetensors", + "model.layers.102.post_attention_layernorm.weight": "model-00078-of-00106.safetensors", + "model.layers.102.self_attn.k_proj.weight": "model-00077-of-00106.safetensors", + "model.layers.102.self_attn.o_proj.weight": "model-00077-of-00106.safetensors", + "model.layers.102.self_attn.q_proj.weight": "model-00077-of-00106.safetensors", + "model.layers.102.self_attn.v_proj.weight": "model-00077-of-00106.safetensors", + "model.layers.103.input_layernorm.weight": "model-00079-of-00106.safetensors", + "model.layers.103.mlp.down_proj.weight": "model-00079-of-00106.safetensors", + "model.layers.103.mlp.gate_proj.weight": "model-00078-of-00106.safetensors", + "model.layers.103.mlp.up_proj.weight": "model-00078-of-00106.safetensors", + "model.layers.103.post_attention_layernorm.weight": "model-00079-of-00106.safetensors", + "model.layers.103.self_attn.k_proj.weight": "model-00078-of-00106.safetensors", + "model.layers.103.self_attn.o_proj.weight": "model-00078-of-00106.safetensors", + "model.layers.103.self_attn.q_proj.weight": "model-00078-of-00106.safetensors", + "model.layers.103.self_attn.v_proj.weight": "model-00078-of-00106.safetensors", + "model.layers.104.input_layernorm.weight": "model-00079-of-00106.safetensors", + "model.layers.104.mlp.down_proj.weight": "model-00079-of-00106.safetensors", + "model.layers.104.mlp.gate_proj.weight": "model-00079-of-00106.safetensors", + "model.layers.104.mlp.up_proj.weight": "model-00079-of-00106.safetensors", + "model.layers.104.post_attention_layernorm.weight": "model-00079-of-00106.safetensors", + "model.layers.104.self_attn.k_proj.weight": "model-00079-of-00106.safetensors", + "model.layers.104.self_attn.o_proj.weight": "model-00079-of-00106.safetensors", + "model.layers.104.self_attn.q_proj.weight": "model-00079-of-00106.safetensors", + "model.layers.104.self_attn.v_proj.weight": "model-00079-of-00106.safetensors", + "model.layers.105.input_layernorm.weight": "model-00080-of-00106.safetensors", + "model.layers.105.mlp.down_proj.weight": "model-00080-of-00106.safetensors", + "model.layers.105.mlp.gate_proj.weight": "model-00080-of-00106.safetensors", + "model.layers.105.mlp.up_proj.weight": "model-00080-of-00106.safetensors", + "model.layers.105.post_attention_layernorm.weight": "model-00080-of-00106.safetensors", + "model.layers.105.self_attn.k_proj.weight": "model-00079-of-00106.safetensors", + "model.layers.105.self_attn.o_proj.weight": "model-00079-of-00106.safetensors", + "model.layers.105.self_attn.q_proj.weight": "model-00079-of-00106.safetensors", + "model.layers.105.self_attn.v_proj.weight": "model-00079-of-00106.safetensors", + "model.layers.106.input_layernorm.weight": "model-00081-of-00106.safetensors", + "model.layers.106.mlp.down_proj.weight": "model-00081-of-00106.safetensors", + "model.layers.106.mlp.gate_proj.weight": "model-00080-of-00106.safetensors", + "model.layers.106.mlp.up_proj.weight": "model-00081-of-00106.safetensors", + "model.layers.106.post_attention_layernorm.weight": "model-00081-of-00106.safetensors", + "model.layers.106.self_attn.k_proj.weight": "model-00080-of-00106.safetensors", + "model.layers.106.self_attn.o_proj.weight": "model-00080-of-00106.safetensors", + "model.layers.106.self_attn.q_proj.weight": "model-00080-of-00106.safetensors", + "model.layers.106.self_attn.v_proj.weight": "model-00080-of-00106.safetensors", + "model.layers.107.input_layernorm.weight": "model-00082-of-00106.safetensors", + "model.layers.107.mlp.down_proj.weight": "model-00082-of-00106.safetensors", + "model.layers.107.mlp.gate_proj.weight": "model-00081-of-00106.safetensors", + "model.layers.107.mlp.up_proj.weight": "model-00081-of-00106.safetensors", + "model.layers.107.post_attention_layernorm.weight": "model-00082-of-00106.safetensors", + "model.layers.107.self_attn.k_proj.weight": "model-00081-of-00106.safetensors", + "model.layers.107.self_attn.o_proj.weight": "model-00081-of-00106.safetensors", + "model.layers.107.self_attn.q_proj.weight": "model-00081-of-00106.safetensors", + "model.layers.107.self_attn.v_proj.weight": "model-00081-of-00106.safetensors", + "model.layers.108.input_layernorm.weight": "model-00082-of-00106.safetensors", + "model.layers.108.mlp.down_proj.weight": "model-00082-of-00106.safetensors", + "model.layers.108.mlp.gate_proj.weight": "model-00082-of-00106.safetensors", + "model.layers.108.mlp.up_proj.weight": "model-00082-of-00106.safetensors", + "model.layers.108.post_attention_layernorm.weight": "model-00082-of-00106.safetensors", + "model.layers.108.self_attn.k_proj.weight": "model-00082-of-00106.safetensors", + "model.layers.108.self_attn.o_proj.weight": "model-00082-of-00106.safetensors", + "model.layers.108.self_attn.q_proj.weight": "model-00082-of-00106.safetensors", + "model.layers.108.self_attn.v_proj.weight": "model-00082-of-00106.safetensors", + "model.layers.109.input_layernorm.weight": "model-00083-of-00106.safetensors", + "model.layers.109.mlp.down_proj.weight": "model-00083-of-00106.safetensors", + "model.layers.109.mlp.gate_proj.weight": "model-00083-of-00106.safetensors", + "model.layers.109.mlp.up_proj.weight": "model-00083-of-00106.safetensors", + "model.layers.109.post_attention_layernorm.weight": "model-00083-of-00106.safetensors", + "model.layers.109.self_attn.k_proj.weight": "model-00082-of-00106.safetensors", + "model.layers.109.self_attn.o_proj.weight": "model-00082-of-00106.safetensors", + "model.layers.109.self_attn.q_proj.weight": "model-00082-of-00106.safetensors", + "model.layers.109.self_attn.v_proj.weight": "model-00082-of-00106.safetensors", + "model.layers.11.input_layernorm.weight": "model-00010-of-00106.safetensors", + "model.layers.11.mlp.down_proj.weight": "model-00010-of-00106.safetensors", + "model.layers.11.mlp.gate_proj.weight": "model-00009-of-00106.safetensors", + "model.layers.11.mlp.up_proj.weight": "model-00009-of-00106.safetensors", + "model.layers.11.post_attention_layernorm.weight": "model-00010-of-00106.safetensors", + "model.layers.11.self_attn.k_proj.weight": "model-00009-of-00106.safetensors", + "model.layers.11.self_attn.o_proj.weight": "model-00009-of-00106.safetensors", + "model.layers.11.self_attn.q_proj.weight": "model-00009-of-00106.safetensors", + "model.layers.11.self_attn.v_proj.weight": "model-00009-of-00106.safetensors", + "model.layers.110.input_layernorm.weight": "model-00084-of-00106.safetensors", + "model.layers.110.mlp.down_proj.weight": "model-00084-of-00106.safetensors", + "model.layers.110.mlp.gate_proj.weight": "model-00083-of-00106.safetensors", + "model.layers.110.mlp.up_proj.weight": "model-00084-of-00106.safetensors", + "model.layers.110.post_attention_layernorm.weight": "model-00084-of-00106.safetensors", + "model.layers.110.self_attn.k_proj.weight": "model-00083-of-00106.safetensors", + "model.layers.110.self_attn.o_proj.weight": "model-00083-of-00106.safetensors", + "model.layers.110.self_attn.q_proj.weight": "model-00083-of-00106.safetensors", + "model.layers.110.self_attn.v_proj.weight": "model-00083-of-00106.safetensors", + "model.layers.111.input_layernorm.weight": "model-00085-of-00106.safetensors", + "model.layers.111.mlp.down_proj.weight": "model-00085-of-00106.safetensors", + "model.layers.111.mlp.gate_proj.weight": "model-00084-of-00106.safetensors", + "model.layers.111.mlp.up_proj.weight": "model-00084-of-00106.safetensors", + "model.layers.111.post_attention_layernorm.weight": "model-00085-of-00106.safetensors", + "model.layers.111.self_attn.k_proj.weight": "model-00084-of-00106.safetensors", + "model.layers.111.self_attn.o_proj.weight": "model-00084-of-00106.safetensors", + "model.layers.111.self_attn.q_proj.weight": "model-00084-of-00106.safetensors", + "model.layers.111.self_attn.v_proj.weight": "model-00084-of-00106.safetensors", + "model.layers.112.input_layernorm.weight": "model-00085-of-00106.safetensors", + "model.layers.112.mlp.down_proj.weight": "model-00085-of-00106.safetensors", + "model.layers.112.mlp.gate_proj.weight": "model-00085-of-00106.safetensors", + "model.layers.112.mlp.up_proj.weight": "model-00085-of-00106.safetensors", + "model.layers.112.post_attention_layernorm.weight": "model-00085-of-00106.safetensors", + "model.layers.112.self_attn.k_proj.weight": "model-00085-of-00106.safetensors", + "model.layers.112.self_attn.o_proj.weight": "model-00085-of-00106.safetensors", + "model.layers.112.self_attn.q_proj.weight": "model-00085-of-00106.safetensors", + "model.layers.112.self_attn.v_proj.weight": "model-00085-of-00106.safetensors", + "model.layers.113.input_layernorm.weight": "model-00086-of-00106.safetensors", + "model.layers.113.mlp.down_proj.weight": "model-00086-of-00106.safetensors", + "model.layers.113.mlp.gate_proj.weight": "model-00086-of-00106.safetensors", + "model.layers.113.mlp.up_proj.weight": "model-00086-of-00106.safetensors", + "model.layers.113.post_attention_layernorm.weight": "model-00086-of-00106.safetensors", + "model.layers.113.self_attn.k_proj.weight": "model-00085-of-00106.safetensors", + "model.layers.113.self_attn.o_proj.weight": "model-00085-of-00106.safetensors", + "model.layers.113.self_attn.q_proj.weight": "model-00085-of-00106.safetensors", + "model.layers.113.self_attn.v_proj.weight": "model-00085-of-00106.safetensors", + "model.layers.114.input_layernorm.weight": "model-00087-of-00106.safetensors", + "model.layers.114.mlp.down_proj.weight": "model-00087-of-00106.safetensors", + "model.layers.114.mlp.gate_proj.weight": "model-00086-of-00106.safetensors", + "model.layers.114.mlp.up_proj.weight": "model-00087-of-00106.safetensors", + "model.layers.114.post_attention_layernorm.weight": "model-00087-of-00106.safetensors", + "model.layers.114.self_attn.k_proj.weight": "model-00086-of-00106.safetensors", + "model.layers.114.self_attn.o_proj.weight": "model-00086-of-00106.safetensors", + "model.layers.114.self_attn.q_proj.weight": "model-00086-of-00106.safetensors", + "model.layers.114.self_attn.v_proj.weight": "model-00086-of-00106.safetensors", + "model.layers.115.input_layernorm.weight": "model-00088-of-00106.safetensors", + "model.layers.115.mlp.down_proj.weight": "model-00088-of-00106.safetensors", + "model.layers.115.mlp.gate_proj.weight": "model-00087-of-00106.safetensors", + "model.layers.115.mlp.up_proj.weight": "model-00087-of-00106.safetensors", + "model.layers.115.post_attention_layernorm.weight": "model-00088-of-00106.safetensors", + "model.layers.115.self_attn.k_proj.weight": "model-00087-of-00106.safetensors", + "model.layers.115.self_attn.o_proj.weight": "model-00087-of-00106.safetensors", + "model.layers.115.self_attn.q_proj.weight": "model-00087-of-00106.safetensors", + "model.layers.115.self_attn.v_proj.weight": "model-00087-of-00106.safetensors", + "model.layers.116.input_layernorm.weight": "model-00088-of-00106.safetensors", + "model.layers.116.mlp.down_proj.weight": "model-00088-of-00106.safetensors", + "model.layers.116.mlp.gate_proj.weight": "model-00088-of-00106.safetensors", + "model.layers.116.mlp.up_proj.weight": "model-00088-of-00106.safetensors", + "model.layers.116.post_attention_layernorm.weight": "model-00088-of-00106.safetensors", + "model.layers.116.self_attn.k_proj.weight": "model-00088-of-00106.safetensors", + "model.layers.116.self_attn.o_proj.weight": "model-00088-of-00106.safetensors", + "model.layers.116.self_attn.q_proj.weight": "model-00088-of-00106.safetensors", + "model.layers.116.self_attn.v_proj.weight": "model-00088-of-00106.safetensors", + "model.layers.117.input_layernorm.weight": "model-00089-of-00106.safetensors", + "model.layers.117.mlp.down_proj.weight": "model-00089-of-00106.safetensors", + "model.layers.117.mlp.gate_proj.weight": "model-00089-of-00106.safetensors", + "model.layers.117.mlp.up_proj.weight": "model-00089-of-00106.safetensors", + "model.layers.117.post_attention_layernorm.weight": "model-00089-of-00106.safetensors", + "model.layers.117.self_attn.k_proj.weight": "model-00088-of-00106.safetensors", + "model.layers.117.self_attn.o_proj.weight": "model-00088-of-00106.safetensors", + "model.layers.117.self_attn.q_proj.weight": "model-00088-of-00106.safetensors", + "model.layers.117.self_attn.v_proj.weight": "model-00088-of-00106.safetensors", + "model.layers.118.input_layernorm.weight": "model-00090-of-00106.safetensors", + "model.layers.118.mlp.down_proj.weight": "model-00090-of-00106.safetensors", + "model.layers.118.mlp.gate_proj.weight": "model-00089-of-00106.safetensors", + "model.layers.118.mlp.up_proj.weight": "model-00090-of-00106.safetensors", + "model.layers.118.post_attention_layernorm.weight": "model-00090-of-00106.safetensors", + "model.layers.118.self_attn.k_proj.weight": "model-00089-of-00106.safetensors", + "model.layers.118.self_attn.o_proj.weight": "model-00089-of-00106.safetensors", + "model.layers.118.self_attn.q_proj.weight": "model-00089-of-00106.safetensors", + "model.layers.118.self_attn.v_proj.weight": "model-00089-of-00106.safetensors", + "model.layers.119.input_layernorm.weight": "model-00091-of-00106.safetensors", + "model.layers.119.mlp.down_proj.weight": "model-00091-of-00106.safetensors", + "model.layers.119.mlp.gate_proj.weight": "model-00090-of-00106.safetensors", + "model.layers.119.mlp.up_proj.weight": "model-00090-of-00106.safetensors", + "model.layers.119.post_attention_layernorm.weight": "model-00091-of-00106.safetensors", + "model.layers.119.self_attn.k_proj.weight": "model-00090-of-00106.safetensors", + "model.layers.119.self_attn.o_proj.weight": "model-00090-of-00106.safetensors", + "model.layers.119.self_attn.q_proj.weight": "model-00090-of-00106.safetensors", + "model.layers.119.self_attn.v_proj.weight": "model-00090-of-00106.safetensors", + "model.layers.12.input_layernorm.weight": "model-00010-of-00106.safetensors", + "model.layers.12.mlp.down_proj.weight": "model-00010-of-00106.safetensors", + "model.layers.12.mlp.gate_proj.weight": "model-00010-of-00106.safetensors", + "model.layers.12.mlp.up_proj.weight": "model-00010-of-00106.safetensors", + "model.layers.12.post_attention_layernorm.weight": "model-00010-of-00106.safetensors", + "model.layers.12.self_attn.k_proj.weight": "model-00010-of-00106.safetensors", + "model.layers.12.self_attn.o_proj.weight": "model-00010-of-00106.safetensors", + "model.layers.12.self_attn.q_proj.weight": "model-00010-of-00106.safetensors", + "model.layers.12.self_attn.v_proj.weight": "model-00010-of-00106.safetensors", + "model.layers.120.input_layernorm.weight": "model-00091-of-00106.safetensors", + "model.layers.120.mlp.down_proj.weight": "model-00091-of-00106.safetensors", + "model.layers.120.mlp.gate_proj.weight": "model-00091-of-00106.safetensors", + "model.layers.120.mlp.up_proj.weight": "model-00091-of-00106.safetensors", + "model.layers.120.post_attention_layernorm.weight": "model-00091-of-00106.safetensors", + "model.layers.120.self_attn.k_proj.weight": "model-00091-of-00106.safetensors", + "model.layers.120.self_attn.o_proj.weight": "model-00091-of-00106.safetensors", + "model.layers.120.self_attn.q_proj.weight": "model-00091-of-00106.safetensors", + "model.layers.120.self_attn.v_proj.weight": "model-00091-of-00106.safetensors", + "model.layers.121.input_layernorm.weight": "model-00092-of-00106.safetensors", + "model.layers.121.mlp.down_proj.weight": "model-00092-of-00106.safetensors", + "model.layers.121.mlp.gate_proj.weight": "model-00092-of-00106.safetensors", + "model.layers.121.mlp.up_proj.weight": "model-00092-of-00106.safetensors", + "model.layers.121.post_attention_layernorm.weight": "model-00092-of-00106.safetensors", + "model.layers.121.self_attn.k_proj.weight": "model-00091-of-00106.safetensors", + "model.layers.121.self_attn.o_proj.weight": "model-00091-of-00106.safetensors", + "model.layers.121.self_attn.q_proj.weight": "model-00091-of-00106.safetensors", + "model.layers.121.self_attn.v_proj.weight": "model-00091-of-00106.safetensors", + "model.layers.122.input_layernorm.weight": "model-00093-of-00106.safetensors", + "model.layers.122.mlp.down_proj.weight": "model-00093-of-00106.safetensors", + "model.layers.122.mlp.gate_proj.weight": "model-00092-of-00106.safetensors", + "model.layers.122.mlp.up_proj.weight": "model-00093-of-00106.safetensors", + "model.layers.122.post_attention_layernorm.weight": "model-00093-of-00106.safetensors", + "model.layers.122.self_attn.k_proj.weight": "model-00092-of-00106.safetensors", + "model.layers.122.self_attn.o_proj.weight": "model-00092-of-00106.safetensors", + "model.layers.122.self_attn.q_proj.weight": "model-00092-of-00106.safetensors", + "model.layers.122.self_attn.v_proj.weight": "model-00092-of-00106.safetensors", + "model.layers.123.input_layernorm.weight": "model-00094-of-00106.safetensors", + "model.layers.123.mlp.down_proj.weight": "model-00094-of-00106.safetensors", + "model.layers.123.mlp.gate_proj.weight": "model-00093-of-00106.safetensors", + "model.layers.123.mlp.up_proj.weight": "model-00093-of-00106.safetensors", + "model.layers.123.post_attention_layernorm.weight": "model-00094-of-00106.safetensors", + "model.layers.123.self_attn.k_proj.weight": "model-00093-of-00106.safetensors", + "model.layers.123.self_attn.o_proj.weight": "model-00093-of-00106.safetensors", + "model.layers.123.self_attn.q_proj.weight": "model-00093-of-00106.safetensors", + "model.layers.123.self_attn.v_proj.weight": "model-00093-of-00106.safetensors", + "model.layers.124.input_layernorm.weight": "model-00094-of-00106.safetensors", + "model.layers.124.mlp.down_proj.weight": "model-00094-of-00106.safetensors", + "model.layers.124.mlp.gate_proj.weight": "model-00094-of-00106.safetensors", + "model.layers.124.mlp.up_proj.weight": "model-00094-of-00106.safetensors", + "model.layers.124.post_attention_layernorm.weight": "model-00094-of-00106.safetensors", + "model.layers.124.self_attn.k_proj.weight": "model-00094-of-00106.safetensors", + "model.layers.124.self_attn.o_proj.weight": "model-00094-of-00106.safetensors", + "model.layers.124.self_attn.q_proj.weight": "model-00094-of-00106.safetensors", + "model.layers.124.self_attn.v_proj.weight": "model-00094-of-00106.safetensors", + "model.layers.125.input_layernorm.weight": "model-00095-of-00106.safetensors", + "model.layers.125.mlp.down_proj.weight": "model-00095-of-00106.safetensors", + "model.layers.125.mlp.gate_proj.weight": "model-00095-of-00106.safetensors", + "model.layers.125.mlp.up_proj.weight": "model-00095-of-00106.safetensors", + "model.layers.125.post_attention_layernorm.weight": "model-00095-of-00106.safetensors", + "model.layers.125.self_attn.k_proj.weight": "model-00094-of-00106.safetensors", + "model.layers.125.self_attn.o_proj.weight": "model-00094-of-00106.safetensors", + "model.layers.125.self_attn.q_proj.weight": "model-00094-of-00106.safetensors", + "model.layers.125.self_attn.v_proj.weight": "model-00094-of-00106.safetensors", + "model.layers.126.input_layernorm.weight": "model-00096-of-00106.safetensors", + "model.layers.126.mlp.down_proj.weight": "model-00096-of-00106.safetensors", + "model.layers.126.mlp.gate_proj.weight": "model-00095-of-00106.safetensors", + "model.layers.126.mlp.up_proj.weight": "model-00096-of-00106.safetensors", + "model.layers.126.post_attention_layernorm.weight": "model-00096-of-00106.safetensors", + "model.layers.126.self_attn.k_proj.weight": "model-00095-of-00106.safetensors", + "model.layers.126.self_attn.o_proj.weight": "model-00095-of-00106.safetensors", + "model.layers.126.self_attn.q_proj.weight": "model-00095-of-00106.safetensors", + "model.layers.126.self_attn.v_proj.weight": "model-00095-of-00106.safetensors", + "model.layers.127.input_layernorm.weight": "model-00097-of-00106.safetensors", + "model.layers.127.mlp.down_proj.weight": "model-00097-of-00106.safetensors", + "model.layers.127.mlp.gate_proj.weight": "model-00096-of-00106.safetensors", + "model.layers.127.mlp.up_proj.weight": "model-00096-of-00106.safetensors", + "model.layers.127.post_attention_layernorm.weight": "model-00097-of-00106.safetensors", + "model.layers.127.self_attn.k_proj.weight": "model-00096-of-00106.safetensors", + "model.layers.127.self_attn.o_proj.weight": "model-00096-of-00106.safetensors", + "model.layers.127.self_attn.q_proj.weight": "model-00096-of-00106.safetensors", + "model.layers.127.self_attn.v_proj.weight": "model-00096-of-00106.safetensors", + "model.layers.128.input_layernorm.weight": "model-00097-of-00106.safetensors", + "model.layers.128.mlp.down_proj.weight": "model-00097-of-00106.safetensors", + "model.layers.128.mlp.gate_proj.weight": "model-00097-of-00106.safetensors", + "model.layers.128.mlp.up_proj.weight": "model-00097-of-00106.safetensors", + "model.layers.128.post_attention_layernorm.weight": "model-00097-of-00106.safetensors", + "model.layers.128.self_attn.k_proj.weight": "model-00097-of-00106.safetensors", + "model.layers.128.self_attn.o_proj.weight": "model-00097-of-00106.safetensors", + "model.layers.128.self_attn.q_proj.weight": "model-00097-of-00106.safetensors", + "model.layers.128.self_attn.v_proj.weight": "model-00097-of-00106.safetensors", + "model.layers.129.input_layernorm.weight": "model-00098-of-00106.safetensors", + "model.layers.129.mlp.down_proj.weight": "model-00098-of-00106.safetensors", + "model.layers.129.mlp.gate_proj.weight": "model-00098-of-00106.safetensors", + "model.layers.129.mlp.up_proj.weight": "model-00098-of-00106.safetensors", + "model.layers.129.post_attention_layernorm.weight": "model-00098-of-00106.safetensors", + "model.layers.129.self_attn.k_proj.weight": "model-00097-of-00106.safetensors", + "model.layers.129.self_attn.o_proj.weight": "model-00097-of-00106.safetensors", + "model.layers.129.self_attn.q_proj.weight": "model-00097-of-00106.safetensors", + "model.layers.129.self_attn.v_proj.weight": "model-00097-of-00106.safetensors", + "model.layers.13.input_layernorm.weight": "model-00011-of-00106.safetensors", + "model.layers.13.mlp.down_proj.weight": "model-00011-of-00106.safetensors", + "model.layers.13.mlp.gate_proj.weight": "model-00011-of-00106.safetensors", + "model.layers.13.mlp.up_proj.weight": "model-00011-of-00106.safetensors", + "model.layers.13.post_attention_layernorm.weight": "model-00011-of-00106.safetensors", + "model.layers.13.self_attn.k_proj.weight": "model-00010-of-00106.safetensors", + "model.layers.13.self_attn.o_proj.weight": "model-00010-of-00106.safetensors", + "model.layers.13.self_attn.q_proj.weight": "model-00010-of-00106.safetensors", + "model.layers.13.self_attn.v_proj.weight": "model-00010-of-00106.safetensors", + "model.layers.130.input_layernorm.weight": "model-00099-of-00106.safetensors", + "model.layers.130.mlp.down_proj.weight": "model-00099-of-00106.safetensors", + "model.layers.130.mlp.gate_proj.weight": "model-00098-of-00106.safetensors", + "model.layers.130.mlp.up_proj.weight": "model-00099-of-00106.safetensors", + "model.layers.130.post_attention_layernorm.weight": "model-00099-of-00106.safetensors", + "model.layers.130.self_attn.k_proj.weight": "model-00098-of-00106.safetensors", + "model.layers.130.self_attn.o_proj.weight": "model-00098-of-00106.safetensors", + "model.layers.130.self_attn.q_proj.weight": "model-00098-of-00106.safetensors", + "model.layers.130.self_attn.v_proj.weight": "model-00098-of-00106.safetensors", + "model.layers.131.input_layernorm.weight": "model-00100-of-00106.safetensors", + "model.layers.131.mlp.down_proj.weight": "model-00100-of-00106.safetensors", + "model.layers.131.mlp.gate_proj.weight": "model-00099-of-00106.safetensors", + "model.layers.131.mlp.up_proj.weight": "model-00099-of-00106.safetensors", + "model.layers.131.post_attention_layernorm.weight": "model-00100-of-00106.safetensors", + "model.layers.131.self_attn.k_proj.weight": "model-00099-of-00106.safetensors", + "model.layers.131.self_attn.o_proj.weight": "model-00099-of-00106.safetensors", + "model.layers.131.self_attn.q_proj.weight": "model-00099-of-00106.safetensors", + "model.layers.131.self_attn.v_proj.weight": "model-00099-of-00106.safetensors", + "model.layers.132.input_layernorm.weight": "model-00100-of-00106.safetensors", + "model.layers.132.mlp.down_proj.weight": "model-00100-of-00106.safetensors", + "model.layers.132.mlp.gate_proj.weight": "model-00100-of-00106.safetensors", + "model.layers.132.mlp.up_proj.weight": "model-00100-of-00106.safetensors", + "model.layers.132.post_attention_layernorm.weight": "model-00100-of-00106.safetensors", + "model.layers.132.self_attn.k_proj.weight": "model-00100-of-00106.safetensors", + "model.layers.132.self_attn.o_proj.weight": "model-00100-of-00106.safetensors", + "model.layers.132.self_attn.q_proj.weight": "model-00100-of-00106.safetensors", + "model.layers.132.self_attn.v_proj.weight": "model-00100-of-00106.safetensors", + "model.layers.133.input_layernorm.weight": "model-00101-of-00106.safetensors", + "model.layers.133.mlp.down_proj.weight": "model-00101-of-00106.safetensors", + "model.layers.133.mlp.gate_proj.weight": "model-00101-of-00106.safetensors", + "model.layers.133.mlp.up_proj.weight": "model-00101-of-00106.safetensors", + "model.layers.133.post_attention_layernorm.weight": "model-00101-of-00106.safetensors", + "model.layers.133.self_attn.k_proj.weight": "model-00100-of-00106.safetensors", + "model.layers.133.self_attn.o_proj.weight": "model-00100-of-00106.safetensors", + "model.layers.133.self_attn.q_proj.weight": "model-00100-of-00106.safetensors", + "model.layers.133.self_attn.v_proj.weight": "model-00100-of-00106.safetensors", + "model.layers.134.input_layernorm.weight": "model-00102-of-00106.safetensors", + "model.layers.134.mlp.down_proj.weight": "model-00102-of-00106.safetensors", + "model.layers.134.mlp.gate_proj.weight": "model-00101-of-00106.safetensors", + "model.layers.134.mlp.up_proj.weight": "model-00102-of-00106.safetensors", + "model.layers.134.post_attention_layernorm.weight": "model-00102-of-00106.safetensors", + "model.layers.134.self_attn.k_proj.weight": "model-00101-of-00106.safetensors", + "model.layers.134.self_attn.o_proj.weight": "model-00101-of-00106.safetensors", + "model.layers.134.self_attn.q_proj.weight": "model-00101-of-00106.safetensors", + "model.layers.134.self_attn.v_proj.weight": "model-00101-of-00106.safetensors", + "model.layers.135.input_layernorm.weight": "model-00103-of-00106.safetensors", + "model.layers.135.mlp.down_proj.weight": "model-00103-of-00106.safetensors", + "model.layers.135.mlp.gate_proj.weight": "model-00102-of-00106.safetensors", + "model.layers.135.mlp.up_proj.weight": "model-00102-of-00106.safetensors", + "model.layers.135.post_attention_layernorm.weight": "model-00103-of-00106.safetensors", + "model.layers.135.self_attn.k_proj.weight": "model-00102-of-00106.safetensors", + "model.layers.135.self_attn.o_proj.weight": "model-00102-of-00106.safetensors", + "model.layers.135.self_attn.q_proj.weight": "model-00102-of-00106.safetensors", + "model.layers.135.self_attn.v_proj.weight": "model-00102-of-00106.safetensors", + "model.layers.136.input_layernorm.weight": "model-00103-of-00106.safetensors", + "model.layers.136.mlp.down_proj.weight": "model-00103-of-00106.safetensors", + "model.layers.136.mlp.gate_proj.weight": "model-00103-of-00106.safetensors", + "model.layers.136.mlp.up_proj.weight": "model-00103-of-00106.safetensors", + "model.layers.136.post_attention_layernorm.weight": "model-00103-of-00106.safetensors", + "model.layers.136.self_attn.k_proj.weight": "model-00103-of-00106.safetensors", + "model.layers.136.self_attn.o_proj.weight": "model-00103-of-00106.safetensors", + "model.layers.136.self_attn.q_proj.weight": "model-00103-of-00106.safetensors", + "model.layers.136.self_attn.v_proj.weight": "model-00103-of-00106.safetensors", + "model.layers.137.input_layernorm.weight": "model-00104-of-00106.safetensors", + "model.layers.137.mlp.down_proj.weight": "model-00104-of-00106.safetensors", + "model.layers.137.mlp.gate_proj.weight": "model-00104-of-00106.safetensors", + "model.layers.137.mlp.up_proj.weight": "model-00104-of-00106.safetensors", + "model.layers.137.post_attention_layernorm.weight": "model-00104-of-00106.safetensors", + "model.layers.137.self_attn.k_proj.weight": "model-00103-of-00106.safetensors", + "model.layers.137.self_attn.o_proj.weight": "model-00103-of-00106.safetensors", + "model.layers.137.self_attn.q_proj.weight": "model-00103-of-00106.safetensors", + "model.layers.137.self_attn.v_proj.weight": "model-00103-of-00106.safetensors", + "model.layers.138.input_layernorm.weight": "model-00105-of-00106.safetensors", + "model.layers.138.mlp.down_proj.weight": "model-00105-of-00106.safetensors", + "model.layers.138.mlp.gate_proj.weight": "model-00104-of-00106.safetensors", + "model.layers.138.mlp.up_proj.weight": "model-00105-of-00106.safetensors", + "model.layers.138.post_attention_layernorm.weight": "model-00105-of-00106.safetensors", + "model.layers.138.self_attn.k_proj.weight": "model-00104-of-00106.safetensors", + "model.layers.138.self_attn.o_proj.weight": "model-00104-of-00106.safetensors", + "model.layers.138.self_attn.q_proj.weight": "model-00104-of-00106.safetensors", + "model.layers.138.self_attn.v_proj.weight": "model-00104-of-00106.safetensors", + "model.layers.139.input_layernorm.weight": "model-00106-of-00106.safetensors", + "model.layers.139.mlp.down_proj.weight": "model-00106-of-00106.safetensors", + "model.layers.139.mlp.gate_proj.weight": "model-00105-of-00106.safetensors", + "model.layers.139.mlp.up_proj.weight": "model-00105-of-00106.safetensors", + "model.layers.139.post_attention_layernorm.weight": "model-00106-of-00106.safetensors", + "model.layers.139.self_attn.k_proj.weight": "model-00105-of-00106.safetensors", + "model.layers.139.self_attn.o_proj.weight": "model-00105-of-00106.safetensors", + "model.layers.139.self_attn.q_proj.weight": "model-00105-of-00106.safetensors", + "model.layers.139.self_attn.v_proj.weight": "model-00105-of-00106.safetensors", + "model.layers.14.input_layernorm.weight": "model-00012-of-00106.safetensors", + "model.layers.14.mlp.down_proj.weight": "model-00012-of-00106.safetensors", + "model.layers.14.mlp.gate_proj.weight": "model-00011-of-00106.safetensors", + "model.layers.14.mlp.up_proj.weight": "model-00012-of-00106.safetensors", + "model.layers.14.post_attention_layernorm.weight": "model-00012-of-00106.safetensors", + "model.layers.14.self_attn.k_proj.weight": "model-00011-of-00106.safetensors", + "model.layers.14.self_attn.o_proj.weight": "model-00011-of-00106.safetensors", + "model.layers.14.self_attn.q_proj.weight": "model-00011-of-00106.safetensors", + "model.layers.14.self_attn.v_proj.weight": "model-00011-of-00106.safetensors", + "model.layers.15.input_layernorm.weight": "model-00013-of-00106.safetensors", + "model.layers.15.mlp.down_proj.weight": "model-00013-of-00106.safetensors", + "model.layers.15.mlp.gate_proj.weight": "model-00012-of-00106.safetensors", + "model.layers.15.mlp.up_proj.weight": "model-00012-of-00106.safetensors", + "model.layers.15.post_attention_layernorm.weight": "model-00013-of-00106.safetensors", + "model.layers.15.self_attn.k_proj.weight": "model-00012-of-00106.safetensors", + "model.layers.15.self_attn.o_proj.weight": "model-00012-of-00106.safetensors", + "model.layers.15.self_attn.q_proj.weight": "model-00012-of-00106.safetensors", + "model.layers.15.self_attn.v_proj.weight": "model-00012-of-00106.safetensors", + "model.layers.16.input_layernorm.weight": "model-00013-of-00106.safetensors", + "model.layers.16.mlp.down_proj.weight": "model-00013-of-00106.safetensors", + "model.layers.16.mlp.gate_proj.weight": "model-00013-of-00106.safetensors", + "model.layers.16.mlp.up_proj.weight": "model-00013-of-00106.safetensors", + "model.layers.16.post_attention_layernorm.weight": "model-00013-of-00106.safetensors", + "model.layers.16.self_attn.k_proj.weight": "model-00013-of-00106.safetensors", + "model.layers.16.self_attn.o_proj.weight": "model-00013-of-00106.safetensors", + "model.layers.16.self_attn.q_proj.weight": "model-00013-of-00106.safetensors", + "model.layers.16.self_attn.v_proj.weight": "model-00013-of-00106.safetensors", + "model.layers.17.input_layernorm.weight": "model-00014-of-00106.safetensors", + "model.layers.17.mlp.down_proj.weight": "model-00014-of-00106.safetensors", + "model.layers.17.mlp.gate_proj.weight": "model-00014-of-00106.safetensors", + "model.layers.17.mlp.up_proj.weight": "model-00014-of-00106.safetensors", + "model.layers.17.post_attention_layernorm.weight": "model-00014-of-00106.safetensors", + "model.layers.17.self_attn.k_proj.weight": "model-00013-of-00106.safetensors", + "model.layers.17.self_attn.o_proj.weight": "model-00013-of-00106.safetensors", + "model.layers.17.self_attn.q_proj.weight": "model-00013-of-00106.safetensors", + "model.layers.17.self_attn.v_proj.weight": "model-00013-of-00106.safetensors", + "model.layers.18.input_layernorm.weight": "model-00015-of-00106.safetensors", + "model.layers.18.mlp.down_proj.weight": "model-00015-of-00106.safetensors", + "model.layers.18.mlp.gate_proj.weight": "model-00014-of-00106.safetensors", + "model.layers.18.mlp.up_proj.weight": "model-00015-of-00106.safetensors", + "model.layers.18.post_attention_layernorm.weight": "model-00015-of-00106.safetensors", + "model.layers.18.self_attn.k_proj.weight": "model-00014-of-00106.safetensors", + "model.layers.18.self_attn.o_proj.weight": "model-00014-of-00106.safetensors", + "model.layers.18.self_attn.q_proj.weight": "model-00014-of-00106.safetensors", + "model.layers.18.self_attn.v_proj.weight": "model-00014-of-00106.safetensors", + "model.layers.19.input_layernorm.weight": "model-00016-of-00106.safetensors", + "model.layers.19.mlp.down_proj.weight": "model-00016-of-00106.safetensors", + "model.layers.19.mlp.gate_proj.weight": "model-00015-of-00106.safetensors", + "model.layers.19.mlp.up_proj.weight": "model-00015-of-00106.safetensors", + "model.layers.19.post_attention_layernorm.weight": "model-00016-of-00106.safetensors", + "model.layers.19.self_attn.k_proj.weight": "model-00015-of-00106.safetensors", + "model.layers.19.self_attn.o_proj.weight": "model-00015-of-00106.safetensors", + "model.layers.19.self_attn.q_proj.weight": "model-00015-of-00106.safetensors", + "model.layers.19.self_attn.v_proj.weight": "model-00015-of-00106.safetensors", + "model.layers.2.input_layernorm.weight": "model-00003-of-00106.safetensors", + "model.layers.2.mlp.down_proj.weight": "model-00003-of-00106.safetensors", + "model.layers.2.mlp.gate_proj.weight": "model-00002-of-00106.safetensors", + "model.layers.2.mlp.up_proj.weight": "model-00003-of-00106.safetensors", + "model.layers.2.post_attention_layernorm.weight": "model-00003-of-00106.safetensors", + "model.layers.2.self_attn.k_proj.weight": "model-00002-of-00106.safetensors", + "model.layers.2.self_attn.o_proj.weight": "model-00002-of-00106.safetensors", + "model.layers.2.self_attn.q_proj.weight": "model-00002-of-00106.safetensors", + "model.layers.2.self_attn.v_proj.weight": "model-00002-of-00106.safetensors", + "model.layers.20.input_layernorm.weight": "model-00016-of-00106.safetensors", + "model.layers.20.mlp.down_proj.weight": "model-00016-of-00106.safetensors", + "model.layers.20.mlp.gate_proj.weight": "model-00016-of-00106.safetensors", + "model.layers.20.mlp.up_proj.weight": "model-00016-of-00106.safetensors", + "model.layers.20.post_attention_layernorm.weight": "model-00016-of-00106.safetensors", + "model.layers.20.self_attn.k_proj.weight": "model-00016-of-00106.safetensors", + "model.layers.20.self_attn.o_proj.weight": "model-00016-of-00106.safetensors", + "model.layers.20.self_attn.q_proj.weight": "model-00016-of-00106.safetensors", + "model.layers.20.self_attn.v_proj.weight": "model-00016-of-00106.safetensors", + "model.layers.21.input_layernorm.weight": "model-00017-of-00106.safetensors", + "model.layers.21.mlp.down_proj.weight": "model-00017-of-00106.safetensors", + "model.layers.21.mlp.gate_proj.weight": "model-00017-of-00106.safetensors", + "model.layers.21.mlp.up_proj.weight": "model-00017-of-00106.safetensors", + "model.layers.21.post_attention_layernorm.weight": "model-00017-of-00106.safetensors", + "model.layers.21.self_attn.k_proj.weight": "model-00016-of-00106.safetensors", + "model.layers.21.self_attn.o_proj.weight": "model-00016-of-00106.safetensors", + "model.layers.21.self_attn.q_proj.weight": "model-00016-of-00106.safetensors", + "model.layers.21.self_attn.v_proj.weight": "model-00016-of-00106.safetensors", + "model.layers.22.input_layernorm.weight": "model-00018-of-00106.safetensors", + "model.layers.22.mlp.down_proj.weight": "model-00018-of-00106.safetensors", + "model.layers.22.mlp.gate_proj.weight": "model-00017-of-00106.safetensors", + "model.layers.22.mlp.up_proj.weight": "model-00018-of-00106.safetensors", + "model.layers.22.post_attention_layernorm.weight": "model-00018-of-00106.safetensors", + "model.layers.22.self_attn.k_proj.weight": "model-00017-of-00106.safetensors", + "model.layers.22.self_attn.o_proj.weight": "model-00017-of-00106.safetensors", + "model.layers.22.self_attn.q_proj.weight": "model-00017-of-00106.safetensors", + "model.layers.22.self_attn.v_proj.weight": "model-00017-of-00106.safetensors", + "model.layers.23.input_layernorm.weight": "model-00019-of-00106.safetensors", + "model.layers.23.mlp.down_proj.weight": "model-00019-of-00106.safetensors", + "model.layers.23.mlp.gate_proj.weight": "model-00018-of-00106.safetensors", + "model.layers.23.mlp.up_proj.weight": "model-00018-of-00106.safetensors", + "model.layers.23.post_attention_layernorm.weight": "model-00019-of-00106.safetensors", + "model.layers.23.self_attn.k_proj.weight": "model-00018-of-00106.safetensors", + "model.layers.23.self_attn.o_proj.weight": "model-00018-of-00106.safetensors", + "model.layers.23.self_attn.q_proj.weight": "model-00018-of-00106.safetensors", + "model.layers.23.self_attn.v_proj.weight": "model-00018-of-00106.safetensors", + "model.layers.24.input_layernorm.weight": "model-00019-of-00106.safetensors", + "model.layers.24.mlp.down_proj.weight": "model-00019-of-00106.safetensors", + "model.layers.24.mlp.gate_proj.weight": "model-00019-of-00106.safetensors", + "model.layers.24.mlp.up_proj.weight": "model-00019-of-00106.safetensors", + "model.layers.24.post_attention_layernorm.weight": "model-00019-of-00106.safetensors", + "model.layers.24.self_attn.k_proj.weight": "model-00019-of-00106.safetensors", + "model.layers.24.self_attn.o_proj.weight": "model-00019-of-00106.safetensors", + "model.layers.24.self_attn.q_proj.weight": "model-00019-of-00106.safetensors", + "model.layers.24.self_attn.v_proj.weight": "model-00019-of-00106.safetensors", + "model.layers.25.input_layernorm.weight": "model-00020-of-00106.safetensors", + "model.layers.25.mlp.down_proj.weight": "model-00020-of-00106.safetensors", + "model.layers.25.mlp.gate_proj.weight": "model-00020-of-00106.safetensors", + "model.layers.25.mlp.up_proj.weight": "model-00020-of-00106.safetensors", + "model.layers.25.post_attention_layernorm.weight": "model-00020-of-00106.safetensors", + "model.layers.25.self_attn.k_proj.weight": "model-00019-of-00106.safetensors", + "model.layers.25.self_attn.o_proj.weight": "model-00019-of-00106.safetensors", + "model.layers.25.self_attn.q_proj.weight": "model-00019-of-00106.safetensors", + "model.layers.25.self_attn.v_proj.weight": "model-00019-of-00106.safetensors", + "model.layers.26.input_layernorm.weight": "model-00021-of-00106.safetensors", + "model.layers.26.mlp.down_proj.weight": "model-00021-of-00106.safetensors", + "model.layers.26.mlp.gate_proj.weight": "model-00020-of-00106.safetensors", + "model.layers.26.mlp.up_proj.weight": "model-00021-of-00106.safetensors", + "model.layers.26.post_attention_layernorm.weight": "model-00021-of-00106.safetensors", + "model.layers.26.self_attn.k_proj.weight": "model-00020-of-00106.safetensors", + "model.layers.26.self_attn.o_proj.weight": "model-00020-of-00106.safetensors", + "model.layers.26.self_attn.q_proj.weight": "model-00020-of-00106.safetensors", + "model.layers.26.self_attn.v_proj.weight": "model-00020-of-00106.safetensors", + "model.layers.27.input_layernorm.weight": "model-00022-of-00106.safetensors", + "model.layers.27.mlp.down_proj.weight": "model-00022-of-00106.safetensors", + "model.layers.27.mlp.gate_proj.weight": "model-00021-of-00106.safetensors", + "model.layers.27.mlp.up_proj.weight": "model-00021-of-00106.safetensors", + "model.layers.27.post_attention_layernorm.weight": "model-00022-of-00106.safetensors", + "model.layers.27.self_attn.k_proj.weight": "model-00021-of-00106.safetensors", + "model.layers.27.self_attn.o_proj.weight": "model-00021-of-00106.safetensors", + "model.layers.27.self_attn.q_proj.weight": "model-00021-of-00106.safetensors", + "model.layers.27.self_attn.v_proj.weight": "model-00021-of-00106.safetensors", + "model.layers.28.input_layernorm.weight": "model-00022-of-00106.safetensors", + "model.layers.28.mlp.down_proj.weight": "model-00022-of-00106.safetensors", + "model.layers.28.mlp.gate_proj.weight": "model-00022-of-00106.safetensors", + "model.layers.28.mlp.up_proj.weight": "model-00022-of-00106.safetensors", + "model.layers.28.post_attention_layernorm.weight": "model-00022-of-00106.safetensors", + "model.layers.28.self_attn.k_proj.weight": "model-00022-of-00106.safetensors", + "model.layers.28.self_attn.o_proj.weight": "model-00022-of-00106.safetensors", + "model.layers.28.self_attn.q_proj.weight": "model-00022-of-00106.safetensors", + "model.layers.28.self_attn.v_proj.weight": "model-00022-of-00106.safetensors", + "model.layers.29.input_layernorm.weight": "model-00023-of-00106.safetensors", + "model.layers.29.mlp.down_proj.weight": "model-00023-of-00106.safetensors", + "model.layers.29.mlp.gate_proj.weight": "model-00023-of-00106.safetensors", + "model.layers.29.mlp.up_proj.weight": "model-00023-of-00106.safetensors", + "model.layers.29.post_attention_layernorm.weight": "model-00023-of-00106.safetensors", + "model.layers.29.self_attn.k_proj.weight": "model-00022-of-00106.safetensors", + "model.layers.29.self_attn.o_proj.weight": "model-00022-of-00106.safetensors", + "model.layers.29.self_attn.q_proj.weight": "model-00022-of-00106.safetensors", + "model.layers.29.self_attn.v_proj.weight": "model-00022-of-00106.safetensors", + "model.layers.3.input_layernorm.weight": "model-00004-of-00106.safetensors", + "model.layers.3.mlp.down_proj.weight": "model-00004-of-00106.safetensors", + "model.layers.3.mlp.gate_proj.weight": "model-00003-of-00106.safetensors", + "model.layers.3.mlp.up_proj.weight": "model-00003-of-00106.safetensors", + "model.layers.3.post_attention_layernorm.weight": "model-00004-of-00106.safetensors", + "model.layers.3.self_attn.k_proj.weight": "model-00003-of-00106.safetensors", + "model.layers.3.self_attn.o_proj.weight": "model-00003-of-00106.safetensors", + "model.layers.3.self_attn.q_proj.weight": "model-00003-of-00106.safetensors", + "model.layers.3.self_attn.v_proj.weight": "model-00003-of-00106.safetensors", + "model.layers.30.input_layernorm.weight": "model-00024-of-00106.safetensors", + "model.layers.30.mlp.down_proj.weight": "model-00024-of-00106.safetensors", + "model.layers.30.mlp.gate_proj.weight": "model-00023-of-00106.safetensors", + "model.layers.30.mlp.up_proj.weight": "model-00024-of-00106.safetensors", + "model.layers.30.post_attention_layernorm.weight": "model-00024-of-00106.safetensors", + "model.layers.30.self_attn.k_proj.weight": "model-00023-of-00106.safetensors", + "model.layers.30.self_attn.o_proj.weight": "model-00023-of-00106.safetensors", + "model.layers.30.self_attn.q_proj.weight": "model-00023-of-00106.safetensors", + "model.layers.30.self_attn.v_proj.weight": "model-00023-of-00106.safetensors", + "model.layers.31.input_layernorm.weight": "model-00025-of-00106.safetensors", + "model.layers.31.mlp.down_proj.weight": "model-00025-of-00106.safetensors", + "model.layers.31.mlp.gate_proj.weight": "model-00024-of-00106.safetensors", + "model.layers.31.mlp.up_proj.weight": "model-00024-of-00106.safetensors", + "model.layers.31.post_attention_layernorm.weight": "model-00025-of-00106.safetensors", + "model.layers.31.self_attn.k_proj.weight": "model-00024-of-00106.safetensors", + "model.layers.31.self_attn.o_proj.weight": "model-00024-of-00106.safetensors", + "model.layers.31.self_attn.q_proj.weight": "model-00024-of-00106.safetensors", + "model.layers.31.self_attn.v_proj.weight": "model-00024-of-00106.safetensors", + "model.layers.32.input_layernorm.weight": "model-00025-of-00106.safetensors", + "model.layers.32.mlp.down_proj.weight": "model-00025-of-00106.safetensors", + "model.layers.32.mlp.gate_proj.weight": "model-00025-of-00106.safetensors", + "model.layers.32.mlp.up_proj.weight": "model-00025-of-00106.safetensors", + "model.layers.32.post_attention_layernorm.weight": "model-00025-of-00106.safetensors", + "model.layers.32.self_attn.k_proj.weight": "model-00025-of-00106.safetensors", + "model.layers.32.self_attn.o_proj.weight": "model-00025-of-00106.safetensors", + "model.layers.32.self_attn.q_proj.weight": "model-00025-of-00106.safetensors", + "model.layers.32.self_attn.v_proj.weight": "model-00025-of-00106.safetensors", + "model.layers.33.input_layernorm.weight": "model-00026-of-00106.safetensors", + "model.layers.33.mlp.down_proj.weight": "model-00026-of-00106.safetensors", + "model.layers.33.mlp.gate_proj.weight": "model-00026-of-00106.safetensors", + "model.layers.33.mlp.up_proj.weight": "model-00026-of-00106.safetensors", + "model.layers.33.post_attention_layernorm.weight": "model-00026-of-00106.safetensors", + "model.layers.33.self_attn.k_proj.weight": "model-00025-of-00106.safetensors", + "model.layers.33.self_attn.o_proj.weight": "model-00025-of-00106.safetensors", + "model.layers.33.self_attn.q_proj.weight": "model-00025-of-00106.safetensors", + "model.layers.33.self_attn.v_proj.weight": "model-00025-of-00106.safetensors", + "model.layers.34.input_layernorm.weight": "model-00027-of-00106.safetensors", + "model.layers.34.mlp.down_proj.weight": "model-00027-of-00106.safetensors", + "model.layers.34.mlp.gate_proj.weight": "model-00026-of-00106.safetensors", + "model.layers.34.mlp.up_proj.weight": "model-00027-of-00106.safetensors", + "model.layers.34.post_attention_layernorm.weight": "model-00027-of-00106.safetensors", + "model.layers.34.self_attn.k_proj.weight": "model-00026-of-00106.safetensors", + "model.layers.34.self_attn.o_proj.weight": "model-00026-of-00106.safetensors", + "model.layers.34.self_attn.q_proj.weight": "model-00026-of-00106.safetensors", + "model.layers.34.self_attn.v_proj.weight": "model-00026-of-00106.safetensors", + "model.layers.35.input_layernorm.weight": "model-00028-of-00106.safetensors", + "model.layers.35.mlp.down_proj.weight": "model-00028-of-00106.safetensors", + "model.layers.35.mlp.gate_proj.weight": "model-00027-of-00106.safetensors", + "model.layers.35.mlp.up_proj.weight": "model-00027-of-00106.safetensors", + "model.layers.35.post_attention_layernorm.weight": "model-00028-of-00106.safetensors", + "model.layers.35.self_attn.k_proj.weight": "model-00027-of-00106.safetensors", + "model.layers.35.self_attn.o_proj.weight": "model-00027-of-00106.safetensors", + "model.layers.35.self_attn.q_proj.weight": "model-00027-of-00106.safetensors", + "model.layers.35.self_attn.v_proj.weight": "model-00027-of-00106.safetensors", + "model.layers.36.input_layernorm.weight": "model-00028-of-00106.safetensors", + "model.layers.36.mlp.down_proj.weight": "model-00028-of-00106.safetensors", + "model.layers.36.mlp.gate_proj.weight": "model-00028-of-00106.safetensors", + "model.layers.36.mlp.up_proj.weight": "model-00028-of-00106.safetensors", + "model.layers.36.post_attention_layernorm.weight": "model-00028-of-00106.safetensors", + "model.layers.36.self_attn.k_proj.weight": "model-00028-of-00106.safetensors", + "model.layers.36.self_attn.o_proj.weight": "model-00028-of-00106.safetensors", + "model.layers.36.self_attn.q_proj.weight": "model-00028-of-00106.safetensors", + "model.layers.36.self_attn.v_proj.weight": "model-00028-of-00106.safetensors", + "model.layers.37.input_layernorm.weight": "model-00029-of-00106.safetensors", + "model.layers.37.mlp.down_proj.weight": "model-00029-of-00106.safetensors", + "model.layers.37.mlp.gate_proj.weight": "model-00029-of-00106.safetensors", + "model.layers.37.mlp.up_proj.weight": "model-00029-of-00106.safetensors", + "model.layers.37.post_attention_layernorm.weight": "model-00029-of-00106.safetensors", + "model.layers.37.self_attn.k_proj.weight": "model-00028-of-00106.safetensors", + "model.layers.37.self_attn.o_proj.weight": "model-00028-of-00106.safetensors", + "model.layers.37.self_attn.q_proj.weight": "model-00028-of-00106.safetensors", + "model.layers.37.self_attn.v_proj.weight": "model-00028-of-00106.safetensors", + "model.layers.38.input_layernorm.weight": "model-00030-of-00106.safetensors", + "model.layers.38.mlp.down_proj.weight": "model-00030-of-00106.safetensors", + "model.layers.38.mlp.gate_proj.weight": "model-00029-of-00106.safetensors", + "model.layers.38.mlp.up_proj.weight": "model-00030-of-00106.safetensors", + "model.layers.38.post_attention_layernorm.weight": "model-00030-of-00106.safetensors", + "model.layers.38.self_attn.k_proj.weight": "model-00029-of-00106.safetensors", + "model.layers.38.self_attn.o_proj.weight": "model-00029-of-00106.safetensors", + "model.layers.38.self_attn.q_proj.weight": "model-00029-of-00106.safetensors", + "model.layers.38.self_attn.v_proj.weight": "model-00029-of-00106.safetensors", + "model.layers.39.input_layernorm.weight": "model-00031-of-00106.safetensors", + "model.layers.39.mlp.down_proj.weight": "model-00031-of-00106.safetensors", + "model.layers.39.mlp.gate_proj.weight": "model-00030-of-00106.safetensors", + "model.layers.39.mlp.up_proj.weight": "model-00030-of-00106.safetensors", + "model.layers.39.post_attention_layernorm.weight": "model-00031-of-00106.safetensors", + "model.layers.39.self_attn.k_proj.weight": "model-00030-of-00106.safetensors", + "model.layers.39.self_attn.o_proj.weight": "model-00030-of-00106.safetensors", + "model.layers.39.self_attn.q_proj.weight": "model-00030-of-00106.safetensors", + "model.layers.39.self_attn.v_proj.weight": "model-00030-of-00106.safetensors", + "model.layers.4.input_layernorm.weight": "model-00004-of-00106.safetensors", + "model.layers.4.mlp.down_proj.weight": "model-00004-of-00106.safetensors", + "model.layers.4.mlp.gate_proj.weight": "model-00004-of-00106.safetensors", + "model.layers.4.mlp.up_proj.weight": "model-00004-of-00106.safetensors", + "model.layers.4.post_attention_layernorm.weight": "model-00004-of-00106.safetensors", + "model.layers.4.self_attn.k_proj.weight": "model-00004-of-00106.safetensors", + "model.layers.4.self_attn.o_proj.weight": "model-00004-of-00106.safetensors", + "model.layers.4.self_attn.q_proj.weight": "model-00004-of-00106.safetensors", + "model.layers.4.self_attn.v_proj.weight": "model-00004-of-00106.safetensors", + "model.layers.40.input_layernorm.weight": "model-00031-of-00106.safetensors", + "model.layers.40.mlp.down_proj.weight": "model-00031-of-00106.safetensors", + "model.layers.40.mlp.gate_proj.weight": "model-00031-of-00106.safetensors", + "model.layers.40.mlp.up_proj.weight": "model-00031-of-00106.safetensors", + "model.layers.40.post_attention_layernorm.weight": "model-00031-of-00106.safetensors", + "model.layers.40.self_attn.k_proj.weight": "model-00031-of-00106.safetensors", + "model.layers.40.self_attn.o_proj.weight": "model-00031-of-00106.safetensors", + "model.layers.40.self_attn.q_proj.weight": "model-00031-of-00106.safetensors", + "model.layers.40.self_attn.v_proj.weight": "model-00031-of-00106.safetensors", + "model.layers.41.input_layernorm.weight": "model-00032-of-00106.safetensors", + "model.layers.41.mlp.down_proj.weight": "model-00032-of-00106.safetensors", + "model.layers.41.mlp.gate_proj.weight": "model-00032-of-00106.safetensors", + "model.layers.41.mlp.up_proj.weight": "model-00032-of-00106.safetensors", + "model.layers.41.post_attention_layernorm.weight": "model-00032-of-00106.safetensors", + "model.layers.41.self_attn.k_proj.weight": "model-00031-of-00106.safetensors", + "model.layers.41.self_attn.o_proj.weight": "model-00031-of-00106.safetensors", + "model.layers.41.self_attn.q_proj.weight": "model-00031-of-00106.safetensors", + "model.layers.41.self_attn.v_proj.weight": "model-00031-of-00106.safetensors", + "model.layers.42.input_layernorm.weight": "model-00033-of-00106.safetensors", + "model.layers.42.mlp.down_proj.weight": "model-00033-of-00106.safetensors", + "model.layers.42.mlp.gate_proj.weight": "model-00032-of-00106.safetensors", + "model.layers.42.mlp.up_proj.weight": "model-00033-of-00106.safetensors", + "model.layers.42.post_attention_layernorm.weight": "model-00033-of-00106.safetensors", + "model.layers.42.self_attn.k_proj.weight": "model-00032-of-00106.safetensors", + "model.layers.42.self_attn.o_proj.weight": "model-00032-of-00106.safetensors", + "model.layers.42.self_attn.q_proj.weight": "model-00032-of-00106.safetensors", + "model.layers.42.self_attn.v_proj.weight": "model-00032-of-00106.safetensors", + "model.layers.43.input_layernorm.weight": "model-00034-of-00106.safetensors", + "model.layers.43.mlp.down_proj.weight": "model-00034-of-00106.safetensors", + "model.layers.43.mlp.gate_proj.weight": "model-00033-of-00106.safetensors", + "model.layers.43.mlp.up_proj.weight": "model-00033-of-00106.safetensors", + "model.layers.43.post_attention_layernorm.weight": "model-00034-of-00106.safetensors", + "model.layers.43.self_attn.k_proj.weight": "model-00033-of-00106.safetensors", + "model.layers.43.self_attn.o_proj.weight": "model-00033-of-00106.safetensors", + "model.layers.43.self_attn.q_proj.weight": "model-00033-of-00106.safetensors", + "model.layers.43.self_attn.v_proj.weight": "model-00033-of-00106.safetensors", + "model.layers.44.input_layernorm.weight": "model-00034-of-00106.safetensors", + "model.layers.44.mlp.down_proj.weight": "model-00034-of-00106.safetensors", + "model.layers.44.mlp.gate_proj.weight": "model-00034-of-00106.safetensors", + "model.layers.44.mlp.up_proj.weight": "model-00034-of-00106.safetensors", + "model.layers.44.post_attention_layernorm.weight": "model-00034-of-00106.safetensors", + "model.layers.44.self_attn.k_proj.weight": "model-00034-of-00106.safetensors", + "model.layers.44.self_attn.o_proj.weight": "model-00034-of-00106.safetensors", + "model.layers.44.self_attn.q_proj.weight": "model-00034-of-00106.safetensors", + "model.layers.44.self_attn.v_proj.weight": "model-00034-of-00106.safetensors", + "model.layers.45.input_layernorm.weight": "model-00035-of-00106.safetensors", + "model.layers.45.mlp.down_proj.weight": "model-00035-of-00106.safetensors", + "model.layers.45.mlp.gate_proj.weight": "model-00035-of-00106.safetensors", + "model.layers.45.mlp.up_proj.weight": "model-00035-of-00106.safetensors", + "model.layers.45.post_attention_layernorm.weight": "model-00035-of-00106.safetensors", + "model.layers.45.self_attn.k_proj.weight": "model-00034-of-00106.safetensors", + "model.layers.45.self_attn.o_proj.weight": "model-00034-of-00106.safetensors", + "model.layers.45.self_attn.q_proj.weight": "model-00034-of-00106.safetensors", + "model.layers.45.self_attn.v_proj.weight": "model-00034-of-00106.safetensors", + "model.layers.46.input_layernorm.weight": "model-00036-of-00106.safetensors", + "model.layers.46.mlp.down_proj.weight": "model-00036-of-00106.safetensors", + "model.layers.46.mlp.gate_proj.weight": "model-00035-of-00106.safetensors", + "model.layers.46.mlp.up_proj.weight": "model-00036-of-00106.safetensors", + "model.layers.46.post_attention_layernorm.weight": "model-00036-of-00106.safetensors", + "model.layers.46.self_attn.k_proj.weight": "model-00035-of-00106.safetensors", + "model.layers.46.self_attn.o_proj.weight": "model-00035-of-00106.safetensors", + "model.layers.46.self_attn.q_proj.weight": "model-00035-of-00106.safetensors", + "model.layers.46.self_attn.v_proj.weight": "model-00035-of-00106.safetensors", + "model.layers.47.input_layernorm.weight": "model-00037-of-00106.safetensors", + "model.layers.47.mlp.down_proj.weight": "model-00037-of-00106.safetensors", + "model.layers.47.mlp.gate_proj.weight": "model-00036-of-00106.safetensors", + "model.layers.47.mlp.up_proj.weight": "model-00036-of-00106.safetensors", + "model.layers.47.post_attention_layernorm.weight": "model-00037-of-00106.safetensors", + "model.layers.47.self_attn.k_proj.weight": "model-00036-of-00106.safetensors", + "model.layers.47.self_attn.o_proj.weight": "model-00036-of-00106.safetensors", + "model.layers.47.self_attn.q_proj.weight": "model-00036-of-00106.safetensors", + "model.layers.47.self_attn.v_proj.weight": "model-00036-of-00106.safetensors", + "model.layers.48.input_layernorm.weight": "model-00037-of-00106.safetensors", + "model.layers.48.mlp.down_proj.weight": "model-00037-of-00106.safetensors", + "model.layers.48.mlp.gate_proj.weight": "model-00037-of-00106.safetensors", + "model.layers.48.mlp.up_proj.weight": "model-00037-of-00106.safetensors", + "model.layers.48.post_attention_layernorm.weight": "model-00037-of-00106.safetensors", + "model.layers.48.self_attn.k_proj.weight": "model-00037-of-00106.safetensors", + "model.layers.48.self_attn.o_proj.weight": "model-00037-of-00106.safetensors", + "model.layers.48.self_attn.q_proj.weight": "model-00037-of-00106.safetensors", + "model.layers.48.self_attn.v_proj.weight": "model-00037-of-00106.safetensors", + "model.layers.49.input_layernorm.weight": "model-00038-of-00106.safetensors", + "model.layers.49.mlp.down_proj.weight": "model-00038-of-00106.safetensors", + "model.layers.49.mlp.gate_proj.weight": "model-00038-of-00106.safetensors", + "model.layers.49.mlp.up_proj.weight": "model-00038-of-00106.safetensors", + "model.layers.49.post_attention_layernorm.weight": "model-00038-of-00106.safetensors", + "model.layers.49.self_attn.k_proj.weight": "model-00037-of-00106.safetensors", + "model.layers.49.self_attn.o_proj.weight": "model-00037-of-00106.safetensors", + "model.layers.49.self_attn.q_proj.weight": "model-00037-of-00106.safetensors", + "model.layers.49.self_attn.v_proj.weight": "model-00037-of-00106.safetensors", + "model.layers.5.input_layernorm.weight": "model-00005-of-00106.safetensors", + "model.layers.5.mlp.down_proj.weight": "model-00005-of-00106.safetensors", + "model.layers.5.mlp.gate_proj.weight": "model-00005-of-00106.safetensors", + "model.layers.5.mlp.up_proj.weight": "model-00005-of-00106.safetensors", + "model.layers.5.post_attention_layernorm.weight": "model-00005-of-00106.safetensors", + "model.layers.5.self_attn.k_proj.weight": "model-00004-of-00106.safetensors", + "model.layers.5.self_attn.o_proj.weight": "model-00004-of-00106.safetensors", + "model.layers.5.self_attn.q_proj.weight": "model-00004-of-00106.safetensors", + "model.layers.5.self_attn.v_proj.weight": "model-00004-of-00106.safetensors", + "model.layers.50.input_layernorm.weight": "model-00039-of-00106.safetensors", + "model.layers.50.mlp.down_proj.weight": "model-00039-of-00106.safetensors", + "model.layers.50.mlp.gate_proj.weight": "model-00038-of-00106.safetensors", + "model.layers.50.mlp.up_proj.weight": "model-00039-of-00106.safetensors", + "model.layers.50.post_attention_layernorm.weight": "model-00039-of-00106.safetensors", + "model.layers.50.self_attn.k_proj.weight": "model-00038-of-00106.safetensors", + "model.layers.50.self_attn.o_proj.weight": "model-00038-of-00106.safetensors", + "model.layers.50.self_attn.q_proj.weight": "model-00038-of-00106.safetensors", + "model.layers.50.self_attn.v_proj.weight": "model-00038-of-00106.safetensors", + "model.layers.51.input_layernorm.weight": "model-00040-of-00106.safetensors", + "model.layers.51.mlp.down_proj.weight": "model-00040-of-00106.safetensors", + "model.layers.51.mlp.gate_proj.weight": "model-00039-of-00106.safetensors", + "model.layers.51.mlp.up_proj.weight": "model-00039-of-00106.safetensors", + "model.layers.51.post_attention_layernorm.weight": "model-00040-of-00106.safetensors", + "model.layers.51.self_attn.k_proj.weight": "model-00039-of-00106.safetensors", + "model.layers.51.self_attn.o_proj.weight": "model-00039-of-00106.safetensors", + "model.layers.51.self_attn.q_proj.weight": "model-00039-of-00106.safetensors", + "model.layers.51.self_attn.v_proj.weight": "model-00039-of-00106.safetensors", + "model.layers.52.input_layernorm.weight": "model-00040-of-00106.safetensors", + "model.layers.52.mlp.down_proj.weight": "model-00040-of-00106.safetensors", + "model.layers.52.mlp.gate_proj.weight": "model-00040-of-00106.safetensors", + "model.layers.52.mlp.up_proj.weight": "model-00040-of-00106.safetensors", + "model.layers.52.post_attention_layernorm.weight": "model-00040-of-00106.safetensors", + "model.layers.52.self_attn.k_proj.weight": "model-00040-of-00106.safetensors", + "model.layers.52.self_attn.o_proj.weight": "model-00040-of-00106.safetensors", + "model.layers.52.self_attn.q_proj.weight": "model-00040-of-00106.safetensors", + "model.layers.52.self_attn.v_proj.weight": "model-00040-of-00106.safetensors", + "model.layers.53.input_layernorm.weight": "model-00041-of-00106.safetensors", + "model.layers.53.mlp.down_proj.weight": "model-00041-of-00106.safetensors", + "model.layers.53.mlp.gate_proj.weight": "model-00041-of-00106.safetensors", + "model.layers.53.mlp.up_proj.weight": "model-00041-of-00106.safetensors", + "model.layers.53.post_attention_layernorm.weight": "model-00041-of-00106.safetensors", + "model.layers.53.self_attn.k_proj.weight": "model-00040-of-00106.safetensors", + "model.layers.53.self_attn.o_proj.weight": "model-00040-of-00106.safetensors", + "model.layers.53.self_attn.q_proj.weight": "model-00040-of-00106.safetensors", + "model.layers.53.self_attn.v_proj.weight": "model-00040-of-00106.safetensors", + "model.layers.54.input_layernorm.weight": "model-00042-of-00106.safetensors", + "model.layers.54.mlp.down_proj.weight": "model-00042-of-00106.safetensors", + "model.layers.54.mlp.gate_proj.weight": "model-00041-of-00106.safetensors", + "model.layers.54.mlp.up_proj.weight": "model-00042-of-00106.safetensors", + "model.layers.54.post_attention_layernorm.weight": "model-00042-of-00106.safetensors", + "model.layers.54.self_attn.k_proj.weight": "model-00041-of-00106.safetensors", + "model.layers.54.self_attn.o_proj.weight": "model-00041-of-00106.safetensors", + "model.layers.54.self_attn.q_proj.weight": "model-00041-of-00106.safetensors", + "model.layers.54.self_attn.v_proj.weight": "model-00041-of-00106.safetensors", + "model.layers.55.input_layernorm.weight": "model-00043-of-00106.safetensors", + "model.layers.55.mlp.down_proj.weight": "model-00043-of-00106.safetensors", + "model.layers.55.mlp.gate_proj.weight": "model-00042-of-00106.safetensors", + "model.layers.55.mlp.up_proj.weight": "model-00042-of-00106.safetensors", + "model.layers.55.post_attention_layernorm.weight": "model-00043-of-00106.safetensors", + "model.layers.55.self_attn.k_proj.weight": "model-00042-of-00106.safetensors", + "model.layers.55.self_attn.o_proj.weight": "model-00042-of-00106.safetensors", + "model.layers.55.self_attn.q_proj.weight": "model-00042-of-00106.safetensors", + "model.layers.55.self_attn.v_proj.weight": "model-00042-of-00106.safetensors", + "model.layers.56.input_layernorm.weight": "model-00043-of-00106.safetensors", + "model.layers.56.mlp.down_proj.weight": "model-00043-of-00106.safetensors", + "model.layers.56.mlp.gate_proj.weight": "model-00043-of-00106.safetensors", + "model.layers.56.mlp.up_proj.weight": "model-00043-of-00106.safetensors", + "model.layers.56.post_attention_layernorm.weight": "model-00043-of-00106.safetensors", + "model.layers.56.self_attn.k_proj.weight": "model-00043-of-00106.safetensors", + "model.layers.56.self_attn.o_proj.weight": "model-00043-of-00106.safetensors", + "model.layers.56.self_attn.q_proj.weight": "model-00043-of-00106.safetensors", + "model.layers.56.self_attn.v_proj.weight": "model-00043-of-00106.safetensors", + "model.layers.57.input_layernorm.weight": "model-00044-of-00106.safetensors", + "model.layers.57.mlp.down_proj.weight": "model-00044-of-00106.safetensors", + "model.layers.57.mlp.gate_proj.weight": "model-00044-of-00106.safetensors", + "model.layers.57.mlp.up_proj.weight": "model-00044-of-00106.safetensors", + "model.layers.57.post_attention_layernorm.weight": "model-00044-of-00106.safetensors", + "model.layers.57.self_attn.k_proj.weight": "model-00043-of-00106.safetensors", + "model.layers.57.self_attn.o_proj.weight": "model-00043-of-00106.safetensors", + "model.layers.57.self_attn.q_proj.weight": "model-00043-of-00106.safetensors", + "model.layers.57.self_attn.v_proj.weight": "model-00043-of-00106.safetensors", + "model.layers.58.input_layernorm.weight": "model-00045-of-00106.safetensors", + "model.layers.58.mlp.down_proj.weight": "model-00045-of-00106.safetensors", + "model.layers.58.mlp.gate_proj.weight": "model-00044-of-00106.safetensors", + "model.layers.58.mlp.up_proj.weight": "model-00045-of-00106.safetensors", + "model.layers.58.post_attention_layernorm.weight": "model-00045-of-00106.safetensors", + "model.layers.58.self_attn.k_proj.weight": "model-00044-of-00106.safetensors", + "model.layers.58.self_attn.o_proj.weight": "model-00044-of-00106.safetensors", + "model.layers.58.self_attn.q_proj.weight": "model-00044-of-00106.safetensors", + "model.layers.58.self_attn.v_proj.weight": "model-00044-of-00106.safetensors", + "model.layers.59.input_layernorm.weight": "model-00046-of-00106.safetensors", + "model.layers.59.mlp.down_proj.weight": "model-00046-of-00106.safetensors", + "model.layers.59.mlp.gate_proj.weight": "model-00045-of-00106.safetensors", + "model.layers.59.mlp.up_proj.weight": "model-00045-of-00106.safetensors", + "model.layers.59.post_attention_layernorm.weight": "model-00046-of-00106.safetensors", + "model.layers.59.self_attn.k_proj.weight": "model-00045-of-00106.safetensors", + "model.layers.59.self_attn.o_proj.weight": "model-00045-of-00106.safetensors", + "model.layers.59.self_attn.q_proj.weight": "model-00045-of-00106.safetensors", + "model.layers.59.self_attn.v_proj.weight": "model-00045-of-00106.safetensors", + "model.layers.6.input_layernorm.weight": "model-00006-of-00106.safetensors", + "model.layers.6.mlp.down_proj.weight": "model-00006-of-00106.safetensors", + "model.layers.6.mlp.gate_proj.weight": "model-00005-of-00106.safetensors", + "model.layers.6.mlp.up_proj.weight": "model-00006-of-00106.safetensors", + "model.layers.6.post_attention_layernorm.weight": "model-00006-of-00106.safetensors", + "model.layers.6.self_attn.k_proj.weight": "model-00005-of-00106.safetensors", + "model.layers.6.self_attn.o_proj.weight": "model-00005-of-00106.safetensors", + "model.layers.6.self_attn.q_proj.weight": "model-00005-of-00106.safetensors", + "model.layers.6.self_attn.v_proj.weight": "model-00005-of-00106.safetensors", + "model.layers.60.input_layernorm.weight": "model-00046-of-00106.safetensors", + "model.layers.60.mlp.down_proj.weight": "model-00046-of-00106.safetensors", + "model.layers.60.mlp.gate_proj.weight": "model-00046-of-00106.safetensors", + "model.layers.60.mlp.up_proj.weight": "model-00046-of-00106.safetensors", + "model.layers.60.post_attention_layernorm.weight": "model-00046-of-00106.safetensors", + "model.layers.60.self_attn.k_proj.weight": "model-00046-of-00106.safetensors", + "model.layers.60.self_attn.o_proj.weight": "model-00046-of-00106.safetensors", + "model.layers.60.self_attn.q_proj.weight": "model-00046-of-00106.safetensors", + "model.layers.60.self_attn.v_proj.weight": "model-00046-of-00106.safetensors", + "model.layers.61.input_layernorm.weight": "model-00047-of-00106.safetensors", + "model.layers.61.mlp.down_proj.weight": "model-00047-of-00106.safetensors", + "model.layers.61.mlp.gate_proj.weight": "model-00047-of-00106.safetensors", + "model.layers.61.mlp.up_proj.weight": "model-00047-of-00106.safetensors", + "model.layers.61.post_attention_layernorm.weight": "model-00047-of-00106.safetensors", + "model.layers.61.self_attn.k_proj.weight": "model-00046-of-00106.safetensors", + "model.layers.61.self_attn.o_proj.weight": "model-00046-of-00106.safetensors", + "model.layers.61.self_attn.q_proj.weight": "model-00046-of-00106.safetensors", + "model.layers.61.self_attn.v_proj.weight": "model-00046-of-00106.safetensors", + "model.layers.62.input_layernorm.weight": "model-00048-of-00106.safetensors", + "model.layers.62.mlp.down_proj.weight": "model-00048-of-00106.safetensors", + "model.layers.62.mlp.gate_proj.weight": "model-00047-of-00106.safetensors", + "model.layers.62.mlp.up_proj.weight": "model-00048-of-00106.safetensors", + "model.layers.62.post_attention_layernorm.weight": "model-00048-of-00106.safetensors", + "model.layers.62.self_attn.k_proj.weight": "model-00047-of-00106.safetensors", + "model.layers.62.self_attn.o_proj.weight": "model-00047-of-00106.safetensors", + "model.layers.62.self_attn.q_proj.weight": "model-00047-of-00106.safetensors", + "model.layers.62.self_attn.v_proj.weight": "model-00047-of-00106.safetensors", + "model.layers.63.input_layernorm.weight": "model-00049-of-00106.safetensors", + "model.layers.63.mlp.down_proj.weight": "model-00049-of-00106.safetensors", + "model.layers.63.mlp.gate_proj.weight": "model-00048-of-00106.safetensors", + "model.layers.63.mlp.up_proj.weight": "model-00048-of-00106.safetensors", + "model.layers.63.post_attention_layernorm.weight": "model-00049-of-00106.safetensors", + "model.layers.63.self_attn.k_proj.weight": "model-00048-of-00106.safetensors", + "model.layers.63.self_attn.o_proj.weight": "model-00048-of-00106.safetensors", + "model.layers.63.self_attn.q_proj.weight": "model-00048-of-00106.safetensors", + "model.layers.63.self_attn.v_proj.weight": "model-00048-of-00106.safetensors", + "model.layers.64.input_layernorm.weight": "model-00049-of-00106.safetensors", + "model.layers.64.mlp.down_proj.weight": "model-00049-of-00106.safetensors", + "model.layers.64.mlp.gate_proj.weight": "model-00049-of-00106.safetensors", + "model.layers.64.mlp.up_proj.weight": "model-00049-of-00106.safetensors", + "model.layers.64.post_attention_layernorm.weight": "model-00049-of-00106.safetensors", + "model.layers.64.self_attn.k_proj.weight": "model-00049-of-00106.safetensors", + "model.layers.64.self_attn.o_proj.weight": "model-00049-of-00106.safetensors", + "model.layers.64.self_attn.q_proj.weight": "model-00049-of-00106.safetensors", + "model.layers.64.self_attn.v_proj.weight": "model-00049-of-00106.safetensors", + "model.layers.65.input_layernorm.weight": "model-00050-of-00106.safetensors", + "model.layers.65.mlp.down_proj.weight": "model-00050-of-00106.safetensors", + "model.layers.65.mlp.gate_proj.weight": "model-00050-of-00106.safetensors", + "model.layers.65.mlp.up_proj.weight": "model-00050-of-00106.safetensors", + "model.layers.65.post_attention_layernorm.weight": "model-00050-of-00106.safetensors", + "model.layers.65.self_attn.k_proj.weight": "model-00049-of-00106.safetensors", + "model.layers.65.self_attn.o_proj.weight": "model-00049-of-00106.safetensors", + "model.layers.65.self_attn.q_proj.weight": "model-00049-of-00106.safetensors", + "model.layers.65.self_attn.v_proj.weight": "model-00049-of-00106.safetensors", + "model.layers.66.input_layernorm.weight": "model-00051-of-00106.safetensors", + "model.layers.66.mlp.down_proj.weight": "model-00051-of-00106.safetensors", + "model.layers.66.mlp.gate_proj.weight": "model-00050-of-00106.safetensors", + "model.layers.66.mlp.up_proj.weight": "model-00051-of-00106.safetensors", + "model.layers.66.post_attention_layernorm.weight": "model-00051-of-00106.safetensors", + "model.layers.66.self_attn.k_proj.weight": "model-00050-of-00106.safetensors", + "model.layers.66.self_attn.o_proj.weight": "model-00050-of-00106.safetensors", + "model.layers.66.self_attn.q_proj.weight": "model-00050-of-00106.safetensors", + "model.layers.66.self_attn.v_proj.weight": "model-00050-of-00106.safetensors", + "model.layers.67.input_layernorm.weight": "model-00052-of-00106.safetensors", + "model.layers.67.mlp.down_proj.weight": "model-00052-of-00106.safetensors", + "model.layers.67.mlp.gate_proj.weight": "model-00051-of-00106.safetensors", + "model.layers.67.mlp.up_proj.weight": "model-00051-of-00106.safetensors", + "model.layers.67.post_attention_layernorm.weight": "model-00052-of-00106.safetensors", + "model.layers.67.self_attn.k_proj.weight": "model-00051-of-00106.safetensors", + "model.layers.67.self_attn.o_proj.weight": "model-00051-of-00106.safetensors", + "model.layers.67.self_attn.q_proj.weight": "model-00051-of-00106.safetensors", + "model.layers.67.self_attn.v_proj.weight": "model-00051-of-00106.safetensors", + "model.layers.68.input_layernorm.weight": "model-00052-of-00106.safetensors", + "model.layers.68.mlp.down_proj.weight": "model-00052-of-00106.safetensors", + "model.layers.68.mlp.gate_proj.weight": "model-00052-of-00106.safetensors", + "model.layers.68.mlp.up_proj.weight": "model-00052-of-00106.safetensors", + "model.layers.68.post_attention_layernorm.weight": "model-00052-of-00106.safetensors", + "model.layers.68.self_attn.k_proj.weight": "model-00052-of-00106.safetensors", + "model.layers.68.self_attn.o_proj.weight": "model-00052-of-00106.safetensors", + "model.layers.68.self_attn.q_proj.weight": "model-00052-of-00106.safetensors", + "model.layers.68.self_attn.v_proj.weight": "model-00052-of-00106.safetensors", + "model.layers.69.input_layernorm.weight": "model-00053-of-00106.safetensors", + "model.layers.69.mlp.down_proj.weight": "model-00053-of-00106.safetensors", + "model.layers.69.mlp.gate_proj.weight": "model-00053-of-00106.safetensors", + "model.layers.69.mlp.up_proj.weight": "model-00053-of-00106.safetensors", + "model.layers.69.post_attention_layernorm.weight": "model-00053-of-00106.safetensors", + "model.layers.69.self_attn.k_proj.weight": "model-00052-of-00106.safetensors", + "model.layers.69.self_attn.o_proj.weight": "model-00052-of-00106.safetensors", + "model.layers.69.self_attn.q_proj.weight": "model-00052-of-00106.safetensors", + "model.layers.69.self_attn.v_proj.weight": "model-00052-of-00106.safetensors", + "model.layers.7.input_layernorm.weight": "model-00007-of-00106.safetensors", + "model.layers.7.mlp.down_proj.weight": "model-00007-of-00106.safetensors", + "model.layers.7.mlp.gate_proj.weight": "model-00006-of-00106.safetensors", + "model.layers.7.mlp.up_proj.weight": "model-00006-of-00106.safetensors", + "model.layers.7.post_attention_layernorm.weight": "model-00007-of-00106.safetensors", + "model.layers.7.self_attn.k_proj.weight": "model-00006-of-00106.safetensors", + "model.layers.7.self_attn.o_proj.weight": "model-00006-of-00106.safetensors", + "model.layers.7.self_attn.q_proj.weight": "model-00006-of-00106.safetensors", + "model.layers.7.self_attn.v_proj.weight": "model-00006-of-00106.safetensors", + "model.layers.70.input_layernorm.weight": "model-00054-of-00106.safetensors", + "model.layers.70.mlp.down_proj.weight": "model-00054-of-00106.safetensors", + "model.layers.70.mlp.gate_proj.weight": "model-00053-of-00106.safetensors", + "model.layers.70.mlp.up_proj.weight": "model-00054-of-00106.safetensors", + "model.layers.70.post_attention_layernorm.weight": "model-00054-of-00106.safetensors", + "model.layers.70.self_attn.k_proj.weight": "model-00053-of-00106.safetensors", + "model.layers.70.self_attn.o_proj.weight": "model-00053-of-00106.safetensors", + "model.layers.70.self_attn.q_proj.weight": "model-00053-of-00106.safetensors", + "model.layers.70.self_attn.v_proj.weight": "model-00053-of-00106.safetensors", + "model.layers.71.input_layernorm.weight": "model-00055-of-00106.safetensors", + "model.layers.71.mlp.down_proj.weight": "model-00055-of-00106.safetensors", + "model.layers.71.mlp.gate_proj.weight": "model-00054-of-00106.safetensors", + "model.layers.71.mlp.up_proj.weight": "model-00054-of-00106.safetensors", + "model.layers.71.post_attention_layernorm.weight": "model-00055-of-00106.safetensors", + "model.layers.71.self_attn.k_proj.weight": "model-00054-of-00106.safetensors", + "model.layers.71.self_attn.o_proj.weight": "model-00054-of-00106.safetensors", + "model.layers.71.self_attn.q_proj.weight": "model-00054-of-00106.safetensors", + "model.layers.71.self_attn.v_proj.weight": "model-00054-of-00106.safetensors", + "model.layers.72.input_layernorm.weight": "model-00055-of-00106.safetensors", + "model.layers.72.mlp.down_proj.weight": "model-00055-of-00106.safetensors", + "model.layers.72.mlp.gate_proj.weight": "model-00055-of-00106.safetensors", + "model.layers.72.mlp.up_proj.weight": "model-00055-of-00106.safetensors", + "model.layers.72.post_attention_layernorm.weight": "model-00055-of-00106.safetensors", + "model.layers.72.self_attn.k_proj.weight": "model-00055-of-00106.safetensors", + "model.layers.72.self_attn.o_proj.weight": "model-00055-of-00106.safetensors", + "model.layers.72.self_attn.q_proj.weight": "model-00055-of-00106.safetensors", + "model.layers.72.self_attn.v_proj.weight": "model-00055-of-00106.safetensors", + "model.layers.73.input_layernorm.weight": "model-00056-of-00106.safetensors", + "model.layers.73.mlp.down_proj.weight": "model-00056-of-00106.safetensors", + "model.layers.73.mlp.gate_proj.weight": "model-00056-of-00106.safetensors", + "model.layers.73.mlp.up_proj.weight": "model-00056-of-00106.safetensors", + "model.layers.73.post_attention_layernorm.weight": "model-00056-of-00106.safetensors", + "model.layers.73.self_attn.k_proj.weight": "model-00055-of-00106.safetensors", + "model.layers.73.self_attn.o_proj.weight": "model-00055-of-00106.safetensors", + "model.layers.73.self_attn.q_proj.weight": "model-00055-of-00106.safetensors", + "model.layers.73.self_attn.v_proj.weight": "model-00055-of-00106.safetensors", + "model.layers.74.input_layernorm.weight": "model-00057-of-00106.safetensors", + "model.layers.74.mlp.down_proj.weight": "model-00057-of-00106.safetensors", + "model.layers.74.mlp.gate_proj.weight": "model-00056-of-00106.safetensors", + "model.layers.74.mlp.up_proj.weight": "model-00057-of-00106.safetensors", + "model.layers.74.post_attention_layernorm.weight": "model-00057-of-00106.safetensors", + "model.layers.74.self_attn.k_proj.weight": "model-00056-of-00106.safetensors", + "model.layers.74.self_attn.o_proj.weight": "model-00056-of-00106.safetensors", + "model.layers.74.self_attn.q_proj.weight": "model-00056-of-00106.safetensors", + "model.layers.74.self_attn.v_proj.weight": "model-00056-of-00106.safetensors", + "model.layers.75.input_layernorm.weight": "model-00058-of-00106.safetensors", + "model.layers.75.mlp.down_proj.weight": "model-00058-of-00106.safetensors", + "model.layers.75.mlp.gate_proj.weight": "model-00057-of-00106.safetensors", + "model.layers.75.mlp.up_proj.weight": "model-00057-of-00106.safetensors", + "model.layers.75.post_attention_layernorm.weight": "model-00058-of-00106.safetensors", + "model.layers.75.self_attn.k_proj.weight": "model-00057-of-00106.safetensors", + "model.layers.75.self_attn.o_proj.weight": "model-00057-of-00106.safetensors", + "model.layers.75.self_attn.q_proj.weight": "model-00057-of-00106.safetensors", + "model.layers.75.self_attn.v_proj.weight": "model-00057-of-00106.safetensors", + "model.layers.76.input_layernorm.weight": "model-00058-of-00106.safetensors", + "model.layers.76.mlp.down_proj.weight": "model-00058-of-00106.safetensors", + "model.layers.76.mlp.gate_proj.weight": "model-00058-of-00106.safetensors", + "model.layers.76.mlp.up_proj.weight": "model-00058-of-00106.safetensors", + "model.layers.76.post_attention_layernorm.weight": "model-00058-of-00106.safetensors", + "model.layers.76.self_attn.k_proj.weight": "model-00058-of-00106.safetensors", + "model.layers.76.self_attn.o_proj.weight": "model-00058-of-00106.safetensors", + "model.layers.76.self_attn.q_proj.weight": "model-00058-of-00106.safetensors", + "model.layers.76.self_attn.v_proj.weight": "model-00058-of-00106.safetensors", + "model.layers.77.input_layernorm.weight": "model-00059-of-00106.safetensors", + "model.layers.77.mlp.down_proj.weight": "model-00059-of-00106.safetensors", + "model.layers.77.mlp.gate_proj.weight": "model-00059-of-00106.safetensors", + "model.layers.77.mlp.up_proj.weight": "model-00059-of-00106.safetensors", + "model.layers.77.post_attention_layernorm.weight": "model-00059-of-00106.safetensors", + "model.layers.77.self_attn.k_proj.weight": "model-00058-of-00106.safetensors", + "model.layers.77.self_attn.o_proj.weight": "model-00058-of-00106.safetensors", + "model.layers.77.self_attn.q_proj.weight": "model-00058-of-00106.safetensors", + "model.layers.77.self_attn.v_proj.weight": "model-00058-of-00106.safetensors", + "model.layers.78.input_layernorm.weight": "model-00060-of-00106.safetensors", + "model.layers.78.mlp.down_proj.weight": "model-00060-of-00106.safetensors", + "model.layers.78.mlp.gate_proj.weight": "model-00059-of-00106.safetensors", + "model.layers.78.mlp.up_proj.weight": "model-00060-of-00106.safetensors", + "model.layers.78.post_attention_layernorm.weight": "model-00060-of-00106.safetensors", + "model.layers.78.self_attn.k_proj.weight": "model-00059-of-00106.safetensors", + "model.layers.78.self_attn.o_proj.weight": "model-00059-of-00106.safetensors", + "model.layers.78.self_attn.q_proj.weight": "model-00059-of-00106.safetensors", + "model.layers.78.self_attn.v_proj.weight": "model-00059-of-00106.safetensors", + "model.layers.79.input_layernorm.weight": "model-00061-of-00106.safetensors", + "model.layers.79.mlp.down_proj.weight": "model-00061-of-00106.safetensors", + "model.layers.79.mlp.gate_proj.weight": "model-00060-of-00106.safetensors", + "model.layers.79.mlp.up_proj.weight": "model-00060-of-00106.safetensors", + "model.layers.79.post_attention_layernorm.weight": "model-00061-of-00106.safetensors", + "model.layers.79.self_attn.k_proj.weight": "model-00060-of-00106.safetensors", + "model.layers.79.self_attn.o_proj.weight": "model-00060-of-00106.safetensors", + "model.layers.79.self_attn.q_proj.weight": "model-00060-of-00106.safetensors", + "model.layers.79.self_attn.v_proj.weight": "model-00060-of-00106.safetensors", + "model.layers.8.input_layernorm.weight": "model-00007-of-00106.safetensors", + "model.layers.8.mlp.down_proj.weight": "model-00007-of-00106.safetensors", + "model.layers.8.mlp.gate_proj.weight": "model-00007-of-00106.safetensors", + "model.layers.8.mlp.up_proj.weight": "model-00007-of-00106.safetensors", + "model.layers.8.post_attention_layernorm.weight": "model-00007-of-00106.safetensors", + "model.layers.8.self_attn.k_proj.weight": "model-00007-of-00106.safetensors", + "model.layers.8.self_attn.o_proj.weight": "model-00007-of-00106.safetensors", + "model.layers.8.self_attn.q_proj.weight": "model-00007-of-00106.safetensors", + "model.layers.8.self_attn.v_proj.weight": "model-00007-of-00106.safetensors", + "model.layers.80.input_layernorm.weight": "model-00061-of-00106.safetensors", + "model.layers.80.mlp.down_proj.weight": "model-00061-of-00106.safetensors", + "model.layers.80.mlp.gate_proj.weight": "model-00061-of-00106.safetensors", + "model.layers.80.mlp.up_proj.weight": "model-00061-of-00106.safetensors", + "model.layers.80.post_attention_layernorm.weight": "model-00061-of-00106.safetensors", + "model.layers.80.self_attn.k_proj.weight": "model-00061-of-00106.safetensors", + "model.layers.80.self_attn.o_proj.weight": "model-00061-of-00106.safetensors", + "model.layers.80.self_attn.q_proj.weight": "model-00061-of-00106.safetensors", + "model.layers.80.self_attn.v_proj.weight": "model-00061-of-00106.safetensors", + "model.layers.81.input_layernorm.weight": "model-00062-of-00106.safetensors", + "model.layers.81.mlp.down_proj.weight": "model-00062-of-00106.safetensors", + "model.layers.81.mlp.gate_proj.weight": "model-00062-of-00106.safetensors", + "model.layers.81.mlp.up_proj.weight": "model-00062-of-00106.safetensors", + "model.layers.81.post_attention_layernorm.weight": "model-00062-of-00106.safetensors", + "model.layers.81.self_attn.k_proj.weight": "model-00061-of-00106.safetensors", + "model.layers.81.self_attn.o_proj.weight": "model-00061-of-00106.safetensors", + "model.layers.81.self_attn.q_proj.weight": "model-00061-of-00106.safetensors", + "model.layers.81.self_attn.v_proj.weight": "model-00061-of-00106.safetensors", + "model.layers.82.input_layernorm.weight": "model-00063-of-00106.safetensors", + "model.layers.82.mlp.down_proj.weight": "model-00063-of-00106.safetensors", + "model.layers.82.mlp.gate_proj.weight": "model-00062-of-00106.safetensors", + "model.layers.82.mlp.up_proj.weight": "model-00063-of-00106.safetensors", + "model.layers.82.post_attention_layernorm.weight": "model-00063-of-00106.safetensors", + "model.layers.82.self_attn.k_proj.weight": "model-00062-of-00106.safetensors", + "model.layers.82.self_attn.o_proj.weight": "model-00062-of-00106.safetensors", + "model.layers.82.self_attn.q_proj.weight": "model-00062-of-00106.safetensors", + "model.layers.82.self_attn.v_proj.weight": "model-00062-of-00106.safetensors", + "model.layers.83.input_layernorm.weight": "model-00064-of-00106.safetensors", + "model.layers.83.mlp.down_proj.weight": "model-00064-of-00106.safetensors", + "model.layers.83.mlp.gate_proj.weight": "model-00063-of-00106.safetensors", + "model.layers.83.mlp.up_proj.weight": "model-00063-of-00106.safetensors", + "model.layers.83.post_attention_layernorm.weight": "model-00064-of-00106.safetensors", + "model.layers.83.self_attn.k_proj.weight": "model-00063-of-00106.safetensors", + "model.layers.83.self_attn.o_proj.weight": "model-00063-of-00106.safetensors", + "model.layers.83.self_attn.q_proj.weight": "model-00063-of-00106.safetensors", + "model.layers.83.self_attn.v_proj.weight": "model-00063-of-00106.safetensors", + "model.layers.84.input_layernorm.weight": "model-00064-of-00106.safetensors", + "model.layers.84.mlp.down_proj.weight": "model-00064-of-00106.safetensors", + "model.layers.84.mlp.gate_proj.weight": "model-00064-of-00106.safetensors", + "model.layers.84.mlp.up_proj.weight": "model-00064-of-00106.safetensors", + "model.layers.84.post_attention_layernorm.weight": "model-00064-of-00106.safetensors", + "model.layers.84.self_attn.k_proj.weight": "model-00064-of-00106.safetensors", + "model.layers.84.self_attn.o_proj.weight": "model-00064-of-00106.safetensors", + "model.layers.84.self_attn.q_proj.weight": "model-00064-of-00106.safetensors", + "model.layers.84.self_attn.v_proj.weight": "model-00064-of-00106.safetensors", + "model.layers.85.input_layernorm.weight": "model-00065-of-00106.safetensors", + "model.layers.85.mlp.down_proj.weight": "model-00065-of-00106.safetensors", + "model.layers.85.mlp.gate_proj.weight": "model-00065-of-00106.safetensors", + "model.layers.85.mlp.up_proj.weight": "model-00065-of-00106.safetensors", + "model.layers.85.post_attention_layernorm.weight": "model-00065-of-00106.safetensors", + "model.layers.85.self_attn.k_proj.weight": "model-00064-of-00106.safetensors", + "model.layers.85.self_attn.o_proj.weight": "model-00064-of-00106.safetensors", + "model.layers.85.self_attn.q_proj.weight": "model-00064-of-00106.safetensors", + "model.layers.85.self_attn.v_proj.weight": "model-00064-of-00106.safetensors", + "model.layers.86.input_layernorm.weight": "model-00066-of-00106.safetensors", + "model.layers.86.mlp.down_proj.weight": "model-00066-of-00106.safetensors", + "model.layers.86.mlp.gate_proj.weight": "model-00065-of-00106.safetensors", + "model.layers.86.mlp.up_proj.weight": "model-00066-of-00106.safetensors", + "model.layers.86.post_attention_layernorm.weight": "model-00066-of-00106.safetensors", + "model.layers.86.self_attn.k_proj.weight": "model-00065-of-00106.safetensors", + "model.layers.86.self_attn.o_proj.weight": "model-00065-of-00106.safetensors", + "model.layers.86.self_attn.q_proj.weight": "model-00065-of-00106.safetensors", + "model.layers.86.self_attn.v_proj.weight": "model-00065-of-00106.safetensors", + "model.layers.87.input_layernorm.weight": "model-00067-of-00106.safetensors", + "model.layers.87.mlp.down_proj.weight": "model-00067-of-00106.safetensors", + "model.layers.87.mlp.gate_proj.weight": "model-00066-of-00106.safetensors", + "model.layers.87.mlp.up_proj.weight": "model-00066-of-00106.safetensors", + "model.layers.87.post_attention_layernorm.weight": "model-00067-of-00106.safetensors", + "model.layers.87.self_attn.k_proj.weight": "model-00066-of-00106.safetensors", + "model.layers.87.self_attn.o_proj.weight": "model-00066-of-00106.safetensors", + "model.layers.87.self_attn.q_proj.weight": "model-00066-of-00106.safetensors", + "model.layers.87.self_attn.v_proj.weight": "model-00066-of-00106.safetensors", + "model.layers.88.input_layernorm.weight": "model-00067-of-00106.safetensors", + "model.layers.88.mlp.down_proj.weight": "model-00067-of-00106.safetensors", + "model.layers.88.mlp.gate_proj.weight": "model-00067-of-00106.safetensors", + "model.layers.88.mlp.up_proj.weight": "model-00067-of-00106.safetensors", + "model.layers.88.post_attention_layernorm.weight": "model-00067-of-00106.safetensors", + "model.layers.88.self_attn.k_proj.weight": "model-00067-of-00106.safetensors", + "model.layers.88.self_attn.o_proj.weight": "model-00067-of-00106.safetensors", + "model.layers.88.self_attn.q_proj.weight": "model-00067-of-00106.safetensors", + "model.layers.88.self_attn.v_proj.weight": "model-00067-of-00106.safetensors", + "model.layers.89.input_layernorm.weight": "model-00068-of-00106.safetensors", + "model.layers.89.mlp.down_proj.weight": "model-00068-of-00106.safetensors", + "model.layers.89.mlp.gate_proj.weight": "model-00068-of-00106.safetensors", + "model.layers.89.mlp.up_proj.weight": "model-00068-of-00106.safetensors", + "model.layers.89.post_attention_layernorm.weight": "model-00068-of-00106.safetensors", + "model.layers.89.self_attn.k_proj.weight": "model-00067-of-00106.safetensors", + "model.layers.89.self_attn.o_proj.weight": "model-00067-of-00106.safetensors", + "model.layers.89.self_attn.q_proj.weight": "model-00067-of-00106.safetensors", + "model.layers.89.self_attn.v_proj.weight": "model-00067-of-00106.safetensors", + "model.layers.9.input_layernorm.weight": "model-00008-of-00106.safetensors", + "model.layers.9.mlp.down_proj.weight": "model-00008-of-00106.safetensors", + "model.layers.9.mlp.gate_proj.weight": "model-00008-of-00106.safetensors", + "model.layers.9.mlp.up_proj.weight": "model-00008-of-00106.safetensors", + "model.layers.9.post_attention_layernorm.weight": "model-00008-of-00106.safetensors", + "model.layers.9.self_attn.k_proj.weight": "model-00007-of-00106.safetensors", + "model.layers.9.self_attn.o_proj.weight": "model-00007-of-00106.safetensors", + "model.layers.9.self_attn.q_proj.weight": "model-00007-of-00106.safetensors", + "model.layers.9.self_attn.v_proj.weight": "model-00007-of-00106.safetensors", + "model.layers.90.input_layernorm.weight": "model-00069-of-00106.safetensors", + "model.layers.90.mlp.down_proj.weight": "model-00069-of-00106.safetensors", + "model.layers.90.mlp.gate_proj.weight": "model-00068-of-00106.safetensors", + "model.layers.90.mlp.up_proj.weight": "model-00069-of-00106.safetensors", + "model.layers.90.post_attention_layernorm.weight": "model-00069-of-00106.safetensors", + "model.layers.90.self_attn.k_proj.weight": "model-00068-of-00106.safetensors", + "model.layers.90.self_attn.o_proj.weight": "model-00068-of-00106.safetensors", + "model.layers.90.self_attn.q_proj.weight": "model-00068-of-00106.safetensors", + "model.layers.90.self_attn.v_proj.weight": "model-00068-of-00106.safetensors", + "model.layers.91.input_layernorm.weight": "model-00070-of-00106.safetensors", + "model.layers.91.mlp.down_proj.weight": "model-00070-of-00106.safetensors", + "model.layers.91.mlp.gate_proj.weight": "model-00069-of-00106.safetensors", + "model.layers.91.mlp.up_proj.weight": "model-00069-of-00106.safetensors", + "model.layers.91.post_attention_layernorm.weight": "model-00070-of-00106.safetensors", + "model.layers.91.self_attn.k_proj.weight": "model-00069-of-00106.safetensors", + "model.layers.91.self_attn.o_proj.weight": "model-00069-of-00106.safetensors", + "model.layers.91.self_attn.q_proj.weight": "model-00069-of-00106.safetensors", + "model.layers.91.self_attn.v_proj.weight": "model-00069-of-00106.safetensors", + "model.layers.92.input_layernorm.weight": "model-00070-of-00106.safetensors", + "model.layers.92.mlp.down_proj.weight": "model-00070-of-00106.safetensors", + "model.layers.92.mlp.gate_proj.weight": "model-00070-of-00106.safetensors", + "model.layers.92.mlp.up_proj.weight": "model-00070-of-00106.safetensors", + "model.layers.92.post_attention_layernorm.weight": "model-00070-of-00106.safetensors", + "model.layers.92.self_attn.k_proj.weight": "model-00070-of-00106.safetensors", + "model.layers.92.self_attn.o_proj.weight": "model-00070-of-00106.safetensors", + "model.layers.92.self_attn.q_proj.weight": "model-00070-of-00106.safetensors", + "model.layers.92.self_attn.v_proj.weight": "model-00070-of-00106.safetensors", + "model.layers.93.input_layernorm.weight": "model-00071-of-00106.safetensors", + "model.layers.93.mlp.down_proj.weight": "model-00071-of-00106.safetensors", + "model.layers.93.mlp.gate_proj.weight": "model-00071-of-00106.safetensors", + "model.layers.93.mlp.up_proj.weight": "model-00071-of-00106.safetensors", + "model.layers.93.post_attention_layernorm.weight": "model-00071-of-00106.safetensors", + "model.layers.93.self_attn.k_proj.weight": "model-00070-of-00106.safetensors", + "model.layers.93.self_attn.o_proj.weight": "model-00070-of-00106.safetensors", + "model.layers.93.self_attn.q_proj.weight": "model-00070-of-00106.safetensors", + "model.layers.93.self_attn.v_proj.weight": "model-00070-of-00106.safetensors", + "model.layers.94.input_layernorm.weight": "model-00072-of-00106.safetensors", + "model.layers.94.mlp.down_proj.weight": "model-00072-of-00106.safetensors", + "model.layers.94.mlp.gate_proj.weight": "model-00071-of-00106.safetensors", + "model.layers.94.mlp.up_proj.weight": "model-00072-of-00106.safetensors", + "model.layers.94.post_attention_layernorm.weight": "model-00072-of-00106.safetensors", + "model.layers.94.self_attn.k_proj.weight": "model-00071-of-00106.safetensors", + "model.layers.94.self_attn.o_proj.weight": "model-00071-of-00106.safetensors", + "model.layers.94.self_attn.q_proj.weight": "model-00071-of-00106.safetensors", + "model.layers.94.self_attn.v_proj.weight": "model-00071-of-00106.safetensors", + "model.layers.95.input_layernorm.weight": "model-00073-of-00106.safetensors", + "model.layers.95.mlp.down_proj.weight": "model-00073-of-00106.safetensors", + "model.layers.95.mlp.gate_proj.weight": "model-00072-of-00106.safetensors", + "model.layers.95.mlp.up_proj.weight": "model-00072-of-00106.safetensors", + "model.layers.95.post_attention_layernorm.weight": "model-00073-of-00106.safetensors", + "model.layers.95.self_attn.k_proj.weight": "model-00072-of-00106.safetensors", + "model.layers.95.self_attn.o_proj.weight": "model-00072-of-00106.safetensors", + "model.layers.95.self_attn.q_proj.weight": "model-00072-of-00106.safetensors", + "model.layers.95.self_attn.v_proj.weight": "model-00072-of-00106.safetensors", + "model.layers.96.input_layernorm.weight": "model-00073-of-00106.safetensors", + "model.layers.96.mlp.down_proj.weight": "model-00073-of-00106.safetensors", + "model.layers.96.mlp.gate_proj.weight": "model-00073-of-00106.safetensors", + "model.layers.96.mlp.up_proj.weight": "model-00073-of-00106.safetensors", + "model.layers.96.post_attention_layernorm.weight": "model-00073-of-00106.safetensors", + "model.layers.96.self_attn.k_proj.weight": "model-00073-of-00106.safetensors", + "model.layers.96.self_attn.o_proj.weight": "model-00073-of-00106.safetensors", + "model.layers.96.self_attn.q_proj.weight": "model-00073-of-00106.safetensors", + "model.layers.96.self_attn.v_proj.weight": "model-00073-of-00106.safetensors", + "model.layers.97.input_layernorm.weight": "model-00074-of-00106.safetensors", + "model.layers.97.mlp.down_proj.weight": "model-00074-of-00106.safetensors", + "model.layers.97.mlp.gate_proj.weight": "model-00074-of-00106.safetensors", + "model.layers.97.mlp.up_proj.weight": "model-00074-of-00106.safetensors", + "model.layers.97.post_attention_layernorm.weight": "model-00074-of-00106.safetensors", + "model.layers.97.self_attn.k_proj.weight": "model-00073-of-00106.safetensors", + "model.layers.97.self_attn.o_proj.weight": "model-00073-of-00106.safetensors", + "model.layers.97.self_attn.q_proj.weight": "model-00073-of-00106.safetensors", + "model.layers.97.self_attn.v_proj.weight": "model-00073-of-00106.safetensors", + "model.layers.98.input_layernorm.weight": "model-00075-of-00106.safetensors", + "model.layers.98.mlp.down_proj.weight": "model-00075-of-00106.safetensors", + "model.layers.98.mlp.gate_proj.weight": "model-00074-of-00106.safetensors", + "model.layers.98.mlp.up_proj.weight": "model-00075-of-00106.safetensors", + "model.layers.98.post_attention_layernorm.weight": "model-00075-of-00106.safetensors", + "model.layers.98.self_attn.k_proj.weight": "model-00074-of-00106.safetensors", + "model.layers.98.self_attn.o_proj.weight": "model-00074-of-00106.safetensors", + "model.layers.98.self_attn.q_proj.weight": "model-00074-of-00106.safetensors", + "model.layers.98.self_attn.v_proj.weight": "model-00074-of-00106.safetensors", + "model.layers.99.input_layernorm.weight": "model-00076-of-00106.safetensors", + "model.layers.99.mlp.down_proj.weight": "model-00076-of-00106.safetensors", + "model.layers.99.mlp.gate_proj.weight": "model-00075-of-00106.safetensors", + "model.layers.99.mlp.up_proj.weight": "model-00075-of-00106.safetensors", + "model.layers.99.post_attention_layernorm.weight": "model-00076-of-00106.safetensors", + "model.layers.99.self_attn.k_proj.weight": "model-00075-of-00106.safetensors", + "model.layers.99.self_attn.o_proj.weight": "model-00075-of-00106.safetensors", + "model.layers.99.self_attn.q_proj.weight": "model-00075-of-00106.safetensors", + "model.layers.99.self_attn.v_proj.weight": "model-00075-of-00106.safetensors", + "model.norm.weight": "model-00106-of-00106.safetensors" + } +}