diff --git a/model-00001-of-00053.safetensors b/model-00001-of-00053.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..d4fd00f07dced3339eb6d7f8b2433a7b40d68948 --- /dev/null +++ b/model-00001-of-00053.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:fc4b51b29aef6a6e38a99236a835c30f56df4864b685c639884d8c445ba34b27 +size 509627776 diff --git a/model-00002-of-00053.safetensors b/model-00002-of-00053.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..7d0ac5d0a62b5fbfadf8230f3a899674014b8a20 --- /dev/null +++ b/model-00002-of-00053.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:f26b0a6dc4a4a444a1bb3361ea71a58ee746bcf3822807fd8bd92407a559e628 +size 509627776 diff --git a/model-00003-of-00053.safetensors b/model-00003-of-00053.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..a3a8378738cf478be90ab76c2d5af71f1cef465c --- /dev/null +++ b/model-00003-of-00053.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:576637d69790814bfcefb3fe1eb191cdb56a3ed573bdd60a0099ec04bd4f5775 +size 509627776 diff --git a/model-00004-of-00053.safetensors b/model-00004-of-00053.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..dbee6f432c4dfd2b98f9eb361de8b98c7d8c2830 --- /dev/null +++ b/model-00004-of-00053.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:d1eb498137dabfcbc13a29ef6c2916357425a1ba64a4cf8610b13c4760f2d304 +size 509627776 diff --git a/model-00005-of-00053.safetensors b/model-00005-of-00053.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..539cb4b08e586411f4f85507b4042485ccced483 --- /dev/null +++ b/model-00005-of-00053.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:7ca904280af02442444b4f870541d2192d99567a62a741435864101fda8d6e53 +size 509627776 diff --git a/model-00006-of-00053.safetensors b/model-00006-of-00053.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..cf1e5d0e7a3fe42af08c79902b5bb6bde79a0989 --- /dev/null +++ b/model-00006-of-00053.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:29828a543977520f0b32cf54fd3b61d1fbf31aa7a7f5430ce498ef2711c4d745 +size 509627776 diff --git a/model-00007-of-00053.safetensors b/model-00007-of-00053.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..9d8d7829417ebac1fb68c7abadd180e42fb67383 --- /dev/null +++ b/model-00007-of-00053.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:35a27ddbe20a2180adac5287608254cffc35c6c17760a6f80e4ed50fc9c2a774 +size 509627776 diff --git a/model-00008-of-00053.safetensors b/model-00008-of-00053.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..2f3cc335774b41d802aee662c3fcae5477a1991d --- /dev/null +++ b/model-00008-of-00053.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:f631352ea66f3387d0d2ed9a66d0dc27f3fa10b9ca652e38b667966fe363ca53 +size 509627776 diff --git a/model-00009-of-00053.safetensors b/model-00009-of-00053.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..608f40f74d2fcfde904e1919f1312a59ced240e0 --- /dev/null +++ b/model-00009-of-00053.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:320428a6dae74c7eec1680d53700cdc32c17e504b118bb0d9190acc538f36ab9 +size 509627776 diff --git a/model-00010-of-00053.safetensors b/model-00010-of-00053.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..da5a9539953c0989c163f2ee76febf32058617b8 --- /dev/null +++ b/model-00010-of-00053.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:36e567b7a7dd89974dec31dc0d3da619adde8bf1893701534184ffa0fff4308d +size 509627776 diff --git a/model-00011-of-00053.safetensors b/model-00011-of-00053.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..719ce04dda05f59772b2bcf7faed0c0c03543d04 --- /dev/null +++ b/model-00011-of-00053.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:faf3900e70595a828e2ae25fbd66457776f8e010dbefac4603c3bd4b61f74250 +size 509627784 diff --git a/model-00012-of-00053.safetensors b/model-00012-of-00053.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..75da9d96331081620f7664c0f94d29d1845a272c --- /dev/null +++ b/model-00012-of-00053.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:57af8ad59c68582947cdc49793788585fb59f2c9125fb9a8ea4e263114141b4d +size 509627784 diff --git a/model-00013-of-00053.safetensors b/model-00013-of-00053.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..a98adcb372654a00a7063a4a31fc81c8004bb880 --- /dev/null +++ b/model-00013-of-00053.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:f7e0beb70cc8497cdb3c6f0e40aceb43628cb1acf0726a991ca3c3edccbfddd0 +size 509627784 diff --git a/model-00014-of-00053.safetensors b/model-00014-of-00053.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..92cdcf9717b1042915cdc8b65d05e658cf486240 --- /dev/null +++ b/model-00014-of-00053.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:2efad6727d0f2481d28d0fff16e1124afb16c0be8a38138dcc05181103c76caf +size 509627784 diff --git a/model-00015-of-00053.safetensors b/model-00015-of-00053.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..8215a6fd4da095d88b92229bd2c99b4d58f0565c --- /dev/null +++ b/model-00015-of-00053.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:2f52c006d867605aa04b3d231bfd68b2048c32d9fdb47354ecc386ae7b9dec47 +size 509627784 diff --git a/model-00016-of-00053.safetensors b/model-00016-of-00053.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..23abc3913ad63cffff51cb5e6f9c890ab0b4c1b4 --- /dev/null +++ b/model-00016-of-00053.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:2ffec8be26d80bedd6fbd189b9dfd7d7cd5080726aa003e71832ee052ea58c80 +size 509627784 diff --git a/model-00017-of-00053.safetensors b/model-00017-of-00053.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..410fa965c5db28b18dcf432c6527bf0400bd715d --- /dev/null +++ b/model-00017-of-00053.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:d3bcaa1640d8377bdea8ba64d014ad44230e01f97f89fbf273014d628fda1a26 +size 509627784 diff --git a/model-00018-of-00053.safetensors b/model-00018-of-00053.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..75cef8fdbc988f78abd53c822ad0e4678168c634 --- /dev/null +++ b/model-00018-of-00053.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:81f51e4aeee4908c8876173bb3447e5bb3d6250d5f17f123e1ae0d10eb6cbdea +size 509627784 diff --git a/model-00019-of-00053.safetensors b/model-00019-of-00053.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..93d5867afe613b442a854ba06b4cda855ab3404d --- /dev/null +++ b/model-00019-of-00053.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:f890a2a9d01d1e8047b5835fd8a08656668efcefcacdee0b714b12b6c2633983 +size 509627784 diff --git a/model-00020-of-00053.safetensors b/model-00020-of-00053.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..34d6b80e7f41596a5fa09b058706f1d7e4d67a97 --- /dev/null +++ b/model-00020-of-00053.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:0a01803357c5304810cebc5b0d189e534ae16b4a4dbae454b40478f5fffb1371 +size 509627784 diff --git a/model-00021-of-00053.safetensors b/model-00021-of-00053.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..4d7c78209906a8d7775fdfe0cddc50213485c5f3 --- /dev/null +++ b/model-00021-of-00053.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:b2b01e6eb39e31c0cea39ef96f35b603be2c5aeeb02a6967c241f97b6705b7bd +size 509627784 diff --git a/model-00022-of-00053.safetensors b/model-00022-of-00053.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..7dea7d8fdf9a0d439aa9bee92ed82b962d9c0554 --- /dev/null +++ b/model-00022-of-00053.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:ab8af7d8f04b1c75bbcc048c69faff291a08e37ef58e16e609b233641e9538d7 +size 509627784 diff --git a/model-00023-of-00053.safetensors b/model-00023-of-00053.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..bc35899354887eb7246f2d98e8f398095ebd44c2 --- /dev/null +++ b/model-00023-of-00053.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:1513886b3d1633384c595e55800888fa60fd1b397f0ec5a67acc74bb1e83561d +size 509627784 diff --git a/model-00024-of-00053.safetensors b/model-00024-of-00053.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..4adc177c7f0a1e26571f7543aa8ec15a9b0cb643 --- /dev/null +++ b/model-00024-of-00053.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:d19798e6d398e222fc32dffbebe3b8b8649cb1764c7b625781bc767c2276dfdf +size 509627784 diff --git a/model-00025-of-00053.safetensors b/model-00025-of-00053.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..ace0bf8086f83120a5e2a2be189907aea83da02e --- /dev/null +++ b/model-00025-of-00053.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:b21b152fd83c00c5721244302ca33964ddd127bbbb902a73f0a706da8583f081 +size 509627784 diff --git a/model-00026-of-00053.safetensors b/model-00026-of-00053.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..8c6ac11823f03c166ad02e3b0f54db1a74e75de4 --- /dev/null +++ b/model-00026-of-00053.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:b83954a006bd1b6b630537d6df9ee0a16a1074f05dbaeddd50367999325c1279 +size 509627784 diff --git a/model-00027-of-00053.safetensors b/model-00027-of-00053.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..5bdf3b5a2b14acc243942a6ec705af34cce53aed --- /dev/null +++ b/model-00027-of-00053.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:f9138a276e23b0a600f96108e258fae802927c78a87ef203c5ddb3e056efb372 +size 509627784 diff --git a/model-00028-of-00053.safetensors b/model-00028-of-00053.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..8ddfeb25c65b6d036e1a259d4f35bbf8078aff62 --- /dev/null +++ b/model-00028-of-00053.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:a416c5bf41f7435d85cbbfdbf91b888dad07ef20b596d7b6b5dacbd7f73e0589 +size 509627784 diff --git a/model-00029-of-00053.safetensors b/model-00029-of-00053.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..2b10f6e024dc0002c491dbbdfbef80c9e0f31625 --- /dev/null +++ b/model-00029-of-00053.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:4a2835db2f0f1c85a606368c66049fa6c999c76998527605fd9196bd1293b942 +size 509627784 diff --git a/model-00030-of-00053.safetensors b/model-00030-of-00053.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..0c745e1f80559486622742e0376245d782b2dbe9 --- /dev/null +++ b/model-00030-of-00053.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:a26cf49db2c43a807df9b96fd60723cfc312b125686cd6dc5fbcab5ef3edf663 +size 509627784 diff --git a/model-00031-of-00053.safetensors b/model-00031-of-00053.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..d27d4660489042463cffae76023e9e02bc7eacba --- /dev/null +++ b/model-00031-of-00053.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:4283579d6de131374aad9b01817efd94f8c648312871a94532f5ee37e890d89e +size 509627784 diff --git a/model-00032-of-00053.safetensors b/model-00032-of-00053.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..4c24dd3357bbacdb5a1967bc0b26da7889476524 --- /dev/null +++ b/model-00032-of-00053.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:f04a35c6d5a467651debd3825be8dd8fad743543fa8b08035f0f92546c1a6501 +size 509627784 diff --git a/model-00033-of-00053.safetensors b/model-00033-of-00053.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..41feaaf13f0b7172c8fef0506c1e5b8ebb0f158f --- /dev/null +++ b/model-00033-of-00053.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:8d485e4b7d10e2b3d4f21629cad73e43ee482430f87e6fd8e879820de663d86c +size 509627784 diff --git a/model-00034-of-00053.safetensors b/model-00034-of-00053.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..3dd1fa9011cbc8d7cf671e35eb973f34dff4d718 --- /dev/null +++ b/model-00034-of-00053.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:ff70d82f2c940f5f34fe8326eec36b8652fd2da13100668ae71c9e76996a07f2 +size 509627784 diff --git a/model-00035-of-00053.safetensors b/model-00035-of-00053.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..fc3de076ebe5d90ac2da51dc816dd83308ae6888 --- /dev/null +++ b/model-00035-of-00053.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:e47b9245228075bd20092f222577bff5d87969f69b287710adc9ff4d4ef2a352 +size 509627784 diff --git a/model-00036-of-00053.safetensors b/model-00036-of-00053.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..bf031a817f591a4b933c37397a976f4a63d38045 --- /dev/null +++ b/model-00036-of-00053.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:be53a1944bc3eb7f49c6c2b9619e9ad0e07ed0510dd56f0073a29b9a073a8d5b +size 509627784 diff --git a/model-00037-of-00053.safetensors b/model-00037-of-00053.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..0c6c881391af95a4def562ac65539e0216ad6919 --- /dev/null +++ b/model-00037-of-00053.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:c90ef8ada948aaa7f64c05725177cbdaa898c6adf3d415a4bfa50b31d837cc68 +size 509627784 diff --git a/model-00038-of-00053.safetensors b/model-00038-of-00053.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..d41207a9e6dae6693105f07dae294f50e21c1068 --- /dev/null +++ b/model-00038-of-00053.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:7bc3f8d35c454141c615080e10adee54bf21c837aa833564f439ab0c21eaba87 +size 509627784 diff --git a/model-00039-of-00053.safetensors b/model-00039-of-00053.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..6295cb3ef183734f871fbda1ffb25ca52b99d7e7 --- /dev/null +++ b/model-00039-of-00053.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:96d953fd741d9a11e65722d8a6e4b5340a2a735d572739135fe8369f05592bbb +size 509627784 diff --git a/model-00040-of-00053.safetensors b/model-00040-of-00053.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..0346015919fda8125d298ec30967777fbd817bfc --- /dev/null +++ b/model-00040-of-00053.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:a32a1414b6dfb6f62fdf0a33087414347df8ce6c6fa63023bc928e4f778bb38f +size 509627784 diff --git a/model-00041-of-00053.safetensors b/model-00041-of-00053.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..e85c7dbdec7176815288e22de51f20117c5fd428 --- /dev/null +++ b/model-00041-of-00053.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:47ea766706de28a75ef4748ee9cab1a40ced7a52fc653ff23a9e720dfcbee149 +size 509627784 diff --git a/model-00042-of-00053.safetensors b/model-00042-of-00053.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..dc572f52cb8376edb106cece886c8cea47223df6 --- /dev/null +++ b/model-00042-of-00053.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:8ddd88e02a819e3fbf20f3073f156cb28bb62e40d59354bc800a66b881f7f683 +size 509627784 diff --git a/model-00043-of-00053.safetensors b/model-00043-of-00053.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..8a06c7f52792d743c48bb6fcf252b0c8cb4f9195 --- /dev/null +++ b/model-00043-of-00053.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:098d2432e7fcdfcfee36125b2646f452d34f7edcb2ea385d3d7121c791465b18 +size 509627784 diff --git a/model-00044-of-00053.safetensors b/model-00044-of-00053.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..8266ba299c99e5e103974afc35c70df720e8bfcf --- /dev/null +++ b/model-00044-of-00053.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:22c62a6d62a83dfb88eeb7326c7fbb80e81bb4338fa2d89598a6e6aa81fac667 +size 509627784 diff --git a/model-00045-of-00053.safetensors b/model-00045-of-00053.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..4cfa565450dbcf5068763473f7e5b6142ce7d95b --- /dev/null +++ b/model-00045-of-00053.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:311abdd6508c135163dede36858fe08105eb0d5f90223cb4022dd863d7d2a9ef +size 509627784 diff --git a/model-00046-of-00053.safetensors b/model-00046-of-00053.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..3a66f906577ebc81d00151fb25bde08c999b4f33 --- /dev/null +++ b/model-00046-of-00053.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:fedd15a525b349bf62502a586cae8cdcb8ff5a30d8649205232bc1789b903231 +size 509627784 diff --git a/model-00047-of-00053.safetensors b/model-00047-of-00053.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..7ce96080501e2c07bc043048f2723c511be6069a --- /dev/null +++ b/model-00047-of-00053.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:2d5b6c8fb33331decd43e4267bb9fc01c46a45f1401a51dcc94986791bd1255b +size 509627784 diff --git a/model-00048-of-00053.safetensors b/model-00048-of-00053.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..0cefab9605fd9fff531cc18bd51bb27fd069c7f4 --- /dev/null +++ b/model-00048-of-00053.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:812fa1fec6fc80e39d8ae9bdc5443d2835771c422f196edafef7d59ce7783acb +size 509627784 diff --git a/model-00049-of-00053.safetensors b/model-00049-of-00053.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..ade20ec73f2a96f9834ba0b6fbc847fc07662ccb --- /dev/null +++ b/model-00049-of-00053.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:2a781a669e4b0cc7391516feed49baa327f4d5c01a00b4a2cbb15f6cbf225d2f +size 509627784 diff --git a/model-00050-of-00053.safetensors b/model-00050-of-00053.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..4074d083f263a34a1f54d50665fb717e281bccc9 --- /dev/null +++ b/model-00050-of-00053.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:38c4979939e7d4ac91e2d1ff4eade04d5591b9206eeea848c0d021278aa679fa +size 509627784 diff --git a/model-00051-of-00053.safetensors b/model-00051-of-00053.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..54ff46a19c0cde356e505f09169a6bfcd6b7af81 --- /dev/null +++ b/model-00051-of-00053.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:66a56d279278389b18eea6ed061158caf9c54d335b151e383628256e4df525bd +size 509627784 diff --git a/model-00052-of-00053.safetensors b/model-00052-of-00053.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..b04b6b9a0ba44e787ab5c753e1708ad86a2d2e64 --- /dev/null +++ b/model-00052-of-00053.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:99dcb247a6be3dad3126b2bb6099083f0e80ada5f7bb11198a8722d55189e80b +size 509627784 diff --git a/model-00053-of-00053.safetensors b/model-00053-of-00053.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..1367c1283c63fd47fe108c137e6fd4f700a65500 --- /dev/null +++ b/model-00053-of-00053.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:9804443bfcd03307bbfcb95528b5d980072aa6eb643543b97dcb685ea17f842f +size 1207969088 diff --git a/model.safetensors.index.json b/model.safetensors.index.json new file mode 100644 index 0000000000000000000000000000000000000000..7f06b691c8b4e17e17152ae227ccddd99d44cb2e --- /dev/null +++ b/model.safetensors.index.json @@ -0,0 +1,530 @@ +{ + "metadata": { + "total_size": 27708553216 + }, + "weight_map": { + "model.layers.0.input_layernorm.weight": "model-00001-of-00053.safetensors", + "model.layers.0.post_attention_layernorm.weight": "model-00001-of-00053.safetensors", + "model.layers.0.self_attn.q_proj.weight": "model-00001-of-00053.safetensors", + "model.layers.0.self_attn.k_proj.weight": "model-00001-of-00053.safetensors", + "model.layers.0.self_attn.v_proj.weight": "model-00001-of-00053.safetensors", + "model.layers.0.self_attn.o_proj.weight": "model-00001-of-00053.safetensors", + "model.layers.0.self_attn.rotary_emb.inv_freq": "model-00001-of-00053.safetensors", + "model.layers.0.mlp.gate_proj.weight": "model-00001-of-00053.safetensors", + "model.layers.0.mlp.up_proj.weight": "model-00001-of-00053.safetensors", + "model.layers.0.mlp.down_proj.weight": "model-00001-of-00053.safetensors", + "model.layers.1.input_layernorm.weight": "model-00002-of-00053.safetensors", + "model.layers.1.post_attention_layernorm.weight": "model-00002-of-00053.safetensors", + "model.layers.1.self_attn.q_proj.weight": "model-00002-of-00053.safetensors", + "model.layers.1.self_attn.k_proj.weight": "model-00002-of-00053.safetensors", + "model.layers.1.self_attn.v_proj.weight": "model-00002-of-00053.safetensors", + "model.layers.1.self_attn.o_proj.weight": "model-00002-of-00053.safetensors", + "model.layers.1.self_attn.rotary_emb.inv_freq": "model-00002-of-00053.safetensors", + "model.layers.1.mlp.gate_proj.weight": "model-00002-of-00053.safetensors", + "model.layers.1.mlp.up_proj.weight": "model-00002-of-00053.safetensors", + "model.layers.1.mlp.down_proj.weight": "model-00002-of-00053.safetensors", + "model.layers.2.input_layernorm.weight": "model-00003-of-00053.safetensors", + "model.layers.2.post_attention_layernorm.weight": "model-00003-of-00053.safetensors", + "model.layers.2.self_attn.q_proj.weight": "model-00003-of-00053.safetensors", + "model.layers.2.self_attn.k_proj.weight": "model-00003-of-00053.safetensors", + "model.layers.2.self_attn.v_proj.weight": "model-00003-of-00053.safetensors", + "model.layers.2.self_attn.o_proj.weight": "model-00003-of-00053.safetensors", + "model.layers.2.self_attn.rotary_emb.inv_freq": "model-00003-of-00053.safetensors", + "model.layers.2.mlp.gate_proj.weight": "model-00003-of-00053.safetensors", + "model.layers.2.mlp.up_proj.weight": "model-00003-of-00053.safetensors", + "model.layers.2.mlp.down_proj.weight": "model-00003-of-00053.safetensors", + "model.layers.3.input_layernorm.weight": "model-00004-of-00053.safetensors", + "model.layers.3.post_attention_layernorm.weight": "model-00004-of-00053.safetensors", + "model.layers.3.self_attn.q_proj.weight": "model-00004-of-00053.safetensors", + "model.layers.3.self_attn.k_proj.weight": "model-00004-of-00053.safetensors", + "model.layers.3.self_attn.v_proj.weight": "model-00004-of-00053.safetensors", + "model.layers.3.self_attn.o_proj.weight": "model-00004-of-00053.safetensors", + "model.layers.3.self_attn.rotary_emb.inv_freq": "model-00004-of-00053.safetensors", + "model.layers.3.mlp.gate_proj.weight": "model-00004-of-00053.safetensors", + "model.layers.3.mlp.up_proj.weight": "model-00004-of-00053.safetensors", + "model.layers.3.mlp.down_proj.weight": "model-00004-of-00053.safetensors", + "model.layers.4.input_layernorm.weight": "model-00005-of-00053.safetensors", + "model.layers.4.post_attention_layernorm.weight": "model-00005-of-00053.safetensors", + "model.layers.4.self_attn.q_proj.weight": "model-00005-of-00053.safetensors", + "model.layers.4.self_attn.k_proj.weight": "model-00005-of-00053.safetensors", + "model.layers.4.self_attn.v_proj.weight": "model-00005-of-00053.safetensors", + "model.layers.4.self_attn.o_proj.weight": "model-00005-of-00053.safetensors", + "model.layers.4.self_attn.rotary_emb.inv_freq": "model-00005-of-00053.safetensors", + "model.layers.4.mlp.gate_proj.weight": "model-00005-of-00053.safetensors", + "model.layers.4.mlp.up_proj.weight": "model-00005-of-00053.safetensors", + "model.layers.4.mlp.down_proj.weight": "model-00005-of-00053.safetensors", + "model.layers.5.input_layernorm.weight": "model-00006-of-00053.safetensors", + "model.layers.5.post_attention_layernorm.weight": "model-00006-of-00053.safetensors", + "model.layers.5.self_attn.q_proj.weight": "model-00006-of-00053.safetensors", + "model.layers.5.self_attn.k_proj.weight": "model-00006-of-00053.safetensors", + "model.layers.5.self_attn.v_proj.weight": "model-00006-of-00053.safetensors", + "model.layers.5.self_attn.o_proj.weight": "model-00006-of-00053.safetensors", + "model.layers.5.self_attn.rotary_emb.inv_freq": "model-00006-of-00053.safetensors", + "model.layers.5.mlp.gate_proj.weight": "model-00006-of-00053.safetensors", + "model.layers.5.mlp.up_proj.weight": "model-00006-of-00053.safetensors", + "model.layers.5.mlp.down_proj.weight": "model-00006-of-00053.safetensors", + "model.layers.6.input_layernorm.weight": "model-00007-of-00053.safetensors", + "model.layers.6.post_attention_layernorm.weight": "model-00007-of-00053.safetensors", + "model.layers.6.self_attn.q_proj.weight": "model-00007-of-00053.safetensors", + "model.layers.6.self_attn.k_proj.weight": "model-00007-of-00053.safetensors", + "model.layers.6.self_attn.v_proj.weight": "model-00007-of-00053.safetensors", + "model.layers.6.self_attn.o_proj.weight": "model-00007-of-00053.safetensors", + "model.layers.6.self_attn.rotary_emb.inv_freq": "model-00007-of-00053.safetensors", + "model.layers.6.mlp.gate_proj.weight": "model-00007-of-00053.safetensors", + "model.layers.6.mlp.up_proj.weight": "model-00007-of-00053.safetensors", + "model.layers.6.mlp.down_proj.weight": "model-00007-of-00053.safetensors", + "model.layers.7.input_layernorm.weight": "model-00008-of-00053.safetensors", + "model.layers.7.post_attention_layernorm.weight": "model-00008-of-00053.safetensors", + "model.layers.7.self_attn.q_proj.weight": "model-00008-of-00053.safetensors", + "model.layers.7.self_attn.k_proj.weight": "model-00008-of-00053.safetensors", + "model.layers.7.self_attn.v_proj.weight": "model-00008-of-00053.safetensors", + "model.layers.7.self_attn.o_proj.weight": "model-00008-of-00053.safetensors", + "model.layers.7.self_attn.rotary_emb.inv_freq": "model-00008-of-00053.safetensors", + "model.layers.7.mlp.gate_proj.weight": "model-00008-of-00053.safetensors", + "model.layers.7.mlp.up_proj.weight": "model-00008-of-00053.safetensors", + "model.layers.7.mlp.down_proj.weight": "model-00008-of-00053.safetensors", + "model.layers.8.input_layernorm.weight": "model-00009-of-00053.safetensors", + "model.layers.8.post_attention_layernorm.weight": "model-00009-of-00053.safetensors", + "model.layers.8.self_attn.q_proj.weight": "model-00009-of-00053.safetensors", + "model.layers.8.self_attn.k_proj.weight": "model-00009-of-00053.safetensors", + "model.layers.8.self_attn.v_proj.weight": "model-00009-of-00053.safetensors", + "model.layers.8.self_attn.o_proj.weight": "model-00009-of-00053.safetensors", + "model.layers.8.self_attn.rotary_emb.inv_freq": "model-00009-of-00053.safetensors", + "model.layers.8.mlp.gate_proj.weight": "model-00009-of-00053.safetensors", + "model.layers.8.mlp.up_proj.weight": "model-00009-of-00053.safetensors", + "model.layers.8.mlp.down_proj.weight": "model-00009-of-00053.safetensors", + "model.layers.9.input_layernorm.weight": "model-00010-of-00053.safetensors", + "model.layers.9.post_attention_layernorm.weight": "model-00010-of-00053.safetensors", + "model.layers.9.self_attn.q_proj.weight": "model-00010-of-00053.safetensors", + "model.layers.9.self_attn.k_proj.weight": "model-00010-of-00053.safetensors", + "model.layers.9.self_attn.v_proj.weight": "model-00010-of-00053.safetensors", + "model.layers.9.self_attn.o_proj.weight": "model-00010-of-00053.safetensors", + "model.layers.9.self_attn.rotary_emb.inv_freq": "model-00010-of-00053.safetensors", + "model.layers.9.mlp.gate_proj.weight": "model-00010-of-00053.safetensors", + "model.layers.9.mlp.up_proj.weight": "model-00010-of-00053.safetensors", + "model.layers.9.mlp.down_proj.weight": "model-00010-of-00053.safetensors", + "model.layers.10.input_layernorm.weight": "model-00011-of-00053.safetensors", + "model.layers.10.post_attention_layernorm.weight": "model-00011-of-00053.safetensors", + "model.layers.10.self_attn.q_proj.weight": "model-00011-of-00053.safetensors", + "model.layers.10.self_attn.k_proj.weight": "model-00011-of-00053.safetensors", + "model.layers.10.self_attn.v_proj.weight": "model-00011-of-00053.safetensors", + "model.layers.10.self_attn.o_proj.weight": "model-00011-of-00053.safetensors", + "model.layers.10.self_attn.rotary_emb.inv_freq": "model-00011-of-00053.safetensors", + "model.layers.10.mlp.gate_proj.weight": "model-00011-of-00053.safetensors", + "model.layers.10.mlp.up_proj.weight": "model-00011-of-00053.safetensors", + "model.layers.10.mlp.down_proj.weight": "model-00011-of-00053.safetensors", + "model.layers.11.input_layernorm.weight": "model-00012-of-00053.safetensors", + "model.layers.11.post_attention_layernorm.weight": "model-00012-of-00053.safetensors", + "model.layers.11.self_attn.q_proj.weight": "model-00012-of-00053.safetensors", + "model.layers.11.self_attn.k_proj.weight": "model-00012-of-00053.safetensors", + "model.layers.11.self_attn.v_proj.weight": "model-00012-of-00053.safetensors", + "model.layers.11.self_attn.o_proj.weight": "model-00012-of-00053.safetensors", + "model.layers.11.self_attn.rotary_emb.inv_freq": "model-00012-of-00053.safetensors", + "model.layers.11.mlp.gate_proj.weight": "model-00012-of-00053.safetensors", + "model.layers.11.mlp.up_proj.weight": "model-00012-of-00053.safetensors", + "model.layers.11.mlp.down_proj.weight": "model-00012-of-00053.safetensors", + "model.layers.12.input_layernorm.weight": "model-00013-of-00053.safetensors", + "model.layers.12.post_attention_layernorm.weight": "model-00013-of-00053.safetensors", + "model.layers.12.self_attn.q_proj.weight": "model-00013-of-00053.safetensors", + "model.layers.12.self_attn.k_proj.weight": "model-00013-of-00053.safetensors", + "model.layers.12.self_attn.v_proj.weight": "model-00013-of-00053.safetensors", + "model.layers.12.self_attn.o_proj.weight": "model-00013-of-00053.safetensors", + "model.layers.12.self_attn.rotary_emb.inv_freq": "model-00013-of-00053.safetensors", + "model.layers.12.mlp.gate_proj.weight": "model-00013-of-00053.safetensors", + "model.layers.12.mlp.up_proj.weight": "model-00013-of-00053.safetensors", + "model.layers.12.mlp.down_proj.weight": "model-00013-of-00053.safetensors", + "model.layers.13.input_layernorm.weight": "model-00014-of-00053.safetensors", + "model.layers.13.post_attention_layernorm.weight": "model-00014-of-00053.safetensors", + "model.layers.13.self_attn.q_proj.weight": "model-00014-of-00053.safetensors", + "model.layers.13.self_attn.k_proj.weight": "model-00014-of-00053.safetensors", + "model.layers.13.self_attn.v_proj.weight": "model-00014-of-00053.safetensors", + "model.layers.13.self_attn.o_proj.weight": "model-00014-of-00053.safetensors", + "model.layers.13.self_attn.rotary_emb.inv_freq": "model-00014-of-00053.safetensors", + "model.layers.13.mlp.gate_proj.weight": "model-00014-of-00053.safetensors", + "model.layers.13.mlp.up_proj.weight": "model-00014-of-00053.safetensors", + "model.layers.13.mlp.down_proj.weight": "model-00014-of-00053.safetensors", + "model.layers.14.input_layernorm.weight": "model-00015-of-00053.safetensors", + "model.layers.14.post_attention_layernorm.weight": "model-00015-of-00053.safetensors", + "model.layers.14.self_attn.q_proj.weight": "model-00015-of-00053.safetensors", + "model.layers.14.self_attn.k_proj.weight": "model-00015-of-00053.safetensors", + "model.layers.14.self_attn.v_proj.weight": "model-00015-of-00053.safetensors", + "model.layers.14.self_attn.o_proj.weight": "model-00015-of-00053.safetensors", + "model.layers.14.self_attn.rotary_emb.inv_freq": "model-00015-of-00053.safetensors", + "model.layers.14.mlp.gate_proj.weight": "model-00015-of-00053.safetensors", + "model.layers.14.mlp.up_proj.weight": "model-00015-of-00053.safetensors", + "model.layers.14.mlp.down_proj.weight": "model-00015-of-00053.safetensors", + "model.layers.15.input_layernorm.weight": "model-00016-of-00053.safetensors", + "model.layers.15.post_attention_layernorm.weight": "model-00016-of-00053.safetensors", + "model.layers.15.self_attn.q_proj.weight": "model-00016-of-00053.safetensors", + "model.layers.15.self_attn.k_proj.weight": "model-00016-of-00053.safetensors", + "model.layers.15.self_attn.v_proj.weight": "model-00016-of-00053.safetensors", + "model.layers.15.self_attn.o_proj.weight": "model-00016-of-00053.safetensors", + "model.layers.15.self_attn.rotary_emb.inv_freq": "model-00016-of-00053.safetensors", + "model.layers.15.mlp.gate_proj.weight": "model-00016-of-00053.safetensors", + "model.layers.15.mlp.up_proj.weight": "model-00016-of-00053.safetensors", + "model.layers.15.mlp.down_proj.weight": "model-00016-of-00053.safetensors", + "model.layers.16.input_layernorm.weight": "model-00017-of-00053.safetensors", + "model.layers.16.post_attention_layernorm.weight": "model-00017-of-00053.safetensors", + "model.layers.16.self_attn.q_proj.weight": "model-00017-of-00053.safetensors", + "model.layers.16.self_attn.k_proj.weight": "model-00017-of-00053.safetensors", + "model.layers.16.self_attn.v_proj.weight": "model-00017-of-00053.safetensors", + "model.layers.16.self_attn.o_proj.weight": "model-00017-of-00053.safetensors", + "model.layers.16.self_attn.rotary_emb.inv_freq": "model-00017-of-00053.safetensors", + "model.layers.16.mlp.gate_proj.weight": "model-00017-of-00053.safetensors", + "model.layers.16.mlp.up_proj.weight": "model-00017-of-00053.safetensors", + "model.layers.16.mlp.down_proj.weight": "model-00017-of-00053.safetensors", + "model.layers.17.input_layernorm.weight": "model-00018-of-00053.safetensors", + "model.layers.17.post_attention_layernorm.weight": "model-00018-of-00053.safetensors", + "model.layers.17.self_attn.q_proj.weight": "model-00018-of-00053.safetensors", + "model.layers.17.self_attn.k_proj.weight": "model-00018-of-00053.safetensors", + "model.layers.17.self_attn.v_proj.weight": "model-00018-of-00053.safetensors", + "model.layers.17.self_attn.o_proj.weight": "model-00018-of-00053.safetensors", + "model.layers.17.self_attn.rotary_emb.inv_freq": "model-00018-of-00053.safetensors", + "model.layers.17.mlp.gate_proj.weight": "model-00018-of-00053.safetensors", + "model.layers.17.mlp.up_proj.weight": "model-00018-of-00053.safetensors", + "model.layers.17.mlp.down_proj.weight": "model-00018-of-00053.safetensors", + "model.layers.18.input_layernorm.weight": "model-00019-of-00053.safetensors", + "model.layers.18.post_attention_layernorm.weight": "model-00019-of-00053.safetensors", + "model.layers.18.self_attn.q_proj.weight": "model-00019-of-00053.safetensors", + "model.layers.18.self_attn.k_proj.weight": "model-00019-of-00053.safetensors", + "model.layers.18.self_attn.v_proj.weight": "model-00019-of-00053.safetensors", + "model.layers.18.self_attn.o_proj.weight": "model-00019-of-00053.safetensors", + "model.layers.18.self_attn.rotary_emb.inv_freq": "model-00019-of-00053.safetensors", + "model.layers.18.mlp.gate_proj.weight": "model-00019-of-00053.safetensors", + "model.layers.18.mlp.up_proj.weight": "model-00019-of-00053.safetensors", + "model.layers.18.mlp.down_proj.weight": "model-00019-of-00053.safetensors", + "model.layers.19.input_layernorm.weight": "model-00020-of-00053.safetensors", + "model.layers.19.post_attention_layernorm.weight": "model-00020-of-00053.safetensors", + "model.layers.19.self_attn.q_proj.weight": "model-00020-of-00053.safetensors", + "model.layers.19.self_attn.k_proj.weight": "model-00020-of-00053.safetensors", + "model.layers.19.self_attn.v_proj.weight": "model-00020-of-00053.safetensors", + "model.layers.19.self_attn.o_proj.weight": "model-00020-of-00053.safetensors", + "model.layers.19.self_attn.rotary_emb.inv_freq": "model-00020-of-00053.safetensors", + "model.layers.19.mlp.gate_proj.weight": "model-00020-of-00053.safetensors", + "model.layers.19.mlp.up_proj.weight": "model-00020-of-00053.safetensors", + "model.layers.19.mlp.down_proj.weight": "model-00020-of-00053.safetensors", + "model.layers.20.input_layernorm.weight": "model-00021-of-00053.safetensors", + "model.layers.20.post_attention_layernorm.weight": "model-00021-of-00053.safetensors", + "model.layers.20.self_attn.q_proj.weight": "model-00021-of-00053.safetensors", + "model.layers.20.self_attn.k_proj.weight": "model-00021-of-00053.safetensors", + "model.layers.20.self_attn.v_proj.weight": "model-00021-of-00053.safetensors", + "model.layers.20.self_attn.o_proj.weight": "model-00021-of-00053.safetensors", + "model.layers.20.self_attn.rotary_emb.inv_freq": "model-00021-of-00053.safetensors", + "model.layers.20.mlp.gate_proj.weight": "model-00021-of-00053.safetensors", + "model.layers.20.mlp.up_proj.weight": "model-00021-of-00053.safetensors", + "model.layers.20.mlp.down_proj.weight": "model-00021-of-00053.safetensors", + "model.layers.21.input_layernorm.weight": "model-00022-of-00053.safetensors", + "model.layers.21.post_attention_layernorm.weight": "model-00022-of-00053.safetensors", + "model.layers.21.self_attn.q_proj.weight": "model-00022-of-00053.safetensors", + "model.layers.21.self_attn.k_proj.weight": "model-00022-of-00053.safetensors", + "model.layers.21.self_attn.v_proj.weight": "model-00022-of-00053.safetensors", + "model.layers.21.self_attn.o_proj.weight": "model-00022-of-00053.safetensors", + "model.layers.21.self_attn.rotary_emb.inv_freq": "model-00022-of-00053.safetensors", + "model.layers.21.mlp.gate_proj.weight": "model-00022-of-00053.safetensors", + "model.layers.21.mlp.up_proj.weight": "model-00022-of-00053.safetensors", + "model.layers.21.mlp.down_proj.weight": "model-00022-of-00053.safetensors", + "model.layers.22.input_layernorm.weight": "model-00023-of-00053.safetensors", + "model.layers.22.post_attention_layernorm.weight": "model-00023-of-00053.safetensors", + "model.layers.22.self_attn.q_proj.weight": "model-00023-of-00053.safetensors", + "model.layers.22.self_attn.k_proj.weight": "model-00023-of-00053.safetensors", + "model.layers.22.self_attn.v_proj.weight": "model-00023-of-00053.safetensors", + "model.layers.22.self_attn.o_proj.weight": "model-00023-of-00053.safetensors", + "model.layers.22.self_attn.rotary_emb.inv_freq": "model-00023-of-00053.safetensors", + "model.layers.22.mlp.gate_proj.weight": "model-00023-of-00053.safetensors", + "model.layers.22.mlp.up_proj.weight": "model-00023-of-00053.safetensors", + "model.layers.22.mlp.down_proj.weight": "model-00023-of-00053.safetensors", + "model.layers.23.input_layernorm.weight": "model-00024-of-00053.safetensors", + "model.layers.23.post_attention_layernorm.weight": "model-00024-of-00053.safetensors", + "model.layers.23.self_attn.q_proj.weight": "model-00024-of-00053.safetensors", + "model.layers.23.self_attn.k_proj.weight": "model-00024-of-00053.safetensors", + "model.layers.23.self_attn.v_proj.weight": "model-00024-of-00053.safetensors", + "model.layers.23.self_attn.o_proj.weight": "model-00024-of-00053.safetensors", + "model.layers.23.self_attn.rotary_emb.inv_freq": "model-00024-of-00053.safetensors", + "model.layers.23.mlp.gate_proj.weight": "model-00024-of-00053.safetensors", + "model.layers.23.mlp.up_proj.weight": "model-00024-of-00053.safetensors", + "model.layers.23.mlp.down_proj.weight": "model-00024-of-00053.safetensors", + "model.layers.24.input_layernorm.weight": "model-00025-of-00053.safetensors", + "model.layers.24.post_attention_layernorm.weight": "model-00025-of-00053.safetensors", + "model.layers.24.self_attn.q_proj.weight": "model-00025-of-00053.safetensors", + "model.layers.24.self_attn.k_proj.weight": "model-00025-of-00053.safetensors", + "model.layers.24.self_attn.v_proj.weight": "model-00025-of-00053.safetensors", + "model.layers.24.self_attn.o_proj.weight": "model-00025-of-00053.safetensors", + "model.layers.24.self_attn.rotary_emb.inv_freq": "model-00025-of-00053.safetensors", + "model.layers.24.mlp.gate_proj.weight": "model-00025-of-00053.safetensors", + "model.layers.24.mlp.up_proj.weight": "model-00025-of-00053.safetensors", + "model.layers.24.mlp.down_proj.weight": "model-00025-of-00053.safetensors", + "model.layers.25.input_layernorm.weight": "model-00026-of-00053.safetensors", + "model.layers.25.post_attention_layernorm.weight": "model-00026-of-00053.safetensors", + "model.layers.25.self_attn.q_proj.weight": "model-00026-of-00053.safetensors", + "model.layers.25.self_attn.k_proj.weight": "model-00026-of-00053.safetensors", + "model.layers.25.self_attn.v_proj.weight": "model-00026-of-00053.safetensors", + "model.layers.25.self_attn.o_proj.weight": "model-00026-of-00053.safetensors", + "model.layers.25.self_attn.rotary_emb.inv_freq": "model-00026-of-00053.safetensors", + "model.layers.25.mlp.gate_proj.weight": "model-00026-of-00053.safetensors", + "model.layers.25.mlp.up_proj.weight": "model-00026-of-00053.safetensors", + "model.layers.25.mlp.down_proj.weight": "model-00026-of-00053.safetensors", + "model.layers.26.input_layernorm.weight": "model-00027-of-00053.safetensors", + "model.layers.26.post_attention_layernorm.weight": "model-00027-of-00053.safetensors", + "model.layers.26.self_attn.q_proj.weight": "model-00027-of-00053.safetensors", + "model.layers.26.self_attn.k_proj.weight": "model-00027-of-00053.safetensors", + "model.layers.26.self_attn.v_proj.weight": "model-00027-of-00053.safetensors", + "model.layers.26.self_attn.o_proj.weight": "model-00027-of-00053.safetensors", + "model.layers.26.self_attn.rotary_emb.inv_freq": "model-00027-of-00053.safetensors", + "model.layers.26.mlp.gate_proj.weight": "model-00027-of-00053.safetensors", + "model.layers.26.mlp.up_proj.weight": "model-00027-of-00053.safetensors", + "model.layers.26.mlp.down_proj.weight": "model-00027-of-00053.safetensors", + "model.layers.27.input_layernorm.weight": "model-00028-of-00053.safetensors", + "model.layers.27.post_attention_layernorm.weight": "model-00028-of-00053.safetensors", + "model.layers.27.self_attn.q_proj.weight": "model-00028-of-00053.safetensors", + "model.layers.27.self_attn.k_proj.weight": "model-00028-of-00053.safetensors", + "model.layers.27.self_attn.v_proj.weight": "model-00028-of-00053.safetensors", + "model.layers.27.self_attn.o_proj.weight": "model-00028-of-00053.safetensors", + "model.layers.27.self_attn.rotary_emb.inv_freq": "model-00028-of-00053.safetensors", + "model.layers.27.mlp.gate_proj.weight": "model-00028-of-00053.safetensors", + "model.layers.27.mlp.up_proj.weight": "model-00028-of-00053.safetensors", + "model.layers.27.mlp.down_proj.weight": "model-00028-of-00053.safetensors", + "model.layers.28.input_layernorm.weight": "model-00029-of-00053.safetensors", + "model.layers.28.post_attention_layernorm.weight": "model-00029-of-00053.safetensors", + "model.layers.28.self_attn.q_proj.weight": "model-00029-of-00053.safetensors", + "model.layers.28.self_attn.k_proj.weight": "model-00029-of-00053.safetensors", + "model.layers.28.self_attn.v_proj.weight": "model-00029-of-00053.safetensors", + "model.layers.28.self_attn.o_proj.weight": "model-00029-of-00053.safetensors", + "model.layers.28.self_attn.rotary_emb.inv_freq": "model-00029-of-00053.safetensors", + "model.layers.28.mlp.gate_proj.weight": "model-00029-of-00053.safetensors", + "model.layers.28.mlp.up_proj.weight": "model-00029-of-00053.safetensors", + "model.layers.28.mlp.down_proj.weight": "model-00029-of-00053.safetensors", + "model.layers.29.input_layernorm.weight": "model-00030-of-00053.safetensors", + "model.layers.29.post_attention_layernorm.weight": "model-00030-of-00053.safetensors", + "model.layers.29.self_attn.q_proj.weight": "model-00030-of-00053.safetensors", + "model.layers.29.self_attn.k_proj.weight": "model-00030-of-00053.safetensors", + "model.layers.29.self_attn.v_proj.weight": "model-00030-of-00053.safetensors", + "model.layers.29.self_attn.o_proj.weight": "model-00030-of-00053.safetensors", + "model.layers.29.self_attn.rotary_emb.inv_freq": "model-00030-of-00053.safetensors", + "model.layers.29.mlp.gate_proj.weight": "model-00030-of-00053.safetensors", + "model.layers.29.mlp.up_proj.weight": "model-00030-of-00053.safetensors", + "model.layers.29.mlp.down_proj.weight": "model-00030-of-00053.safetensors", + "model.layers.30.input_layernorm.weight": "model-00031-of-00053.safetensors", + "model.layers.30.post_attention_layernorm.weight": "model-00031-of-00053.safetensors", + "model.layers.30.self_attn.q_proj.weight": "model-00031-of-00053.safetensors", + "model.layers.30.self_attn.k_proj.weight": "model-00031-of-00053.safetensors", + "model.layers.30.self_attn.v_proj.weight": "model-00031-of-00053.safetensors", + "model.layers.30.self_attn.o_proj.weight": "model-00031-of-00053.safetensors", + "model.layers.30.self_attn.rotary_emb.inv_freq": "model-00031-of-00053.safetensors", + "model.layers.30.mlp.gate_proj.weight": "model-00031-of-00053.safetensors", + "model.layers.30.mlp.up_proj.weight": "model-00031-of-00053.safetensors", + "model.layers.30.mlp.down_proj.weight": "model-00031-of-00053.safetensors", + "model.layers.31.input_layernorm.weight": "model-00032-of-00053.safetensors", + "model.layers.31.post_attention_layernorm.weight": "model-00032-of-00053.safetensors", + "model.layers.31.self_attn.q_proj.weight": "model-00032-of-00053.safetensors", + "model.layers.31.self_attn.k_proj.weight": "model-00032-of-00053.safetensors", + "model.layers.31.self_attn.v_proj.weight": "model-00032-of-00053.safetensors", + "model.layers.31.self_attn.o_proj.weight": "model-00032-of-00053.safetensors", + "model.layers.31.self_attn.rotary_emb.inv_freq": "model-00032-of-00053.safetensors", + "model.layers.31.mlp.gate_proj.weight": "model-00032-of-00053.safetensors", + "model.layers.31.mlp.up_proj.weight": "model-00032-of-00053.safetensors", + "model.layers.31.mlp.down_proj.weight": "model-00032-of-00053.safetensors", + "model.layers.32.input_layernorm.weight": "model-00033-of-00053.safetensors", + "model.layers.32.post_attention_layernorm.weight": "model-00033-of-00053.safetensors", + "model.layers.32.self_attn.q_proj.weight": "model-00033-of-00053.safetensors", + "model.layers.32.self_attn.k_proj.weight": "model-00033-of-00053.safetensors", + "model.layers.32.self_attn.v_proj.weight": "model-00033-of-00053.safetensors", + "model.layers.32.self_attn.o_proj.weight": "model-00033-of-00053.safetensors", + "model.layers.32.self_attn.rotary_emb.inv_freq": "model-00033-of-00053.safetensors", + "model.layers.32.mlp.gate_proj.weight": "model-00033-of-00053.safetensors", + "model.layers.32.mlp.up_proj.weight": "model-00033-of-00053.safetensors", + "model.layers.32.mlp.down_proj.weight": "model-00033-of-00053.safetensors", + "model.layers.33.input_layernorm.weight": "model-00034-of-00053.safetensors", + "model.layers.33.post_attention_layernorm.weight": "model-00034-of-00053.safetensors", + "model.layers.33.self_attn.q_proj.weight": "model-00034-of-00053.safetensors", + "model.layers.33.self_attn.k_proj.weight": "model-00034-of-00053.safetensors", + "model.layers.33.self_attn.v_proj.weight": "model-00034-of-00053.safetensors", + "model.layers.33.self_attn.o_proj.weight": "model-00034-of-00053.safetensors", + "model.layers.33.self_attn.rotary_emb.inv_freq": "model-00034-of-00053.safetensors", + "model.layers.33.mlp.gate_proj.weight": "model-00034-of-00053.safetensors", + "model.layers.33.mlp.up_proj.weight": "model-00034-of-00053.safetensors", + "model.layers.33.mlp.down_proj.weight": "model-00034-of-00053.safetensors", + "model.layers.34.input_layernorm.weight": "model-00035-of-00053.safetensors", + "model.layers.34.post_attention_layernorm.weight": "model-00035-of-00053.safetensors", + "model.layers.34.self_attn.q_proj.weight": "model-00035-of-00053.safetensors", + "model.layers.34.self_attn.k_proj.weight": "model-00035-of-00053.safetensors", + "model.layers.34.self_attn.v_proj.weight": "model-00035-of-00053.safetensors", + "model.layers.34.self_attn.o_proj.weight": "model-00035-of-00053.safetensors", + "model.layers.34.self_attn.rotary_emb.inv_freq": "model-00035-of-00053.safetensors", + "model.layers.34.mlp.gate_proj.weight": "model-00035-of-00053.safetensors", + "model.layers.34.mlp.up_proj.weight": "model-00035-of-00053.safetensors", + "model.layers.34.mlp.down_proj.weight": "model-00035-of-00053.safetensors", + "model.layers.35.input_layernorm.weight": "model-00036-of-00053.safetensors", + "model.layers.35.post_attention_layernorm.weight": "model-00036-of-00053.safetensors", + "model.layers.35.self_attn.q_proj.weight": "model-00036-of-00053.safetensors", + "model.layers.35.self_attn.k_proj.weight": "model-00036-of-00053.safetensors", + "model.layers.35.self_attn.v_proj.weight": "model-00036-of-00053.safetensors", + "model.layers.35.self_attn.o_proj.weight": "model-00036-of-00053.safetensors", + "model.layers.35.self_attn.rotary_emb.inv_freq": "model-00036-of-00053.safetensors", + "model.layers.35.mlp.gate_proj.weight": "model-00036-of-00053.safetensors", + "model.layers.35.mlp.up_proj.weight": "model-00036-of-00053.safetensors", + "model.layers.35.mlp.down_proj.weight": "model-00036-of-00053.safetensors", + "model.layers.36.input_layernorm.weight": "model-00037-of-00053.safetensors", + "model.layers.36.post_attention_layernorm.weight": "model-00037-of-00053.safetensors", + "model.layers.36.self_attn.q_proj.weight": "model-00037-of-00053.safetensors", + "model.layers.36.self_attn.k_proj.weight": "model-00037-of-00053.safetensors", + "model.layers.36.self_attn.v_proj.weight": "model-00037-of-00053.safetensors", + "model.layers.36.self_attn.o_proj.weight": "model-00037-of-00053.safetensors", + "model.layers.36.self_attn.rotary_emb.inv_freq": "model-00037-of-00053.safetensors", + "model.layers.36.mlp.gate_proj.weight": "model-00037-of-00053.safetensors", + "model.layers.36.mlp.up_proj.weight": "model-00037-of-00053.safetensors", + "model.layers.36.mlp.down_proj.weight": "model-00037-of-00053.safetensors", + "model.layers.37.input_layernorm.weight": "model-00038-of-00053.safetensors", + "model.layers.37.post_attention_layernorm.weight": "model-00038-of-00053.safetensors", + "model.layers.37.self_attn.q_proj.weight": "model-00038-of-00053.safetensors", + "model.layers.37.self_attn.k_proj.weight": "model-00038-of-00053.safetensors", + "model.layers.37.self_attn.v_proj.weight": "model-00038-of-00053.safetensors", + "model.layers.37.self_attn.o_proj.weight": "model-00038-of-00053.safetensors", + "model.layers.37.self_attn.rotary_emb.inv_freq": "model-00038-of-00053.safetensors", + "model.layers.37.mlp.gate_proj.weight": "model-00038-of-00053.safetensors", + "model.layers.37.mlp.up_proj.weight": "model-00038-of-00053.safetensors", + "model.layers.37.mlp.down_proj.weight": "model-00038-of-00053.safetensors", + "model.layers.38.input_layernorm.weight": "model-00039-of-00053.safetensors", + "model.layers.38.post_attention_layernorm.weight": "model-00039-of-00053.safetensors", + "model.layers.38.self_attn.q_proj.weight": "model-00039-of-00053.safetensors", + "model.layers.38.self_attn.k_proj.weight": "model-00039-of-00053.safetensors", + "model.layers.38.self_attn.v_proj.weight": "model-00039-of-00053.safetensors", + "model.layers.38.self_attn.o_proj.weight": "model-00039-of-00053.safetensors", + "model.layers.38.self_attn.rotary_emb.inv_freq": "model-00039-of-00053.safetensors", + "model.layers.38.mlp.gate_proj.weight": "model-00039-of-00053.safetensors", + "model.layers.38.mlp.up_proj.weight": "model-00039-of-00053.safetensors", + "model.layers.38.mlp.down_proj.weight": "model-00039-of-00053.safetensors", + "model.layers.39.input_layernorm.weight": "model-00040-of-00053.safetensors", + "model.layers.39.post_attention_layernorm.weight": "model-00040-of-00053.safetensors", + "model.layers.39.self_attn.q_proj.weight": "model-00040-of-00053.safetensors", + "model.layers.39.self_attn.k_proj.weight": "model-00040-of-00053.safetensors", + "model.layers.39.self_attn.v_proj.weight": "model-00040-of-00053.safetensors", + "model.layers.39.self_attn.o_proj.weight": "model-00040-of-00053.safetensors", + "model.layers.39.self_attn.rotary_emb.inv_freq": "model-00040-of-00053.safetensors", + "model.layers.39.mlp.gate_proj.weight": "model-00040-of-00053.safetensors", + "model.layers.39.mlp.up_proj.weight": "model-00040-of-00053.safetensors", + "model.layers.39.mlp.down_proj.weight": "model-00040-of-00053.safetensors", + "model.layers.40.input_layernorm.weight": "model-00041-of-00053.safetensors", + "model.layers.40.post_attention_layernorm.weight": "model-00041-of-00053.safetensors", + "model.layers.40.self_attn.q_proj.weight": "model-00041-of-00053.safetensors", + "model.layers.40.self_attn.k_proj.weight": "model-00041-of-00053.safetensors", + "model.layers.40.self_attn.v_proj.weight": "model-00041-of-00053.safetensors", + "model.layers.40.self_attn.o_proj.weight": "model-00041-of-00053.safetensors", + "model.layers.40.self_attn.rotary_emb.inv_freq": "model-00041-of-00053.safetensors", + "model.layers.40.mlp.gate_proj.weight": "model-00041-of-00053.safetensors", + "model.layers.40.mlp.up_proj.weight": "model-00041-of-00053.safetensors", + "model.layers.40.mlp.down_proj.weight": "model-00041-of-00053.safetensors", + "model.layers.41.input_layernorm.weight": "model-00042-of-00053.safetensors", + "model.layers.41.post_attention_layernorm.weight": "model-00042-of-00053.safetensors", + "model.layers.41.self_attn.q_proj.weight": "model-00042-of-00053.safetensors", + "model.layers.41.self_attn.k_proj.weight": "model-00042-of-00053.safetensors", + "model.layers.41.self_attn.v_proj.weight": "model-00042-of-00053.safetensors", + "model.layers.41.self_attn.o_proj.weight": "model-00042-of-00053.safetensors", + "model.layers.41.self_attn.rotary_emb.inv_freq": "model-00042-of-00053.safetensors", + "model.layers.41.mlp.gate_proj.weight": "model-00042-of-00053.safetensors", + "model.layers.41.mlp.up_proj.weight": "model-00042-of-00053.safetensors", + "model.layers.41.mlp.down_proj.weight": "model-00042-of-00053.safetensors", + "model.layers.42.input_layernorm.weight": "model-00043-of-00053.safetensors", + "model.layers.42.post_attention_layernorm.weight": "model-00043-of-00053.safetensors", + "model.layers.42.self_attn.q_proj.weight": "model-00043-of-00053.safetensors", + "model.layers.42.self_attn.k_proj.weight": "model-00043-of-00053.safetensors", + "model.layers.42.self_attn.v_proj.weight": "model-00043-of-00053.safetensors", + "model.layers.42.self_attn.o_proj.weight": "model-00043-of-00053.safetensors", + "model.layers.42.self_attn.rotary_emb.inv_freq": "model-00043-of-00053.safetensors", + "model.layers.42.mlp.gate_proj.weight": "model-00043-of-00053.safetensors", + "model.layers.42.mlp.up_proj.weight": "model-00043-of-00053.safetensors", + "model.layers.42.mlp.down_proj.weight": "model-00043-of-00053.safetensors", + "model.layers.43.input_layernorm.weight": "model-00044-of-00053.safetensors", + "model.layers.43.post_attention_layernorm.weight": "model-00044-of-00053.safetensors", + "model.layers.43.self_attn.q_proj.weight": "model-00044-of-00053.safetensors", + "model.layers.43.self_attn.k_proj.weight": "model-00044-of-00053.safetensors", + "model.layers.43.self_attn.v_proj.weight": "model-00044-of-00053.safetensors", + "model.layers.43.self_attn.o_proj.weight": "model-00044-of-00053.safetensors", + "model.layers.43.self_attn.rotary_emb.inv_freq": "model-00044-of-00053.safetensors", + "model.layers.43.mlp.gate_proj.weight": "model-00044-of-00053.safetensors", + "model.layers.43.mlp.up_proj.weight": "model-00044-of-00053.safetensors", + "model.layers.43.mlp.down_proj.weight": "model-00044-of-00053.safetensors", + "model.layers.44.input_layernorm.weight": "model-00045-of-00053.safetensors", + "model.layers.44.post_attention_layernorm.weight": "model-00045-of-00053.safetensors", + "model.layers.44.self_attn.q_proj.weight": "model-00045-of-00053.safetensors", + "model.layers.44.self_attn.k_proj.weight": "model-00045-of-00053.safetensors", + "model.layers.44.self_attn.v_proj.weight": "model-00045-of-00053.safetensors", + "model.layers.44.self_attn.o_proj.weight": "model-00045-of-00053.safetensors", + "model.layers.44.self_attn.rotary_emb.inv_freq": "model-00045-of-00053.safetensors", + "model.layers.44.mlp.gate_proj.weight": "model-00045-of-00053.safetensors", + "model.layers.44.mlp.up_proj.weight": "model-00045-of-00053.safetensors", + "model.layers.44.mlp.down_proj.weight": "model-00045-of-00053.safetensors", + "model.layers.45.input_layernorm.weight": "model-00046-of-00053.safetensors", + "model.layers.45.post_attention_layernorm.weight": "model-00046-of-00053.safetensors", + "model.layers.45.self_attn.q_proj.weight": "model-00046-of-00053.safetensors", + "model.layers.45.self_attn.k_proj.weight": "model-00046-of-00053.safetensors", + "model.layers.45.self_attn.v_proj.weight": "model-00046-of-00053.safetensors", + "model.layers.45.self_attn.o_proj.weight": "model-00046-of-00053.safetensors", + "model.layers.45.self_attn.rotary_emb.inv_freq": "model-00046-of-00053.safetensors", + "model.layers.45.mlp.gate_proj.weight": "model-00046-of-00053.safetensors", + "model.layers.45.mlp.up_proj.weight": "model-00046-of-00053.safetensors", + "model.layers.45.mlp.down_proj.weight": "model-00046-of-00053.safetensors", + "model.layers.46.input_layernorm.weight": "model-00047-of-00053.safetensors", + "model.layers.46.post_attention_layernorm.weight": "model-00047-of-00053.safetensors", + "model.layers.46.self_attn.q_proj.weight": "model-00047-of-00053.safetensors", + "model.layers.46.self_attn.k_proj.weight": "model-00047-of-00053.safetensors", + "model.layers.46.self_attn.v_proj.weight": "model-00047-of-00053.safetensors", + "model.layers.46.self_attn.o_proj.weight": "model-00047-of-00053.safetensors", + "model.layers.46.self_attn.rotary_emb.inv_freq": "model-00047-of-00053.safetensors", + "model.layers.46.mlp.gate_proj.weight": "model-00047-of-00053.safetensors", + "model.layers.46.mlp.up_proj.weight": "model-00047-of-00053.safetensors", + "model.layers.46.mlp.down_proj.weight": "model-00047-of-00053.safetensors", + "model.layers.47.input_layernorm.weight": "model-00048-of-00053.safetensors", + "model.layers.47.post_attention_layernorm.weight": "model-00048-of-00053.safetensors", + "model.layers.47.self_attn.q_proj.weight": "model-00048-of-00053.safetensors", + "model.layers.47.self_attn.k_proj.weight": "model-00048-of-00053.safetensors", + "model.layers.47.self_attn.v_proj.weight": "model-00048-of-00053.safetensors", + "model.layers.47.self_attn.o_proj.weight": "model-00048-of-00053.safetensors", + "model.layers.47.self_attn.rotary_emb.inv_freq": "model-00048-of-00053.safetensors", + "model.layers.47.mlp.gate_proj.weight": "model-00048-of-00053.safetensors", + "model.layers.47.mlp.up_proj.weight": "model-00048-of-00053.safetensors", + "model.layers.47.mlp.down_proj.weight": "model-00048-of-00053.safetensors", + "model.layers.48.input_layernorm.weight": "model-00049-of-00053.safetensors", + "model.layers.48.post_attention_layernorm.weight": "model-00049-of-00053.safetensors", + "model.layers.48.self_attn.q_proj.weight": "model-00049-of-00053.safetensors", + "model.layers.48.self_attn.k_proj.weight": "model-00049-of-00053.safetensors", + "model.layers.48.self_attn.v_proj.weight": "model-00049-of-00053.safetensors", + "model.layers.48.self_attn.o_proj.weight": "model-00049-of-00053.safetensors", + "model.layers.48.self_attn.rotary_emb.inv_freq": "model-00049-of-00053.safetensors", + "model.layers.48.mlp.gate_proj.weight": "model-00049-of-00053.safetensors", + "model.layers.48.mlp.up_proj.weight": "model-00049-of-00053.safetensors", + "model.layers.48.mlp.down_proj.weight": "model-00049-of-00053.safetensors", + "model.layers.49.input_layernorm.weight": "model-00050-of-00053.safetensors", + "model.layers.49.post_attention_layernorm.weight": "model-00050-of-00053.safetensors", + "model.layers.49.self_attn.q_proj.weight": "model-00050-of-00053.safetensors", + "model.layers.49.self_attn.k_proj.weight": "model-00050-of-00053.safetensors", + "model.layers.49.self_attn.v_proj.weight": "model-00050-of-00053.safetensors", + "model.layers.49.self_attn.o_proj.weight": "model-00050-of-00053.safetensors", + "model.layers.49.self_attn.rotary_emb.inv_freq": "model-00050-of-00053.safetensors", + "model.layers.49.mlp.gate_proj.weight": "model-00050-of-00053.safetensors", + "model.layers.49.mlp.up_proj.weight": "model-00050-of-00053.safetensors", + "model.layers.49.mlp.down_proj.weight": "model-00050-of-00053.safetensors", + "model.layers.50.input_layernorm.weight": "model-00051-of-00053.safetensors", + "model.layers.50.post_attention_layernorm.weight": "model-00051-of-00053.safetensors", + "model.layers.50.self_attn.q_proj.weight": "model-00051-of-00053.safetensors", + "model.layers.50.self_attn.k_proj.weight": "model-00051-of-00053.safetensors", + "model.layers.50.self_attn.v_proj.weight": "model-00051-of-00053.safetensors", + "model.layers.50.self_attn.o_proj.weight": "model-00051-of-00053.safetensors", + "model.layers.50.self_attn.rotary_emb.inv_freq": "model-00051-of-00053.safetensors", + "model.layers.50.mlp.gate_proj.weight": "model-00051-of-00053.safetensors", + "model.layers.50.mlp.up_proj.weight": "model-00051-of-00053.safetensors", + "model.layers.50.mlp.down_proj.weight": "model-00051-of-00053.safetensors", + "model.layers.51.input_layernorm.weight": "model-00052-of-00053.safetensors", + "model.layers.51.post_attention_layernorm.weight": "model-00052-of-00053.safetensors", + "model.layers.51.self_attn.q_proj.weight": "model-00052-of-00053.safetensors", + "model.layers.51.self_attn.k_proj.weight": "model-00052-of-00053.safetensors", + "model.layers.51.self_attn.v_proj.weight": "model-00052-of-00053.safetensors", + "model.layers.51.self_attn.o_proj.weight": "model-00052-of-00053.safetensors", + "model.layers.51.self_attn.rotary_emb.inv_freq": "model-00052-of-00053.safetensors", + "model.layers.51.mlp.gate_proj.weight": "model-00052-of-00053.safetensors", + "model.layers.51.mlp.up_proj.weight": "model-00052-of-00053.safetensors", + "model.layers.51.mlp.down_proj.weight": "model-00052-of-00053.safetensors", + "model.norm.weight": "model-00053-of-00053.safetensors", + "model.embed_tokens.weight": "model-00053-of-00053.safetensors", + "lm_head.weight": "model-00053-of-00053.safetensors" + } +} \ No newline at end of file