diff --git a/config.json b/config.json index 43ec3420667962f974c7ceb840c4524f6bb0bd27..cf4ff966ebc5ba01a988f5cb3fff5d433deeaf0f 100644 --- a/config.json +++ b/config.json @@ -1 +1 @@ -{"architectures": ["LLaMAForCausalLM"], "bos_token_id": 0, "eos_token_id": 1, "hidden_act": "silu", "hidden_size": 6656, "intermediate_size": 17920, "initializer_range": 0.02, "max_sequence_length": 2048, "model_type": "llama", "num_attention_heads": 52, "num_hidden_layers": 60, "pad_token_id": -1, "rms_norm_eps": 1e-06, "torch_dtype": "float16", "transformers_version": "4.27.0.dev0", "use_cache": true, "vocab_size": 32000} \ No newline at end of file +{"architectures": ["LlamaForCausalLM"], "bos_token_id": 1, "eos_token_id": 2, "hidden_act": "silu", "hidden_size": 6656, "intermediate_size": 17920, "initializer_range": 0.02, "max_sequence_length": 2048, "model_type": "llama", "num_attention_heads": 52, "num_hidden_layers": 60, "pad_token_id": 0, "rms_norm_eps": 1e-06, "torch_dtype": "float16", "transformers_version": "4.27.0.dev0", "use_cache": true, "vocab_size": 32000} \ No newline at end of file diff --git a/generation_config.json b/generation_config.json index 8fdfba7b10dbe7d959d961b28f9e67433c0c0e9c..dcc93f979bea8947377955cbddf1031c3ca0b083 100644 --- a/generation_config.json +++ b/generation_config.json @@ -1 +1 @@ -{"_from_model_config": true, "bos_token_id": 0, "eos_token_id": 1, "pad_token_id": -1, "transformers_version": "4.27.0.dev0"} \ No newline at end of file +{"_from_model_config": true, "bos_token_id": 1, "eos_token_id": 2, "pad_token_id": 0, "transformers_version": "4.27.0.dev0"} \ No newline at end of file diff --git a/pytorch_model-00001-of-00061.bin b/pytorch_model-00001-of-00061.bin index 0a3d4bc87f38fc17d1c32a9819e4e14f71a74602..922b128d6ae3947a22f5bdf83c17d4440344d539 100644 --- a/pytorch_model-00001-of-00061.bin +++ b/pytorch_model-00001-of-00061.bin @@ -1,3 +1,3 @@ version https://git-lfs.github.com/spec/v1 -oid sha256:e26dcd72e7967812038ef957ba58d2884678db1e5e9336d8335619ed0c4970b1 +oid sha256:84b38f219d6db60150a0d56d7cda793889d41445e7a8f68df68f45000400d3b8 size 1337620210 diff --git a/pytorch_model-00002-of-00061.bin b/pytorch_model-00002-of-00061.bin index 199161236c484273109d89f2f6fdb1016450f922..f8100bf3525f36f9003fb78c9dde0a36d0c34185 100644 --- a/pytorch_model-00002-of-00061.bin +++ b/pytorch_model-00002-of-00061.bin @@ -1,3 +1,3 @@ version https://git-lfs.github.com/spec/v1 -oid sha256:12e65f293c8da08e96183c91b31f13d1c552406abc26f8f526d5bb06d8bd2fac +oid sha256:16bef8d1ac616ff816b63b391dc40efa546ae27a8e81a1180f09916f354a8345 size 1337620210 diff --git a/pytorch_model-00003-of-00061.bin b/pytorch_model-00003-of-00061.bin index 7f4fc40ca0d7b473616309329e39460ba8c00556..09909e46868a3602f262414d0e86ea762df2497e 100644 --- a/pytorch_model-00003-of-00061.bin +++ b/pytorch_model-00003-of-00061.bin @@ -1,3 +1,3 @@ version https://git-lfs.github.com/spec/v1 -oid sha256:44a082646bea5726c1805760c8f34c1b6996996f3c50146c3a1c801a7413f25f +oid sha256:caed8568039ae4f5cd78fb65bad94009dcad5b57161f1b597d6f8911e735dac2 size 1337620210 diff --git a/pytorch_model-00004-of-00061.bin b/pytorch_model-00004-of-00061.bin index 92f6e04458199b09addf7e4ae47e0653bc44ad49..d824ca42f0b2f151c3c19429faf0957141005394 100644 --- a/pytorch_model-00004-of-00061.bin +++ b/pytorch_model-00004-of-00061.bin @@ -1,3 +1,3 @@ version https://git-lfs.github.com/spec/v1 -oid sha256:bf2540cad38b58f4a2e073eb01d84e75309af2861174142342634ca6562e7f27 +oid sha256:541c7be3fda50c6d808fed285038d1ada750b5d499cc2927842e9d14202a58da size 1337620210 diff --git a/pytorch_model-00005-of-00061.bin b/pytorch_model-00005-of-00061.bin index a17a84dede43b5e0eb7bb8414c3b2626b64c946e..1b6e1c781a6a3a2e5c1781493a2579a6e972184f 100644 --- a/pytorch_model-00005-of-00061.bin +++ b/pytorch_model-00005-of-00061.bin @@ -1,3 +1,3 @@ version https://git-lfs.github.com/spec/v1 -oid sha256:8cd18d38d5157f4c749ed1211c883c41cce4934b0f428a156b31622ab8ea2358 +oid sha256:845621deadc97cc7fafbe7beab5123c25aeb670bee14b461a91d624c24ba24f6 size 1337620210 diff --git a/pytorch_model-00006-of-00061.bin b/pytorch_model-00006-of-00061.bin index 71f664d4c0e34fcd76cec49e6d456816cf2efcf9..95b450fcf0f98d370af9849b7ff00f1bf556ae33 100644 --- a/pytorch_model-00006-of-00061.bin +++ b/pytorch_model-00006-of-00061.bin @@ -1,3 +1,3 @@ version https://git-lfs.github.com/spec/v1 -oid sha256:8e6d265ae64af0c0e186dedb4b315a7a6ad59ec5a40ee1bfd46753d574c11470 +oid sha256:fb91e736d6b3f71e8797490a9c2fb7d2ab3e9f5ff4b1db70881a4ec06bd762c2 size 1337620210 diff --git a/pytorch_model-00007-of-00061.bin b/pytorch_model-00007-of-00061.bin index 31682d5a6c13b7741eb35f0579259ae9f5393c18..344db5127d9ef5aa783fa1a373aebd8d14426408 100644 --- a/pytorch_model-00007-of-00061.bin +++ b/pytorch_model-00007-of-00061.bin @@ -1,3 +1,3 @@ version https://git-lfs.github.com/spec/v1 -oid sha256:131b3fab3c020445a72184d17a50cf1cdd77899719efeb7692625c8c4b4098d8 +oid sha256:bf43842889c368073054ca9a0780f3b72085043a3af36d8bfd5ad2dc8e95a53e size 1337620210 diff --git a/pytorch_model-00008-of-00061.bin b/pytorch_model-00008-of-00061.bin index de57fa4278dd9822011dca13fe298a7ef67d2116..a48b3786e0f45e9b3b66b2920aea0fa13ba2be7a 100644 --- a/pytorch_model-00008-of-00061.bin +++ b/pytorch_model-00008-of-00061.bin @@ -1,3 +1,3 @@ version https://git-lfs.github.com/spec/v1 -oid sha256:3d6359a4f6c102b229ada093a009547707627d1bb9af175c9b389960749c4396 +oid sha256:8ce60a011ea34d0a03dbd3a5b071582bf21d6878b40e0b6ec83828b1da385956 size 1337620210 diff --git a/pytorch_model-00009-of-00061.bin b/pytorch_model-00009-of-00061.bin index fcde3946fde84306285224caf76a304d7279ec7f..954c078c3d579dd1758d34e0beb8e63221b6f616 100644 --- a/pytorch_model-00009-of-00061.bin +++ b/pytorch_model-00009-of-00061.bin @@ -1,3 +1,3 @@ version https://git-lfs.github.com/spec/v1 -oid sha256:44ea0e16e60983bfac93f85b38bb3cfecc588b772a20e1b50fda25179dd60940 +oid sha256:2ef0d71732b372c4fb253a2f9642f811f615004baca907bb9fc788d1a0262b8a size 1337620210 diff --git a/pytorch_model-00010-of-00061.bin b/pytorch_model-00010-of-00061.bin index 9cb0a88d0eeed45fa07306fa2e0c0c204ec095b3..931de77e6c80fb03977e302a016307bb7f9e9d15 100644 --- a/pytorch_model-00010-of-00061.bin +++ b/pytorch_model-00010-of-00061.bin @@ -1,3 +1,3 @@ version https://git-lfs.github.com/spec/v1 -oid sha256:b67de5d47023ad413905a9de9cc472fd344e5081795588e8eca199c9407c97eb +oid sha256:d3d104efad6885a5dcba84e309fb865d6c46e663cbc27266760ba541b68eac9e size 1337620210 diff --git a/pytorch_model-00011-of-00061.bin b/pytorch_model-00011-of-00061.bin index 62dfea56263317bf27ebb4ee18ade35f0fe543e6..d7f32ee67c102b96cee9708bcbfbd818e5b4d924 100644 --- a/pytorch_model-00011-of-00061.bin +++ b/pytorch_model-00011-of-00061.bin @@ -1,3 +1,3 @@ version https://git-lfs.github.com/spec/v1 -oid sha256:9079204376feed571a2de1e1ae12981a683addfad5c00e4b6144a41d2694a614 +oid sha256:29aa8aedb550b36e41a1a1d5f4ffdefbe742d7897ee98da4455ad709ee7bc32f size 1337620210 diff --git a/pytorch_model-00012-of-00061.bin b/pytorch_model-00012-of-00061.bin index 43820148aedf369477a81d71841c4f2dc8cd1155..725a675649a59b0d935c1eea6e2f89127f9a3a4e 100644 --- a/pytorch_model-00012-of-00061.bin +++ b/pytorch_model-00012-of-00061.bin @@ -1,3 +1,3 @@ version https://git-lfs.github.com/spec/v1 -oid sha256:ba7b4327d1de943347101a582e7c3aaedceb8d0846d3b962990f08400066aedf +oid sha256:dd60e84e2df19666feb171a761798df42af87a64afa3adb2d8086e5ccb05e934 size 1337620210 diff --git a/pytorch_model-00013-of-00061.bin b/pytorch_model-00013-of-00061.bin index 683b6d594774b8e8faf8b21bda5308d50bab7390..18820b82c775bdae36e4a285bf8f336fcf12df88 100644 --- a/pytorch_model-00013-of-00061.bin +++ b/pytorch_model-00013-of-00061.bin @@ -1,3 +1,3 @@ version https://git-lfs.github.com/spec/v1 -oid sha256:fdd48972a3108e2b61e159311540523e13e22249552e64b8e806417cafcd24a2 +oid sha256:cb6cfb9a161f930e19b610911082567d509a8c2567afbb80a1e53c835efc2f0f size 1337620210 diff --git a/pytorch_model-00014-of-00061.bin b/pytorch_model-00014-of-00061.bin index d081031aa934a221628b6bb35b2576518d2915d7..9b59570efa94d01d6afd52dbbc35aad6f78a977e 100644 --- a/pytorch_model-00014-of-00061.bin +++ b/pytorch_model-00014-of-00061.bin @@ -1,3 +1,3 @@ version https://git-lfs.github.com/spec/v1 -oid sha256:ef8f104efce6d490d1677e7a6717d43494b41eb6be83aaa6a7454a297fc9c7b6 +oid sha256:fcc2acb1544250149d7d3ca4f5a2fbe1be240c37fd300ad01084b1db9bd1d006 size 1337620210 diff --git a/pytorch_model-00015-of-00061.bin b/pytorch_model-00015-of-00061.bin index 5765d709578d180b0a484ae750ed67f564fb3b2a..0906fcc406e05bbef8b3c5ad01d38b4982a48163 100644 --- a/pytorch_model-00015-of-00061.bin +++ b/pytorch_model-00015-of-00061.bin @@ -1,3 +1,3 @@ version https://git-lfs.github.com/spec/v1 -oid sha256:5f5961cd4d6b44eaf6540ae6d3c7160a96fdc754738c482a4f84f2ca39e6d571 +oid sha256:3ae030f6e8e99a68160d92e0e7051b159b8e70f051fa3f2eb72295eb4ea31c33 size 1337620210 diff --git a/pytorch_model-00016-of-00061.bin b/pytorch_model-00016-of-00061.bin index baceec24e30e3d69c1de7ecdfb5492671fce9170..67ea4259909f090de985fcb06e4cc3a142c9f53e 100644 --- a/pytorch_model-00016-of-00061.bin +++ b/pytorch_model-00016-of-00061.bin @@ -1,3 +1,3 @@ version https://git-lfs.github.com/spec/v1 -oid sha256:dd4c065c013ea5857153d09aa99df93b4fe63e69e6b7d2a3132d7c3a5041a35a +oid sha256:a6f6fab6b362aab59a7b054e6e4d5222a65ee7b039f398dfb71c8ae9b8fc9f04 size 1337620210 diff --git a/pytorch_model-00017-of-00061.bin b/pytorch_model-00017-of-00061.bin index 3b8359d95494040b80ecd1635a3fc3970f3eaa8b..31f157e41eddecd768171a0038adbf0af27cefcb 100644 --- a/pytorch_model-00017-of-00061.bin +++ b/pytorch_model-00017-of-00061.bin @@ -1,3 +1,3 @@ version https://git-lfs.github.com/spec/v1 -oid sha256:c2473a2bb09eabb5236aa7f831df684e977d9da3134b44a0c3f2a49f72ca8fa2 +oid sha256:f241f2aa12befd8e40fbd4542d5aa3ee0cb683f37d3bd0842a173d815656a721 size 1337620210 diff --git a/pytorch_model-00018-of-00061.bin b/pytorch_model-00018-of-00061.bin index 8216f6bfbb7d50d8a3136c00003f79603ac5cec7..9a8b267078cbce95c7854a8bca6378abd14a0a4f 100644 --- a/pytorch_model-00018-of-00061.bin +++ b/pytorch_model-00018-of-00061.bin @@ -1,3 +1,3 @@ version https://git-lfs.github.com/spec/v1 -oid sha256:ef02b9cfc02fdc3cfbc812a1d49a9221a9fde9032e84b91d21faf7415c1acbd9 +oid sha256:b268b4a2ecce321bc91cc56dd5930bd44e5e352ba1c774771dce213bacbcb252 size 1337620210 diff --git a/pytorch_model-00019-of-00061.bin b/pytorch_model-00019-of-00061.bin index 0d9cd4a772a6b7487530b22893a5d5e2cdd1f899..c6ec93fecb36e33bbd024bd763b3d221d5c55d9e 100644 --- a/pytorch_model-00019-of-00061.bin +++ b/pytorch_model-00019-of-00061.bin @@ -1,3 +1,3 @@ version https://git-lfs.github.com/spec/v1 -oid sha256:31b1d06ccb074168f0ccf7cb2ebc6078d1cdae6a1c6ab97ad5d97f7600fde0d0 +oid sha256:5ecbee54e0f47a68e5c6c94298df06803d5ac8c94960174c322cd4892b6ee3f7 size 1337620210 diff --git a/pytorch_model-00020-of-00061.bin b/pytorch_model-00020-of-00061.bin index 0fde24d6526141c883d696300ce1b9009128a951..7bc5787eca6853da462b52071d71938be7cdb5c2 100644 --- a/pytorch_model-00020-of-00061.bin +++ b/pytorch_model-00020-of-00061.bin @@ -1,3 +1,3 @@ version https://git-lfs.github.com/spec/v1 -oid sha256:45d800d8dee1986a34153c396b2d5cd501affa4ac2986082028b5f619058ef52 +oid sha256:6dddd0872788fac1461687b33a266b17e6fde7757e9c13bd3b37f46b3002acbc size 1337620210 diff --git a/pytorch_model-00021-of-00061.bin b/pytorch_model-00021-of-00061.bin index 2d3e970406db0aa4cb9591d3748d0ddeacc7d65e..29c4a0aa2894595ae4ceacb29a1ed5aeacacf814 100644 --- a/pytorch_model-00021-of-00061.bin +++ b/pytorch_model-00021-of-00061.bin @@ -1,3 +1,3 @@ version https://git-lfs.github.com/spec/v1 -oid sha256:d9721f07a51ee49279428edac3c6644a1c6bde44d713a4d0175e7733dbdc9c0c +oid sha256:339dcb85e12a2209fe7ff807eb4f20cc9982babf63e36b44eb9ff655ed2504d8 size 1337620210 diff --git a/pytorch_model-00022-of-00061.bin b/pytorch_model-00022-of-00061.bin index 899705fb6228128db8e6a7897a03a6a853911a23..bc098f28c37a5ddf8c5f553f0fb8fd943927c113 100644 --- a/pytorch_model-00022-of-00061.bin +++ b/pytorch_model-00022-of-00061.bin @@ -1,3 +1,3 @@ version https://git-lfs.github.com/spec/v1 -oid sha256:77c93f28cc7c39e908e3f692d4ad4238d71de9cb82be51d9f200bf3c873747dd +oid sha256:cea8b8c9dab69f0f27aae835aaaf4b3560310c62e5bd4261f43785535a703fb6 size 1337620210 diff --git a/pytorch_model-00023-of-00061.bin b/pytorch_model-00023-of-00061.bin index 0b3b2590d26f0657737477bb2db3ad06d6c7df9e..9d28b069d78020b2468fb86efd4b2b03875297f4 100644 --- a/pytorch_model-00023-of-00061.bin +++ b/pytorch_model-00023-of-00061.bin @@ -1,3 +1,3 @@ version https://git-lfs.github.com/spec/v1 -oid sha256:d46f1dee410916317d907d54e81771654bbab78aff17abbcde256ddb0c258752 +oid sha256:0eee2c1933d6cab3c9bbd4abbdb7d4d091584809233020a8039752b49ad331a9 size 1337620210 diff --git a/pytorch_model-00024-of-00061.bin b/pytorch_model-00024-of-00061.bin index 402dc8f86d7fb4171d96f990a25032c08aa6810c..5be94adbf85d3e3bb565be6dd38863215eb667fb 100644 --- a/pytorch_model-00024-of-00061.bin +++ b/pytorch_model-00024-of-00061.bin @@ -1,3 +1,3 @@ version https://git-lfs.github.com/spec/v1 -oid sha256:5ff77cde942aeab61126a8a0c8c36dac8e8a93e9240a5508e4ba2c69bf749b31 +oid sha256:71f37f697b8d3154bad279d37c3ed27b6c8adb1f8275ae14407af213f8ad8986 size 1337620210 diff --git a/pytorch_model-00025-of-00061.bin b/pytorch_model-00025-of-00061.bin index fcdf4f874102d8d0b6211de626d78e8a652205b5..9d84b607efc247a69cfede2b3becb87daf29115e 100644 --- a/pytorch_model-00025-of-00061.bin +++ b/pytorch_model-00025-of-00061.bin @@ -1,3 +1,3 @@ version https://git-lfs.github.com/spec/v1 -oid sha256:6fa48f916cd163028082776dbeebd9bc9173e3adc9116f913df0a2b174c35c4c +oid sha256:868850b27eea9acca89f0963e064e4e09c8d289fd24f1ed4dfc1eea01c9e7e9d size 1337620210 diff --git a/pytorch_model-00026-of-00061.bin b/pytorch_model-00026-of-00061.bin index 4fc900762a18377a4fe0023a205b15b7c2b3b1a9..8b130b669de290602c798d3a63b8925942e26ba2 100644 --- a/pytorch_model-00026-of-00061.bin +++ b/pytorch_model-00026-of-00061.bin @@ -1,3 +1,3 @@ version https://git-lfs.github.com/spec/v1 -oid sha256:5dcda4e8049bb3dcd344dcdf8e669f276606a2aec9528a4f1454f9a2409aaa54 +oid sha256:b07e076e2085906d3a6547312de3312d2776fa3138c6b542a8fbf62264445a75 size 1337620210 diff --git a/pytorch_model-00027-of-00061.bin b/pytorch_model-00027-of-00061.bin index f634e9eb14c29be99d95cfbec2432bc0fcafcdea..04a1b46bc1e7b0ea4a0be8dd71ffb09be0403c5b 100644 --- a/pytorch_model-00027-of-00061.bin +++ b/pytorch_model-00027-of-00061.bin @@ -1,3 +1,3 @@ version https://git-lfs.github.com/spec/v1 -oid sha256:1cdeacbb55277d3c3c5b213eb5ce3509eca91cef0cb22bf23b6e22bc7b867e30 +oid sha256:635b25c13e861e7c3d1d7eeb9aecbecd7efd80c6e1865cb1966d1dfb426a3818 size 1337620210 diff --git a/pytorch_model-00028-of-00061.bin b/pytorch_model-00028-of-00061.bin index ae14ad7dddd57227298353bf6201743f64eff472..f5011ca1a5aea6151d5dd52cc06adc6df9772a7c 100644 --- a/pytorch_model-00028-of-00061.bin +++ b/pytorch_model-00028-of-00061.bin @@ -1,3 +1,3 @@ version https://git-lfs.github.com/spec/v1 -oid sha256:507a0fa36bbeac4c5fec340cb000139d76c01ada14f04de004027aa8d7bd4ac3 +oid sha256:96778d7209585bf63a2b143f02e53a1e0011d684581ba23689c212e8a0757c8b size 1337620210 diff --git a/pytorch_model-00029-of-00061.bin b/pytorch_model-00029-of-00061.bin index 41451ff59f0ba1af62440a9910d7858d9112d29f..ea724bca4141146dfaf27906252ad111fe494dde 100644 --- a/pytorch_model-00029-of-00061.bin +++ b/pytorch_model-00029-of-00061.bin @@ -1,3 +1,3 @@ version https://git-lfs.github.com/spec/v1 -oid sha256:62746a327ba66f3f887c848be0644aca6bae43f54921fc4cf743ab9bfd536725 +oid sha256:4930c8a432935e442b8072ea8be1fae020a5599b0b3b6e9ca7d0abc006119d02 size 1337620210 diff --git a/pytorch_model-00030-of-00061.bin b/pytorch_model-00030-of-00061.bin index 9bf7e625ff1095231ab06b193a7d9f488fbf1024..9269d296934feb396f3826910f0762f7322ebe1c 100644 --- a/pytorch_model-00030-of-00061.bin +++ b/pytorch_model-00030-of-00061.bin @@ -1,3 +1,3 @@ version https://git-lfs.github.com/spec/v1 -oid sha256:8842a1d7c11391261f053053b37cdcd848bf84d7db77835e1d192ffcd6e2099b +oid sha256:d03949ae9504fd3ad9d7a587a933ba04a4f9b14014a9ad0ba969e36bb7e7ae88 size 1337620210 diff --git a/pytorch_model-00031-of-00061.bin b/pytorch_model-00031-of-00061.bin index 510c518948614f33faeb88267001aae9eb106edd..84a27972b9521a321970e928e95cc3e9a27e6146 100644 --- a/pytorch_model-00031-of-00061.bin +++ b/pytorch_model-00031-of-00061.bin @@ -1,3 +1,3 @@ version https://git-lfs.github.com/spec/v1 -oid sha256:4f0594d8cf0bb8821a92e6bc2967545183a68f718d62b3553ca83a1164c98355 +oid sha256:eda51342afe38df77cea10514090dafbc3bd562b058336973071d583236f9441 size 1337620210 diff --git a/pytorch_model-00032-of-00061.bin b/pytorch_model-00032-of-00061.bin index 8e068d382204fc50683be829c91de8c7fad6fcc5..668137de8a9b2309c4e9749ee9b23461510ea16d 100644 --- a/pytorch_model-00032-of-00061.bin +++ b/pytorch_model-00032-of-00061.bin @@ -1,3 +1,3 @@ version https://git-lfs.github.com/spec/v1 -oid sha256:91b09a16a9b5245d518c396ae745552557359f698690b963361078c9a342eaeb +oid sha256:697a920414740c152309cdfaa0dbeb33c85575aa7716528b3c24ba3be4c36253 size 1337620210 diff --git a/pytorch_model-00033-of-00061.bin b/pytorch_model-00033-of-00061.bin index e821e5d6b5bf3a523d55ce9b1133b29f0f3f83e4..4960cdcb750f055a8b1a5bb6d21e53913b2a10b6 100644 --- a/pytorch_model-00033-of-00061.bin +++ b/pytorch_model-00033-of-00061.bin @@ -1,3 +1,3 @@ version https://git-lfs.github.com/spec/v1 -oid sha256:3349ea7eb1e54ca2d603e7fa194334d53bad6f579fbbe3565a063a4b81f2f281 +oid sha256:0ee849ef7c4343288a520991507a396401e186f23fd2ab7e9605d03406220d5a size 1337620210 diff --git a/pytorch_model-00034-of-00061.bin b/pytorch_model-00034-of-00061.bin index f452bbb4399b825bec53a841f3238351b024585e..63f587e63292661a0bb34f55abedd706bf470732 100644 --- a/pytorch_model-00034-of-00061.bin +++ b/pytorch_model-00034-of-00061.bin @@ -1,3 +1,3 @@ version https://git-lfs.github.com/spec/v1 -oid sha256:71695b7c46939277b452eb36a50a95ad91d70395cb97d7db36618dc74074ac88 +oid sha256:096c9e4a203f059a4639199516a98ce7c92310b2084b8bff4f6be26aa84135ad size 1337620210 diff --git a/pytorch_model-00035-of-00061.bin b/pytorch_model-00035-of-00061.bin index 1505686e293760b3ef1fdbafa886f147ae9514c0..e25916d43781a5dd8a48eaee536dd614e22ef4af 100644 --- a/pytorch_model-00035-of-00061.bin +++ b/pytorch_model-00035-of-00061.bin @@ -1,3 +1,3 @@ version https://git-lfs.github.com/spec/v1 -oid sha256:aacf8dcf079bb9430178cc2c87ded6af05263063b0e1cb51ceb4e0000557a071 +oid sha256:57d76c4ae0f20e11809cf46d2c46947265e03fd086612a20b4b7271fe3a149bc size 1337620210 diff --git a/pytorch_model-00036-of-00061.bin b/pytorch_model-00036-of-00061.bin index 382b07861a3b7ff6f5fc31178fb1c160c193e9fb..258a64f3920dd5c7a07f0255ab930d8acccba913 100644 --- a/pytorch_model-00036-of-00061.bin +++ b/pytorch_model-00036-of-00061.bin @@ -1,3 +1,3 @@ version https://git-lfs.github.com/spec/v1 -oid sha256:10c85f6a7d207624073faa4e68d47a88750b41cd8170ada2eccfc85b9ed37cd1 +oid sha256:feea3c809f8ec9ac68e1a832276d01ccfdf3510d698d77b1c98ffc0536d904ab size 1337620210 diff --git a/pytorch_model-00037-of-00061.bin b/pytorch_model-00037-of-00061.bin index fb4eee72e270acabd604abc7edb0ad591ef6de28..63a963def47061d22ea3ce02aca0d5b2c31e1162 100644 --- a/pytorch_model-00037-of-00061.bin +++ b/pytorch_model-00037-of-00061.bin @@ -1,3 +1,3 @@ version https://git-lfs.github.com/spec/v1 -oid sha256:6de0d9974b26e983567b0387d7b87f76f67edf1fb2c0ea03fcb1e2f4e5dc658f +oid sha256:587737533dd5f02b98a04ee982d3a72512e6786332f153aeff249e3a9eef2255 size 1337620210 diff --git a/pytorch_model-00038-of-00061.bin b/pytorch_model-00038-of-00061.bin index 1054e11e6e64ca18d894ab2db333b324e09cd279..ce22472e0ddc5a7cbb4b1a50d0e143706d0cfc26 100644 --- a/pytorch_model-00038-of-00061.bin +++ b/pytorch_model-00038-of-00061.bin @@ -1,3 +1,3 @@ version https://git-lfs.github.com/spec/v1 -oid sha256:79406f3923a8cdd9e79f7fbffde3549c701cdda665367a4e846fefab7652bb61 +oid sha256:5974033c38e43a11184521ea5f1e07db2080330926ab40a2e00bb9f640fc1f99 size 1337620210 diff --git a/pytorch_model-00039-of-00061.bin b/pytorch_model-00039-of-00061.bin index 1ed8322bb5dbfcfef66e965e19bbaf56136ea0bb..89356656acb6cd223847b0bac76e36864c6ec3fb 100644 --- a/pytorch_model-00039-of-00061.bin +++ b/pytorch_model-00039-of-00061.bin @@ -1,3 +1,3 @@ version https://git-lfs.github.com/spec/v1 -oid sha256:b8484077be7a361eff286ec89856984c7818e93eb776154eb939d7c551167599 +oid sha256:a69f378ed418597a2d70198c3b15f6851eda4d7a8eb1c8073bb04a62f115ac18 size 1337620210 diff --git a/pytorch_model-00040-of-00061.bin b/pytorch_model-00040-of-00061.bin index 4d29f125d121e379c12b509563ff94038b90c558..c0fdb2f69f8282644ffab31a6736ebd23be3c736 100644 --- a/pytorch_model-00040-of-00061.bin +++ b/pytorch_model-00040-of-00061.bin @@ -1,3 +1,3 @@ version https://git-lfs.github.com/spec/v1 -oid sha256:7746057dfea547498f902c49d9da4fcbe310578d038739a1d29cc8097a704913 +oid sha256:0144f41b1c85f06f75f1d3bd1c854047c6494e532b4f0b63c260c71b9688f952 size 1337620210 diff --git a/pytorch_model-00041-of-00061.bin b/pytorch_model-00041-of-00061.bin index 784402d977eeea83cafd8eeab174583dcd004a25..c56d8e8ad52fcb16d616a34833c2100eee26f4ee 100644 --- a/pytorch_model-00041-of-00061.bin +++ b/pytorch_model-00041-of-00061.bin @@ -1,3 +1,3 @@ version https://git-lfs.github.com/spec/v1 -oid sha256:a7b982f94acfae8fb32812aa211fd28f140528c3a12621b090a8c4edc03d8b81 +oid sha256:3e517621ef87cce8f756ca07a92ebca0c867cecaa64ee4fb75dd1a736ab77b92 size 1337620210 diff --git a/pytorch_model-00042-of-00061.bin b/pytorch_model-00042-of-00061.bin index a53149bf9c939b90995b3f5941df83ff4a957c7e..679142ff3c4af77622a9c9dce69e0e1ddcaa6507 100644 --- a/pytorch_model-00042-of-00061.bin +++ b/pytorch_model-00042-of-00061.bin @@ -1,3 +1,3 @@ version https://git-lfs.github.com/spec/v1 -oid sha256:394b8614fcb115753e4fbdc768029c3686628d03d8a0fe40d7ca491d486bb48a +oid sha256:848d0e4e88466fbe86b86160f9f14d79a000620f3aa760303a90135038729012 size 1337620210 diff --git a/pytorch_model-00043-of-00061.bin b/pytorch_model-00043-of-00061.bin index 95e9c089076fcaac00d472c85e4092ca512b7578..c702e2794a7ca73a679b28fbec540e1b0492f5b4 100644 --- a/pytorch_model-00043-of-00061.bin +++ b/pytorch_model-00043-of-00061.bin @@ -1,3 +1,3 @@ version https://git-lfs.github.com/spec/v1 -oid sha256:9f75fc52f1f482f30979119eb91a5ed8f15bb2a8b593987545f09a4b26ec2720 +oid sha256:60bd7630ce3818d30ea7de2a411381679cb37fa21d5a68ace5035b8dda62e209 size 1337620210 diff --git a/pytorch_model-00044-of-00061.bin b/pytorch_model-00044-of-00061.bin index aaf97b4bc522740a403fd5ec938f014678434856..9cfbe4355ad0303b303c99e83d8308dc20a860c2 100644 --- a/pytorch_model-00044-of-00061.bin +++ b/pytorch_model-00044-of-00061.bin @@ -1,3 +1,3 @@ version https://git-lfs.github.com/spec/v1 -oid sha256:8a1469cd8392417220d312e4b0c3d6e6c9afd976425a1cf125997ae2ec744c35 +oid sha256:66967d069ccdb9637e3fa5ef66e13dff9c39267cd386150c95a01206555c7527 size 1337620210 diff --git a/pytorch_model-00045-of-00061.bin b/pytorch_model-00045-of-00061.bin index 214eb99a6f286c28d529768cf519df84ec724009..0567ce4ceb97422263dcf96775ead368c992b396 100644 --- a/pytorch_model-00045-of-00061.bin +++ b/pytorch_model-00045-of-00061.bin @@ -1,3 +1,3 @@ version https://git-lfs.github.com/spec/v1 -oid sha256:51c62a84277bdc35d1ab4b5b65a87e1f204ff4103413bfb3525cf2c3ada59ccd +oid sha256:66a7196e2b5b04c400ce3897c74d9a865b4271f4c29bebef2b95b5863b23cf9a size 1337620210 diff --git a/pytorch_model-00046-of-00061.bin b/pytorch_model-00046-of-00061.bin index 80cd16d567c791ef34792aa376430d5faaf1d0c2..d64ec087cb5747d4338de9c6cc165dc81dd52e1e 100644 --- a/pytorch_model-00046-of-00061.bin +++ b/pytorch_model-00046-of-00061.bin @@ -1,3 +1,3 @@ version https://git-lfs.github.com/spec/v1 -oid sha256:6ba8b192bd6a3322cb71a4c04b7c6a76f161ae94b2dbf27b9788d5c7753751cc +oid sha256:e4ad771e247b7976a4e87f7c32dcc3ee78534b61c1ac750945beff5ced03696c size 1337620210 diff --git a/pytorch_model-00047-of-00061.bin b/pytorch_model-00047-of-00061.bin index 0df29cef65870934f295d61a463ae575bd93cfaa..74819280aecf40caa35c7e520ec5d369e2a5a0e9 100644 --- a/pytorch_model-00047-of-00061.bin +++ b/pytorch_model-00047-of-00061.bin @@ -1,3 +1,3 @@ version https://git-lfs.github.com/spec/v1 -oid sha256:cc9603f655e80ccb5d9393c7e43936a1bce9e38438876729e1cd3c67b001e6fd +oid sha256:65745063e80d130e8d5bbf171ee83377b052cdd2f18bfb7862c1ceb865047eb1 size 1337620210 diff --git a/pytorch_model-00048-of-00061.bin b/pytorch_model-00048-of-00061.bin index ebe44978ec8afde6f70b598c03cc38260a05d5c9..f55cbb6da5a406ebd62817f26a41f7832bc50861 100644 --- a/pytorch_model-00048-of-00061.bin +++ b/pytorch_model-00048-of-00061.bin @@ -1,3 +1,3 @@ version https://git-lfs.github.com/spec/v1 -oid sha256:353d1a8c0f8d39f55637ea22e20d6ef429bf8c0fba87be79b572a54183348e29 +oid sha256:ea1e484a09dd5125005a617750f61d9d7049c1765b58e606417ae778923db796 size 1337620210 diff --git a/pytorch_model-00049-of-00061.bin b/pytorch_model-00049-of-00061.bin index 83a1d872cea6cf59590dba6a73b57d2d064fb8d9..a1e4fed2cb2a593d822c7a0635ce375bd76c80de 100644 --- a/pytorch_model-00049-of-00061.bin +++ b/pytorch_model-00049-of-00061.bin @@ -1,3 +1,3 @@ version https://git-lfs.github.com/spec/v1 -oid sha256:e4994a4bc02d5300d58da999dc6cb12d4d1c8e5c159f3f5ba92349940c2f969e +oid sha256:a39ff509c1d1a2e7df5c63fdfc0e6b7fa9997c490d2f73114130b054aa950c8f size 1337620210 diff --git a/pytorch_model-00050-of-00061.bin b/pytorch_model-00050-of-00061.bin index b7ceeb6382925ed1e1282fd6a657486d9bc40b1c..48f46404edacbee834c0cb2dae551508a36a0fc6 100644 --- a/pytorch_model-00050-of-00061.bin +++ b/pytorch_model-00050-of-00061.bin @@ -1,3 +1,3 @@ version https://git-lfs.github.com/spec/v1 -oid sha256:0b3cb3a3cb1d1e0800803ab3e6f214e2214ff4c8d02e7140fae5a57a4f9be8fc +oid sha256:780abb43773a015d9040628ffaee36d4c9de72bac0eeb034be9e45fea78cab6b size 1337620210 diff --git a/pytorch_model-00051-of-00061.bin b/pytorch_model-00051-of-00061.bin index 1dce3e2851200c415124a3d0cd0e62a25992cef5..bb2f4417888d5f0da46ec5644a087b8457ff0bbe 100644 --- a/pytorch_model-00051-of-00061.bin +++ b/pytorch_model-00051-of-00061.bin @@ -1,3 +1,3 @@ version https://git-lfs.github.com/spec/v1 -oid sha256:79a71fcb296bd0dfa123bad6195f3f48e47a3f47d327367aaa6df8b19e82573b +oid sha256:25fed9e0cfb76407b29fd4017328f8f10497c1f377d5743497f3c8fbb29d351b size 1337620210 diff --git a/pytorch_model-00052-of-00061.bin b/pytorch_model-00052-of-00061.bin index 7c4b29cd04e960f955b3707d1d3a5a1f066d6fdc..2568f986dcc731975f64ccbcf7dfc910649b971a 100644 --- a/pytorch_model-00052-of-00061.bin +++ b/pytorch_model-00052-of-00061.bin @@ -1,3 +1,3 @@ version https://git-lfs.github.com/spec/v1 -oid sha256:4ed64fd83c7c389452bfa36c3f23b6da450b22cbdf9cb63b31da7454786ef158 +oid sha256:c09b7829ac208037e220eabfa7058417bb1e03224b4de0b754d000b159a25385 size 1337620210 diff --git a/pytorch_model-00053-of-00061.bin b/pytorch_model-00053-of-00061.bin index 82353375964be7aeae4fec4959196a87223ecd61..0a6c2a9c39878617b48f505b44df6e00d921c034 100644 --- a/pytorch_model-00053-of-00061.bin +++ b/pytorch_model-00053-of-00061.bin @@ -1,3 +1,3 @@ version https://git-lfs.github.com/spec/v1 -oid sha256:2dc4b785351dd33416d7741f1b4b33fc8839578d68f40dbaea86557a150d8f94 +oid sha256:2badbc153e3a21a538cf8a90a82aca1690739aae6368b0b417eff6c1d1d726b5 size 1337620210 diff --git a/pytorch_model-00054-of-00061.bin b/pytorch_model-00054-of-00061.bin index b79369b350f789ec818fdce711e075fabb0cf617..c5ceff2e052e9024c8de7162325b96bf8a664752 100644 --- a/pytorch_model-00054-of-00061.bin +++ b/pytorch_model-00054-of-00061.bin @@ -1,3 +1,3 @@ version https://git-lfs.github.com/spec/v1 -oid sha256:6b3ff736ce6eb4c328c5d01a05b7fc2996000b68911bf11c43780c2bbce5e376 +oid sha256:eb484625b10fa1618c13ed1daea996bfb6a73645d74e06615601cea7591db9a1 size 1337620210 diff --git a/pytorch_model-00055-of-00061.bin b/pytorch_model-00055-of-00061.bin index 8244eb1ce49c451eff90a42fb8d3add44dadf500..f32e36a07c4fda7a75a5f4796725be1e42deff99 100644 --- a/pytorch_model-00055-of-00061.bin +++ b/pytorch_model-00055-of-00061.bin @@ -1,3 +1,3 @@ version https://git-lfs.github.com/spec/v1 -oid sha256:ff70a42c747bbcec0ce3d0a811181058a398d989504087b10ac5c7118538dd4d +oid sha256:88e2988c41dacf447879055c2eb9c5cf2f3669345bcecb88aaf66a6d67b166af size 1337620210 diff --git a/pytorch_model-00056-of-00061.bin b/pytorch_model-00056-of-00061.bin index 06e4502462a50d25cc7a6218e72b20cf5b31497f..5ae4c262d1377d5e08ee4ac3d4a70961d20d144e 100644 --- a/pytorch_model-00056-of-00061.bin +++ b/pytorch_model-00056-of-00061.bin @@ -1,3 +1,3 @@ version https://git-lfs.github.com/spec/v1 -oid sha256:8f30751c0a02123c9527f5136ecb9330d3ab791669705f7a4a176a58e6c30a21 +oid sha256:58e8ca0b20b1c01596866f5818fe459255e0eb7492c232a74f189997590b94e6 size 1337620210 diff --git a/pytorch_model-00057-of-00061.bin b/pytorch_model-00057-of-00061.bin index 5594a7298d7b57b124ac36c930f525d80e8de221..80891fa6beaf42cda041872153d94619344f9b12 100644 --- a/pytorch_model-00057-of-00061.bin +++ b/pytorch_model-00057-of-00061.bin @@ -1,3 +1,3 @@ version https://git-lfs.github.com/spec/v1 -oid sha256:afd7ba782e05564ed215656f1bfe9d72e27e01379740ee312cc1736670131437 +oid sha256:3ecbf8434d130fe6f1f8f5f77800793ba7b810423dabc1a8bc976d68b425cce7 size 1337620210 diff --git a/pytorch_model-00058-of-00061.bin b/pytorch_model-00058-of-00061.bin index 1c5e37d6da8eac67d8b5c66394b10eb9ca1112a1..268c0d83a39aceefcb4260646ab34421080e4e8b 100644 --- a/pytorch_model-00058-of-00061.bin +++ b/pytorch_model-00058-of-00061.bin @@ -1,3 +1,3 @@ version https://git-lfs.github.com/spec/v1 -oid sha256:7f353a577561e159f98f60d12348e666ebd0eab27ed10c5804c43ac01966ca90 +oid sha256:ff76d3b8b19127fe2f7ce8ee341acf3329637d9fde27f4797df47cb8c326267f size 1337620210 diff --git a/pytorch_model-00059-of-00061.bin b/pytorch_model-00059-of-00061.bin index 67a621deb744c2e639184f7743f6d3a5a9c1cd7e..3d3684d25d0a1e6b01ecdd77af34b8b01fc0c5a5 100644 --- a/pytorch_model-00059-of-00061.bin +++ b/pytorch_model-00059-of-00061.bin @@ -1,3 +1,3 @@ version https://git-lfs.github.com/spec/v1 -oid sha256:4778f514afa332126841cab26fdfe7b352f5bec0f929cbcdbea962ee1d21faa8 +oid sha256:c417249d5f00a480869f691759d119483e2b2a484c93f224b5029745de28d3ad size 1337620210 diff --git a/pytorch_model-00060-of-00061.bin b/pytorch_model-00060-of-00061.bin index abbe0decbb0cbbdfe018af75df3f8ade3b0aeb1a..09585720f39734a3844c9d8b1e1ace0276e0a999 100644 --- a/pytorch_model-00060-of-00061.bin +++ b/pytorch_model-00060-of-00061.bin @@ -1,3 +1,3 @@ version https://git-lfs.github.com/spec/v1 -oid sha256:7742a45080931322a60806b8064eaef97e8b526442151e7dfff433425c5e8f14 -size 1064974796 +oid sha256:5f4d92b1b01e14b94050cbe9362edcd3b0f25c829d01ce15e3e2852079053030 +size 1337620210 diff --git a/pytorch_model-00061-of-00061.bin b/pytorch_model-00061-of-00061.bin new file mode 100644 index 0000000000000000000000000000000000000000..23f5729380f822db3d4d8f5135c39f2018ef17d3 --- /dev/null +++ b/pytorch_model-00061-of-00061.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:02b1177e6cc569c777e076092f1d6d143d7f843ed540019d867628f342e06e55 +size 1064974796 diff --git a/pytorch_model.bin.index.json b/pytorch_model.bin.index.json index 1cf70c46d585389bc0d525fb4de81d5cb84f011c..8ac59a0844c29745e9702bfa8cac73a6d8988ff5 100644 --- a/pytorch_model.bin.index.json +++ b/pytorch_model.bin.index.json @@ -1 +1 @@ -{"weight_map": {"model.layers.0.input_layernorm.weight": "pytorch_model-00000-of-00061.bin", "model.layers.0.post_attention_layernorm.weight": "pytorch_model-00000-of-00061.bin", "model.layers.0.self_attn.q_proj.weight": "pytorch_model-00000-of-00061.bin", "model.layers.0.self_attn.k_proj.weight": "pytorch_model-00000-of-00061.bin", "model.layers.0.self_attn.v_proj.weight": "pytorch_model-00000-of-00061.bin", "model.layers.0.self_attn.o_proj.weight": "pytorch_model-00000-of-00061.bin", "model.layers.0.mlp.gate_proj.weight": "pytorch_model-00000-of-00061.bin", "model.layers.0.mlp.down_proj.weight": "pytorch_model-00000-of-00061.bin", "model.layers.0.mlp.up_proj.weight": "pytorch_model-00000-of-00061.bin", "model.layers.0.self_attn.rotary_emb.inv_freq": "pytorch_model-00000-of-00061.bin", "model.layers.1.input_layernorm.weight": "pytorch_model-00001-of-00061.bin", "model.layers.1.post_attention_layernorm.weight": "pytorch_model-00001-of-00061.bin", "model.layers.1.self_attn.q_proj.weight": "pytorch_model-00001-of-00061.bin", "model.layers.1.self_attn.k_proj.weight": "pytorch_model-00001-of-00061.bin", "model.layers.1.self_attn.v_proj.weight": "pytorch_model-00001-of-00061.bin", "model.layers.1.self_attn.o_proj.weight": "pytorch_model-00001-of-00061.bin", "model.layers.1.mlp.gate_proj.weight": "pytorch_model-00001-of-00061.bin", "model.layers.1.mlp.down_proj.weight": "pytorch_model-00001-of-00061.bin", "model.layers.1.mlp.up_proj.weight": "pytorch_model-00001-of-00061.bin", "model.layers.1.self_attn.rotary_emb.inv_freq": "pytorch_model-00001-of-00061.bin", "model.layers.2.input_layernorm.weight": "pytorch_model-00002-of-00061.bin", "model.layers.2.post_attention_layernorm.weight": "pytorch_model-00002-of-00061.bin", "model.layers.2.self_attn.q_proj.weight": "pytorch_model-00002-of-00061.bin", "model.layers.2.self_attn.k_proj.weight": "pytorch_model-00002-of-00061.bin", "model.layers.2.self_attn.v_proj.weight": "pytorch_model-00002-of-00061.bin", "model.layers.2.self_attn.o_proj.weight": "pytorch_model-00002-of-00061.bin", "model.layers.2.mlp.gate_proj.weight": "pytorch_model-00002-of-00061.bin", "model.layers.2.mlp.down_proj.weight": "pytorch_model-00002-of-00061.bin", "model.layers.2.mlp.up_proj.weight": "pytorch_model-00002-of-00061.bin", "model.layers.2.self_attn.rotary_emb.inv_freq": "pytorch_model-00002-of-00061.bin", "model.layers.3.input_layernorm.weight": "pytorch_model-00003-of-00061.bin", "model.layers.3.post_attention_layernorm.weight": "pytorch_model-00003-of-00061.bin", "model.layers.3.self_attn.q_proj.weight": "pytorch_model-00003-of-00061.bin", "model.layers.3.self_attn.k_proj.weight": "pytorch_model-00003-of-00061.bin", "model.layers.3.self_attn.v_proj.weight": "pytorch_model-00003-of-00061.bin", "model.layers.3.self_attn.o_proj.weight": "pytorch_model-00003-of-00061.bin", "model.layers.3.mlp.gate_proj.weight": "pytorch_model-00003-of-00061.bin", "model.layers.3.mlp.down_proj.weight": "pytorch_model-00003-of-00061.bin", "model.layers.3.mlp.up_proj.weight": "pytorch_model-00003-of-00061.bin", "model.layers.3.self_attn.rotary_emb.inv_freq": "pytorch_model-00003-of-00061.bin", "model.layers.4.input_layernorm.weight": "pytorch_model-00004-of-00061.bin", "model.layers.4.post_attention_layernorm.weight": "pytorch_model-00004-of-00061.bin", "model.layers.4.self_attn.q_proj.weight": "pytorch_model-00004-of-00061.bin", "model.layers.4.self_attn.k_proj.weight": "pytorch_model-00004-of-00061.bin", "model.layers.4.self_attn.v_proj.weight": "pytorch_model-00004-of-00061.bin", "model.layers.4.self_attn.o_proj.weight": "pytorch_model-00004-of-00061.bin", "model.layers.4.mlp.gate_proj.weight": "pytorch_model-00004-of-00061.bin", "model.layers.4.mlp.down_proj.weight": "pytorch_model-00004-of-00061.bin", "model.layers.4.mlp.up_proj.weight": "pytorch_model-00004-of-00061.bin", "model.layers.4.self_attn.rotary_emb.inv_freq": "pytorch_model-00004-of-00061.bin", "model.layers.5.input_layernorm.weight": "pytorch_model-00005-of-00061.bin", "model.layers.5.post_attention_layernorm.weight": "pytorch_model-00005-of-00061.bin", "model.layers.5.self_attn.q_proj.weight": "pytorch_model-00005-of-00061.bin", "model.layers.5.self_attn.k_proj.weight": "pytorch_model-00005-of-00061.bin", "model.layers.5.self_attn.v_proj.weight": "pytorch_model-00005-of-00061.bin", "model.layers.5.self_attn.o_proj.weight": "pytorch_model-00005-of-00061.bin", "model.layers.5.mlp.gate_proj.weight": "pytorch_model-00005-of-00061.bin", "model.layers.5.mlp.down_proj.weight": "pytorch_model-00005-of-00061.bin", "model.layers.5.mlp.up_proj.weight": "pytorch_model-00005-of-00061.bin", "model.layers.5.self_attn.rotary_emb.inv_freq": "pytorch_model-00005-of-00061.bin", "model.layers.6.input_layernorm.weight": "pytorch_model-00006-of-00061.bin", "model.layers.6.post_attention_layernorm.weight": "pytorch_model-00006-of-00061.bin", "model.layers.6.self_attn.q_proj.weight": "pytorch_model-00006-of-00061.bin", "model.layers.6.self_attn.k_proj.weight": "pytorch_model-00006-of-00061.bin", "model.layers.6.self_attn.v_proj.weight": "pytorch_model-00006-of-00061.bin", "model.layers.6.self_attn.o_proj.weight": "pytorch_model-00006-of-00061.bin", "model.layers.6.mlp.gate_proj.weight": "pytorch_model-00006-of-00061.bin", "model.layers.6.mlp.down_proj.weight": "pytorch_model-00006-of-00061.bin", "model.layers.6.mlp.up_proj.weight": "pytorch_model-00006-of-00061.bin", "model.layers.6.self_attn.rotary_emb.inv_freq": "pytorch_model-00006-of-00061.bin", "model.layers.7.input_layernorm.weight": "pytorch_model-00007-of-00061.bin", "model.layers.7.post_attention_layernorm.weight": "pytorch_model-00007-of-00061.bin", "model.layers.7.self_attn.q_proj.weight": "pytorch_model-00007-of-00061.bin", "model.layers.7.self_attn.k_proj.weight": "pytorch_model-00007-of-00061.bin", "model.layers.7.self_attn.v_proj.weight": "pytorch_model-00007-of-00061.bin", "model.layers.7.self_attn.o_proj.weight": "pytorch_model-00007-of-00061.bin", "model.layers.7.mlp.gate_proj.weight": "pytorch_model-00007-of-00061.bin", "model.layers.7.mlp.down_proj.weight": "pytorch_model-00007-of-00061.bin", "model.layers.7.mlp.up_proj.weight": "pytorch_model-00007-of-00061.bin", "model.layers.7.self_attn.rotary_emb.inv_freq": "pytorch_model-00007-of-00061.bin", "model.layers.8.input_layernorm.weight": "pytorch_model-00008-of-00061.bin", "model.layers.8.post_attention_layernorm.weight": "pytorch_model-00008-of-00061.bin", "model.layers.8.self_attn.q_proj.weight": "pytorch_model-00008-of-00061.bin", "model.layers.8.self_attn.k_proj.weight": "pytorch_model-00008-of-00061.bin", "model.layers.8.self_attn.v_proj.weight": "pytorch_model-00008-of-00061.bin", "model.layers.8.self_attn.o_proj.weight": "pytorch_model-00008-of-00061.bin", "model.layers.8.mlp.gate_proj.weight": "pytorch_model-00008-of-00061.bin", "model.layers.8.mlp.down_proj.weight": "pytorch_model-00008-of-00061.bin", "model.layers.8.mlp.up_proj.weight": "pytorch_model-00008-of-00061.bin", "model.layers.8.self_attn.rotary_emb.inv_freq": "pytorch_model-00008-of-00061.bin", "model.layers.9.input_layernorm.weight": "pytorch_model-00009-of-00061.bin", "model.layers.9.post_attention_layernorm.weight": "pytorch_model-00009-of-00061.bin", "model.layers.9.self_attn.q_proj.weight": "pytorch_model-00009-of-00061.bin", "model.layers.9.self_attn.k_proj.weight": "pytorch_model-00009-of-00061.bin", "model.layers.9.self_attn.v_proj.weight": "pytorch_model-00009-of-00061.bin", "model.layers.9.self_attn.o_proj.weight": "pytorch_model-00009-of-00061.bin", "model.layers.9.mlp.gate_proj.weight": "pytorch_model-00009-of-00061.bin", "model.layers.9.mlp.down_proj.weight": "pytorch_model-00009-of-00061.bin", "model.layers.9.mlp.up_proj.weight": "pytorch_model-00009-of-00061.bin", "model.layers.9.self_attn.rotary_emb.inv_freq": "pytorch_model-00009-of-00061.bin", "model.layers.10.input_layernorm.weight": "pytorch_model-00010-of-00061.bin", "model.layers.10.post_attention_layernorm.weight": "pytorch_model-00010-of-00061.bin", "model.layers.10.self_attn.q_proj.weight": "pytorch_model-00010-of-00061.bin", "model.layers.10.self_attn.k_proj.weight": "pytorch_model-00010-of-00061.bin", "model.layers.10.self_attn.v_proj.weight": "pytorch_model-00010-of-00061.bin", "model.layers.10.self_attn.o_proj.weight": "pytorch_model-00010-of-00061.bin", "model.layers.10.mlp.gate_proj.weight": "pytorch_model-00010-of-00061.bin", "model.layers.10.mlp.down_proj.weight": "pytorch_model-00010-of-00061.bin", "model.layers.10.mlp.up_proj.weight": "pytorch_model-00010-of-00061.bin", "model.layers.10.self_attn.rotary_emb.inv_freq": "pytorch_model-00010-of-00061.bin", "model.layers.11.input_layernorm.weight": "pytorch_model-00011-of-00061.bin", "model.layers.11.post_attention_layernorm.weight": "pytorch_model-00011-of-00061.bin", "model.layers.11.self_attn.q_proj.weight": "pytorch_model-00011-of-00061.bin", "model.layers.11.self_attn.k_proj.weight": "pytorch_model-00011-of-00061.bin", "model.layers.11.self_attn.v_proj.weight": "pytorch_model-00011-of-00061.bin", "model.layers.11.self_attn.o_proj.weight": "pytorch_model-00011-of-00061.bin", "model.layers.11.mlp.gate_proj.weight": "pytorch_model-00011-of-00061.bin", "model.layers.11.mlp.down_proj.weight": "pytorch_model-00011-of-00061.bin", "model.layers.11.mlp.up_proj.weight": "pytorch_model-00011-of-00061.bin", "model.layers.11.self_attn.rotary_emb.inv_freq": "pytorch_model-00011-of-00061.bin", "model.layers.12.input_layernorm.weight": "pytorch_model-00012-of-00061.bin", "model.layers.12.post_attention_layernorm.weight": "pytorch_model-00012-of-00061.bin", "model.layers.12.self_attn.q_proj.weight": "pytorch_model-00012-of-00061.bin", "model.layers.12.self_attn.k_proj.weight": "pytorch_model-00012-of-00061.bin", "model.layers.12.self_attn.v_proj.weight": "pytorch_model-00012-of-00061.bin", "model.layers.12.self_attn.o_proj.weight": "pytorch_model-00012-of-00061.bin", "model.layers.12.mlp.gate_proj.weight": "pytorch_model-00012-of-00061.bin", "model.layers.12.mlp.down_proj.weight": "pytorch_model-00012-of-00061.bin", "model.layers.12.mlp.up_proj.weight": "pytorch_model-00012-of-00061.bin", "model.layers.12.self_attn.rotary_emb.inv_freq": "pytorch_model-00012-of-00061.bin", "model.layers.13.input_layernorm.weight": "pytorch_model-00013-of-00061.bin", "model.layers.13.post_attention_layernorm.weight": "pytorch_model-00013-of-00061.bin", "model.layers.13.self_attn.q_proj.weight": "pytorch_model-00013-of-00061.bin", "model.layers.13.self_attn.k_proj.weight": "pytorch_model-00013-of-00061.bin", "model.layers.13.self_attn.v_proj.weight": "pytorch_model-00013-of-00061.bin", "model.layers.13.self_attn.o_proj.weight": "pytorch_model-00013-of-00061.bin", "model.layers.13.mlp.gate_proj.weight": "pytorch_model-00013-of-00061.bin", "model.layers.13.mlp.down_proj.weight": "pytorch_model-00013-of-00061.bin", "model.layers.13.mlp.up_proj.weight": "pytorch_model-00013-of-00061.bin", "model.layers.13.self_attn.rotary_emb.inv_freq": "pytorch_model-00013-of-00061.bin", "model.layers.14.input_layernorm.weight": "pytorch_model-00014-of-00061.bin", "model.layers.14.post_attention_layernorm.weight": "pytorch_model-00014-of-00061.bin", "model.layers.14.self_attn.q_proj.weight": "pytorch_model-00014-of-00061.bin", "model.layers.14.self_attn.k_proj.weight": "pytorch_model-00014-of-00061.bin", "model.layers.14.self_attn.v_proj.weight": "pytorch_model-00014-of-00061.bin", "model.layers.14.self_attn.o_proj.weight": "pytorch_model-00014-of-00061.bin", "model.layers.14.mlp.gate_proj.weight": "pytorch_model-00014-of-00061.bin", "model.layers.14.mlp.down_proj.weight": "pytorch_model-00014-of-00061.bin", "model.layers.14.mlp.up_proj.weight": "pytorch_model-00014-of-00061.bin", "model.layers.14.self_attn.rotary_emb.inv_freq": "pytorch_model-00014-of-00061.bin", "model.layers.15.input_layernorm.weight": "pytorch_model-00015-of-00061.bin", "model.layers.15.post_attention_layernorm.weight": "pytorch_model-00015-of-00061.bin", "model.layers.15.self_attn.q_proj.weight": "pytorch_model-00015-of-00061.bin", "model.layers.15.self_attn.k_proj.weight": "pytorch_model-00015-of-00061.bin", "model.layers.15.self_attn.v_proj.weight": "pytorch_model-00015-of-00061.bin", "model.layers.15.self_attn.o_proj.weight": "pytorch_model-00015-of-00061.bin", "model.layers.15.mlp.gate_proj.weight": "pytorch_model-00015-of-00061.bin", "model.layers.15.mlp.down_proj.weight": "pytorch_model-00015-of-00061.bin", "model.layers.15.mlp.up_proj.weight": "pytorch_model-00015-of-00061.bin", "model.layers.15.self_attn.rotary_emb.inv_freq": "pytorch_model-00015-of-00061.bin", "model.layers.16.input_layernorm.weight": "pytorch_model-00016-of-00061.bin", "model.layers.16.post_attention_layernorm.weight": "pytorch_model-00016-of-00061.bin", "model.layers.16.self_attn.q_proj.weight": "pytorch_model-00016-of-00061.bin", "model.layers.16.self_attn.k_proj.weight": "pytorch_model-00016-of-00061.bin", "model.layers.16.self_attn.v_proj.weight": "pytorch_model-00016-of-00061.bin", "model.layers.16.self_attn.o_proj.weight": "pytorch_model-00016-of-00061.bin", "model.layers.16.mlp.gate_proj.weight": "pytorch_model-00016-of-00061.bin", "model.layers.16.mlp.down_proj.weight": "pytorch_model-00016-of-00061.bin", "model.layers.16.mlp.up_proj.weight": "pytorch_model-00016-of-00061.bin", "model.layers.16.self_attn.rotary_emb.inv_freq": "pytorch_model-00016-of-00061.bin", "model.layers.17.input_layernorm.weight": "pytorch_model-00017-of-00061.bin", "model.layers.17.post_attention_layernorm.weight": "pytorch_model-00017-of-00061.bin", "model.layers.17.self_attn.q_proj.weight": "pytorch_model-00017-of-00061.bin", "model.layers.17.self_attn.k_proj.weight": "pytorch_model-00017-of-00061.bin", "model.layers.17.self_attn.v_proj.weight": "pytorch_model-00017-of-00061.bin", "model.layers.17.self_attn.o_proj.weight": "pytorch_model-00017-of-00061.bin", "model.layers.17.mlp.gate_proj.weight": "pytorch_model-00017-of-00061.bin", "model.layers.17.mlp.down_proj.weight": "pytorch_model-00017-of-00061.bin", "model.layers.17.mlp.up_proj.weight": "pytorch_model-00017-of-00061.bin", "model.layers.17.self_attn.rotary_emb.inv_freq": "pytorch_model-00017-of-00061.bin", "model.layers.18.input_layernorm.weight": "pytorch_model-00018-of-00061.bin", "model.layers.18.post_attention_layernorm.weight": "pytorch_model-00018-of-00061.bin", "model.layers.18.self_attn.q_proj.weight": "pytorch_model-00018-of-00061.bin", "model.layers.18.self_attn.k_proj.weight": "pytorch_model-00018-of-00061.bin", "model.layers.18.self_attn.v_proj.weight": "pytorch_model-00018-of-00061.bin", "model.layers.18.self_attn.o_proj.weight": "pytorch_model-00018-of-00061.bin", "model.layers.18.mlp.gate_proj.weight": "pytorch_model-00018-of-00061.bin", "model.layers.18.mlp.down_proj.weight": "pytorch_model-00018-of-00061.bin", "model.layers.18.mlp.up_proj.weight": "pytorch_model-00018-of-00061.bin", "model.layers.18.self_attn.rotary_emb.inv_freq": "pytorch_model-00018-of-00061.bin", "model.layers.19.input_layernorm.weight": "pytorch_model-00019-of-00061.bin", "model.layers.19.post_attention_layernorm.weight": "pytorch_model-00019-of-00061.bin", "model.layers.19.self_attn.q_proj.weight": "pytorch_model-00019-of-00061.bin", "model.layers.19.self_attn.k_proj.weight": "pytorch_model-00019-of-00061.bin", "model.layers.19.self_attn.v_proj.weight": "pytorch_model-00019-of-00061.bin", "model.layers.19.self_attn.o_proj.weight": "pytorch_model-00019-of-00061.bin", "model.layers.19.mlp.gate_proj.weight": "pytorch_model-00019-of-00061.bin", "model.layers.19.mlp.down_proj.weight": "pytorch_model-00019-of-00061.bin", "model.layers.19.mlp.up_proj.weight": "pytorch_model-00019-of-00061.bin", "model.layers.19.self_attn.rotary_emb.inv_freq": "pytorch_model-00019-of-00061.bin", "model.layers.20.input_layernorm.weight": "pytorch_model-00020-of-00061.bin", "model.layers.20.post_attention_layernorm.weight": "pytorch_model-00020-of-00061.bin", "model.layers.20.self_attn.q_proj.weight": "pytorch_model-00020-of-00061.bin", "model.layers.20.self_attn.k_proj.weight": "pytorch_model-00020-of-00061.bin", "model.layers.20.self_attn.v_proj.weight": "pytorch_model-00020-of-00061.bin", "model.layers.20.self_attn.o_proj.weight": "pytorch_model-00020-of-00061.bin", "model.layers.20.mlp.gate_proj.weight": "pytorch_model-00020-of-00061.bin", "model.layers.20.mlp.down_proj.weight": "pytorch_model-00020-of-00061.bin", "model.layers.20.mlp.up_proj.weight": "pytorch_model-00020-of-00061.bin", "model.layers.20.self_attn.rotary_emb.inv_freq": "pytorch_model-00020-of-00061.bin", "model.layers.21.input_layernorm.weight": "pytorch_model-00021-of-00061.bin", "model.layers.21.post_attention_layernorm.weight": "pytorch_model-00021-of-00061.bin", "model.layers.21.self_attn.q_proj.weight": "pytorch_model-00021-of-00061.bin", "model.layers.21.self_attn.k_proj.weight": "pytorch_model-00021-of-00061.bin", "model.layers.21.self_attn.v_proj.weight": "pytorch_model-00021-of-00061.bin", "model.layers.21.self_attn.o_proj.weight": "pytorch_model-00021-of-00061.bin", "model.layers.21.mlp.gate_proj.weight": "pytorch_model-00021-of-00061.bin", "model.layers.21.mlp.down_proj.weight": "pytorch_model-00021-of-00061.bin", "model.layers.21.mlp.up_proj.weight": "pytorch_model-00021-of-00061.bin", "model.layers.21.self_attn.rotary_emb.inv_freq": "pytorch_model-00021-of-00061.bin", "model.layers.22.input_layernorm.weight": "pytorch_model-00022-of-00061.bin", "model.layers.22.post_attention_layernorm.weight": "pytorch_model-00022-of-00061.bin", "model.layers.22.self_attn.q_proj.weight": "pytorch_model-00022-of-00061.bin", "model.layers.22.self_attn.k_proj.weight": "pytorch_model-00022-of-00061.bin", "model.layers.22.self_attn.v_proj.weight": "pytorch_model-00022-of-00061.bin", "model.layers.22.self_attn.o_proj.weight": "pytorch_model-00022-of-00061.bin", "model.layers.22.mlp.gate_proj.weight": "pytorch_model-00022-of-00061.bin", "model.layers.22.mlp.down_proj.weight": "pytorch_model-00022-of-00061.bin", "model.layers.22.mlp.up_proj.weight": "pytorch_model-00022-of-00061.bin", "model.layers.22.self_attn.rotary_emb.inv_freq": "pytorch_model-00022-of-00061.bin", "model.layers.23.input_layernorm.weight": "pytorch_model-00023-of-00061.bin", "model.layers.23.post_attention_layernorm.weight": "pytorch_model-00023-of-00061.bin", "model.layers.23.self_attn.q_proj.weight": "pytorch_model-00023-of-00061.bin", "model.layers.23.self_attn.k_proj.weight": "pytorch_model-00023-of-00061.bin", "model.layers.23.self_attn.v_proj.weight": "pytorch_model-00023-of-00061.bin", "model.layers.23.self_attn.o_proj.weight": "pytorch_model-00023-of-00061.bin", "model.layers.23.mlp.gate_proj.weight": "pytorch_model-00023-of-00061.bin", "model.layers.23.mlp.down_proj.weight": "pytorch_model-00023-of-00061.bin", "model.layers.23.mlp.up_proj.weight": "pytorch_model-00023-of-00061.bin", "model.layers.23.self_attn.rotary_emb.inv_freq": "pytorch_model-00023-of-00061.bin", "model.layers.24.input_layernorm.weight": "pytorch_model-00024-of-00061.bin", "model.layers.24.post_attention_layernorm.weight": "pytorch_model-00024-of-00061.bin", "model.layers.24.self_attn.q_proj.weight": "pytorch_model-00024-of-00061.bin", "model.layers.24.self_attn.k_proj.weight": "pytorch_model-00024-of-00061.bin", "model.layers.24.self_attn.v_proj.weight": "pytorch_model-00024-of-00061.bin", "model.layers.24.self_attn.o_proj.weight": "pytorch_model-00024-of-00061.bin", "model.layers.24.mlp.gate_proj.weight": "pytorch_model-00024-of-00061.bin", "model.layers.24.mlp.down_proj.weight": "pytorch_model-00024-of-00061.bin", "model.layers.24.mlp.up_proj.weight": "pytorch_model-00024-of-00061.bin", "model.layers.24.self_attn.rotary_emb.inv_freq": "pytorch_model-00024-of-00061.bin", "model.layers.25.input_layernorm.weight": "pytorch_model-00025-of-00061.bin", "model.layers.25.post_attention_layernorm.weight": "pytorch_model-00025-of-00061.bin", "model.layers.25.self_attn.q_proj.weight": "pytorch_model-00025-of-00061.bin", "model.layers.25.self_attn.k_proj.weight": "pytorch_model-00025-of-00061.bin", "model.layers.25.self_attn.v_proj.weight": "pytorch_model-00025-of-00061.bin", "model.layers.25.self_attn.o_proj.weight": "pytorch_model-00025-of-00061.bin", "model.layers.25.mlp.gate_proj.weight": "pytorch_model-00025-of-00061.bin", "model.layers.25.mlp.down_proj.weight": "pytorch_model-00025-of-00061.bin", "model.layers.25.mlp.up_proj.weight": "pytorch_model-00025-of-00061.bin", "model.layers.25.self_attn.rotary_emb.inv_freq": "pytorch_model-00025-of-00061.bin", "model.layers.26.input_layernorm.weight": "pytorch_model-00026-of-00061.bin", "model.layers.26.post_attention_layernorm.weight": "pytorch_model-00026-of-00061.bin", "model.layers.26.self_attn.q_proj.weight": "pytorch_model-00026-of-00061.bin", "model.layers.26.self_attn.k_proj.weight": "pytorch_model-00026-of-00061.bin", "model.layers.26.self_attn.v_proj.weight": "pytorch_model-00026-of-00061.bin", "model.layers.26.self_attn.o_proj.weight": "pytorch_model-00026-of-00061.bin", "model.layers.26.mlp.gate_proj.weight": "pytorch_model-00026-of-00061.bin", "model.layers.26.mlp.down_proj.weight": "pytorch_model-00026-of-00061.bin", "model.layers.26.mlp.up_proj.weight": "pytorch_model-00026-of-00061.bin", "model.layers.26.self_attn.rotary_emb.inv_freq": "pytorch_model-00026-of-00061.bin", "model.layers.27.input_layernorm.weight": "pytorch_model-00027-of-00061.bin", "model.layers.27.post_attention_layernorm.weight": "pytorch_model-00027-of-00061.bin", "model.layers.27.self_attn.q_proj.weight": "pytorch_model-00027-of-00061.bin", "model.layers.27.self_attn.k_proj.weight": "pytorch_model-00027-of-00061.bin", "model.layers.27.self_attn.v_proj.weight": "pytorch_model-00027-of-00061.bin", "model.layers.27.self_attn.o_proj.weight": "pytorch_model-00027-of-00061.bin", "model.layers.27.mlp.gate_proj.weight": "pytorch_model-00027-of-00061.bin", "model.layers.27.mlp.down_proj.weight": "pytorch_model-00027-of-00061.bin", "model.layers.27.mlp.up_proj.weight": "pytorch_model-00027-of-00061.bin", "model.layers.27.self_attn.rotary_emb.inv_freq": "pytorch_model-00027-of-00061.bin", "model.layers.28.input_layernorm.weight": "pytorch_model-00028-of-00061.bin", "model.layers.28.post_attention_layernorm.weight": "pytorch_model-00028-of-00061.bin", "model.layers.28.self_attn.q_proj.weight": "pytorch_model-00028-of-00061.bin", "model.layers.28.self_attn.k_proj.weight": "pytorch_model-00028-of-00061.bin", "model.layers.28.self_attn.v_proj.weight": "pytorch_model-00028-of-00061.bin", "model.layers.28.self_attn.o_proj.weight": "pytorch_model-00028-of-00061.bin", "model.layers.28.mlp.gate_proj.weight": "pytorch_model-00028-of-00061.bin", "model.layers.28.mlp.down_proj.weight": "pytorch_model-00028-of-00061.bin", "model.layers.28.mlp.up_proj.weight": "pytorch_model-00028-of-00061.bin", "model.layers.28.self_attn.rotary_emb.inv_freq": "pytorch_model-00028-of-00061.bin", "model.layers.29.input_layernorm.weight": "pytorch_model-00029-of-00061.bin", "model.layers.29.post_attention_layernorm.weight": "pytorch_model-00029-of-00061.bin", "model.layers.29.self_attn.q_proj.weight": "pytorch_model-00029-of-00061.bin", "model.layers.29.self_attn.k_proj.weight": "pytorch_model-00029-of-00061.bin", "model.layers.29.self_attn.v_proj.weight": "pytorch_model-00029-of-00061.bin", "model.layers.29.self_attn.o_proj.weight": "pytorch_model-00029-of-00061.bin", "model.layers.29.mlp.gate_proj.weight": "pytorch_model-00029-of-00061.bin", "model.layers.29.mlp.down_proj.weight": "pytorch_model-00029-of-00061.bin", "model.layers.29.mlp.up_proj.weight": "pytorch_model-00029-of-00061.bin", "model.layers.29.self_attn.rotary_emb.inv_freq": "pytorch_model-00029-of-00061.bin", "model.layers.30.input_layernorm.weight": "pytorch_model-00030-of-00061.bin", "model.layers.30.post_attention_layernorm.weight": "pytorch_model-00030-of-00061.bin", "model.layers.30.self_attn.q_proj.weight": "pytorch_model-00030-of-00061.bin", "model.layers.30.self_attn.k_proj.weight": "pytorch_model-00030-of-00061.bin", "model.layers.30.self_attn.v_proj.weight": "pytorch_model-00030-of-00061.bin", "model.layers.30.self_attn.o_proj.weight": "pytorch_model-00030-of-00061.bin", "model.layers.30.mlp.gate_proj.weight": "pytorch_model-00030-of-00061.bin", "model.layers.30.mlp.down_proj.weight": "pytorch_model-00030-of-00061.bin", "model.layers.30.mlp.up_proj.weight": "pytorch_model-00030-of-00061.bin", "model.layers.30.self_attn.rotary_emb.inv_freq": "pytorch_model-00030-of-00061.bin", "model.layers.31.input_layernorm.weight": "pytorch_model-00031-of-00061.bin", "model.layers.31.post_attention_layernorm.weight": "pytorch_model-00031-of-00061.bin", "model.layers.31.self_attn.q_proj.weight": "pytorch_model-00031-of-00061.bin", "model.layers.31.self_attn.k_proj.weight": "pytorch_model-00031-of-00061.bin", "model.layers.31.self_attn.v_proj.weight": "pytorch_model-00031-of-00061.bin", "model.layers.31.self_attn.o_proj.weight": "pytorch_model-00031-of-00061.bin", "model.layers.31.mlp.gate_proj.weight": "pytorch_model-00031-of-00061.bin", "model.layers.31.mlp.down_proj.weight": "pytorch_model-00031-of-00061.bin", "model.layers.31.mlp.up_proj.weight": "pytorch_model-00031-of-00061.bin", "model.layers.31.self_attn.rotary_emb.inv_freq": "pytorch_model-00031-of-00061.bin", "model.layers.32.input_layernorm.weight": "pytorch_model-00032-of-00061.bin", "model.layers.32.post_attention_layernorm.weight": "pytorch_model-00032-of-00061.bin", "model.layers.32.self_attn.q_proj.weight": "pytorch_model-00032-of-00061.bin", "model.layers.32.self_attn.k_proj.weight": "pytorch_model-00032-of-00061.bin", "model.layers.32.self_attn.v_proj.weight": "pytorch_model-00032-of-00061.bin", "model.layers.32.self_attn.o_proj.weight": "pytorch_model-00032-of-00061.bin", "model.layers.32.mlp.gate_proj.weight": "pytorch_model-00032-of-00061.bin", "model.layers.32.mlp.down_proj.weight": "pytorch_model-00032-of-00061.bin", "model.layers.32.mlp.up_proj.weight": "pytorch_model-00032-of-00061.bin", "model.layers.32.self_attn.rotary_emb.inv_freq": "pytorch_model-00032-of-00061.bin", "model.layers.33.input_layernorm.weight": "pytorch_model-00033-of-00061.bin", "model.layers.33.post_attention_layernorm.weight": "pytorch_model-00033-of-00061.bin", "model.layers.33.self_attn.q_proj.weight": "pytorch_model-00033-of-00061.bin", "model.layers.33.self_attn.k_proj.weight": "pytorch_model-00033-of-00061.bin", "model.layers.33.self_attn.v_proj.weight": "pytorch_model-00033-of-00061.bin", "model.layers.33.self_attn.o_proj.weight": "pytorch_model-00033-of-00061.bin", "model.layers.33.mlp.gate_proj.weight": "pytorch_model-00033-of-00061.bin", "model.layers.33.mlp.down_proj.weight": "pytorch_model-00033-of-00061.bin", "model.layers.33.mlp.up_proj.weight": "pytorch_model-00033-of-00061.bin", "model.layers.33.self_attn.rotary_emb.inv_freq": "pytorch_model-00033-of-00061.bin", "model.layers.34.input_layernorm.weight": "pytorch_model-00034-of-00061.bin", "model.layers.34.post_attention_layernorm.weight": "pytorch_model-00034-of-00061.bin", "model.layers.34.self_attn.q_proj.weight": "pytorch_model-00034-of-00061.bin", "model.layers.34.self_attn.k_proj.weight": "pytorch_model-00034-of-00061.bin", "model.layers.34.self_attn.v_proj.weight": "pytorch_model-00034-of-00061.bin", "model.layers.34.self_attn.o_proj.weight": "pytorch_model-00034-of-00061.bin", "model.layers.34.mlp.gate_proj.weight": "pytorch_model-00034-of-00061.bin", "model.layers.34.mlp.down_proj.weight": "pytorch_model-00034-of-00061.bin", "model.layers.34.mlp.up_proj.weight": "pytorch_model-00034-of-00061.bin", "model.layers.34.self_attn.rotary_emb.inv_freq": "pytorch_model-00034-of-00061.bin", "model.layers.35.input_layernorm.weight": "pytorch_model-00035-of-00061.bin", "model.layers.35.post_attention_layernorm.weight": "pytorch_model-00035-of-00061.bin", "model.layers.35.self_attn.q_proj.weight": "pytorch_model-00035-of-00061.bin", "model.layers.35.self_attn.k_proj.weight": "pytorch_model-00035-of-00061.bin", "model.layers.35.self_attn.v_proj.weight": "pytorch_model-00035-of-00061.bin", "model.layers.35.self_attn.o_proj.weight": "pytorch_model-00035-of-00061.bin", "model.layers.35.mlp.gate_proj.weight": "pytorch_model-00035-of-00061.bin", "model.layers.35.mlp.down_proj.weight": "pytorch_model-00035-of-00061.bin", "model.layers.35.mlp.up_proj.weight": "pytorch_model-00035-of-00061.bin", "model.layers.35.self_attn.rotary_emb.inv_freq": "pytorch_model-00035-of-00061.bin", "model.layers.36.input_layernorm.weight": "pytorch_model-00036-of-00061.bin", "model.layers.36.post_attention_layernorm.weight": "pytorch_model-00036-of-00061.bin", "model.layers.36.self_attn.q_proj.weight": "pytorch_model-00036-of-00061.bin", "model.layers.36.self_attn.k_proj.weight": "pytorch_model-00036-of-00061.bin", "model.layers.36.self_attn.v_proj.weight": "pytorch_model-00036-of-00061.bin", "model.layers.36.self_attn.o_proj.weight": "pytorch_model-00036-of-00061.bin", "model.layers.36.mlp.gate_proj.weight": "pytorch_model-00036-of-00061.bin", "model.layers.36.mlp.down_proj.weight": "pytorch_model-00036-of-00061.bin", "model.layers.36.mlp.up_proj.weight": "pytorch_model-00036-of-00061.bin", "model.layers.36.self_attn.rotary_emb.inv_freq": "pytorch_model-00036-of-00061.bin", "model.layers.37.input_layernorm.weight": "pytorch_model-00037-of-00061.bin", "model.layers.37.post_attention_layernorm.weight": "pytorch_model-00037-of-00061.bin", "model.layers.37.self_attn.q_proj.weight": "pytorch_model-00037-of-00061.bin", "model.layers.37.self_attn.k_proj.weight": "pytorch_model-00037-of-00061.bin", "model.layers.37.self_attn.v_proj.weight": "pytorch_model-00037-of-00061.bin", "model.layers.37.self_attn.o_proj.weight": "pytorch_model-00037-of-00061.bin", "model.layers.37.mlp.gate_proj.weight": "pytorch_model-00037-of-00061.bin", "model.layers.37.mlp.down_proj.weight": "pytorch_model-00037-of-00061.bin", "model.layers.37.mlp.up_proj.weight": "pytorch_model-00037-of-00061.bin", "model.layers.37.self_attn.rotary_emb.inv_freq": "pytorch_model-00037-of-00061.bin", "model.layers.38.input_layernorm.weight": "pytorch_model-00038-of-00061.bin", "model.layers.38.post_attention_layernorm.weight": "pytorch_model-00038-of-00061.bin", "model.layers.38.self_attn.q_proj.weight": "pytorch_model-00038-of-00061.bin", "model.layers.38.self_attn.k_proj.weight": "pytorch_model-00038-of-00061.bin", "model.layers.38.self_attn.v_proj.weight": "pytorch_model-00038-of-00061.bin", "model.layers.38.self_attn.o_proj.weight": "pytorch_model-00038-of-00061.bin", "model.layers.38.mlp.gate_proj.weight": "pytorch_model-00038-of-00061.bin", "model.layers.38.mlp.down_proj.weight": "pytorch_model-00038-of-00061.bin", "model.layers.38.mlp.up_proj.weight": "pytorch_model-00038-of-00061.bin", "model.layers.38.self_attn.rotary_emb.inv_freq": "pytorch_model-00038-of-00061.bin", "model.layers.39.input_layernorm.weight": "pytorch_model-00039-of-00061.bin", "model.layers.39.post_attention_layernorm.weight": "pytorch_model-00039-of-00061.bin", "model.layers.39.self_attn.q_proj.weight": "pytorch_model-00039-of-00061.bin", "model.layers.39.self_attn.k_proj.weight": "pytorch_model-00039-of-00061.bin", "model.layers.39.self_attn.v_proj.weight": "pytorch_model-00039-of-00061.bin", "model.layers.39.self_attn.o_proj.weight": "pytorch_model-00039-of-00061.bin", "model.layers.39.mlp.gate_proj.weight": "pytorch_model-00039-of-00061.bin", "model.layers.39.mlp.down_proj.weight": "pytorch_model-00039-of-00061.bin", "model.layers.39.mlp.up_proj.weight": "pytorch_model-00039-of-00061.bin", "model.layers.39.self_attn.rotary_emb.inv_freq": "pytorch_model-00039-of-00061.bin", "model.layers.40.input_layernorm.weight": "pytorch_model-00040-of-00061.bin", "model.layers.40.post_attention_layernorm.weight": "pytorch_model-00040-of-00061.bin", "model.layers.40.self_attn.q_proj.weight": "pytorch_model-00040-of-00061.bin", "model.layers.40.self_attn.k_proj.weight": "pytorch_model-00040-of-00061.bin", "model.layers.40.self_attn.v_proj.weight": "pytorch_model-00040-of-00061.bin", "model.layers.40.self_attn.o_proj.weight": "pytorch_model-00040-of-00061.bin", "model.layers.40.mlp.gate_proj.weight": "pytorch_model-00040-of-00061.bin", "model.layers.40.mlp.down_proj.weight": "pytorch_model-00040-of-00061.bin", "model.layers.40.mlp.up_proj.weight": "pytorch_model-00040-of-00061.bin", "model.layers.40.self_attn.rotary_emb.inv_freq": "pytorch_model-00040-of-00061.bin", "model.layers.41.input_layernorm.weight": "pytorch_model-00041-of-00061.bin", "model.layers.41.post_attention_layernorm.weight": "pytorch_model-00041-of-00061.bin", "model.layers.41.self_attn.q_proj.weight": "pytorch_model-00041-of-00061.bin", "model.layers.41.self_attn.k_proj.weight": "pytorch_model-00041-of-00061.bin", "model.layers.41.self_attn.v_proj.weight": "pytorch_model-00041-of-00061.bin", "model.layers.41.self_attn.o_proj.weight": "pytorch_model-00041-of-00061.bin", "model.layers.41.mlp.gate_proj.weight": "pytorch_model-00041-of-00061.bin", "model.layers.41.mlp.down_proj.weight": "pytorch_model-00041-of-00061.bin", "model.layers.41.mlp.up_proj.weight": "pytorch_model-00041-of-00061.bin", "model.layers.41.self_attn.rotary_emb.inv_freq": "pytorch_model-00041-of-00061.bin", "model.layers.42.input_layernorm.weight": "pytorch_model-00042-of-00061.bin", "model.layers.42.post_attention_layernorm.weight": "pytorch_model-00042-of-00061.bin", "model.layers.42.self_attn.q_proj.weight": "pytorch_model-00042-of-00061.bin", "model.layers.42.self_attn.k_proj.weight": "pytorch_model-00042-of-00061.bin", "model.layers.42.self_attn.v_proj.weight": "pytorch_model-00042-of-00061.bin", "model.layers.42.self_attn.o_proj.weight": "pytorch_model-00042-of-00061.bin", "model.layers.42.mlp.gate_proj.weight": "pytorch_model-00042-of-00061.bin", "model.layers.42.mlp.down_proj.weight": "pytorch_model-00042-of-00061.bin", "model.layers.42.mlp.up_proj.weight": "pytorch_model-00042-of-00061.bin", "model.layers.42.self_attn.rotary_emb.inv_freq": "pytorch_model-00042-of-00061.bin", "model.layers.43.input_layernorm.weight": "pytorch_model-00043-of-00061.bin", "model.layers.43.post_attention_layernorm.weight": "pytorch_model-00043-of-00061.bin", "model.layers.43.self_attn.q_proj.weight": "pytorch_model-00043-of-00061.bin", "model.layers.43.self_attn.k_proj.weight": "pytorch_model-00043-of-00061.bin", "model.layers.43.self_attn.v_proj.weight": "pytorch_model-00043-of-00061.bin", "model.layers.43.self_attn.o_proj.weight": "pytorch_model-00043-of-00061.bin", "model.layers.43.mlp.gate_proj.weight": "pytorch_model-00043-of-00061.bin", "model.layers.43.mlp.down_proj.weight": "pytorch_model-00043-of-00061.bin", "model.layers.43.mlp.up_proj.weight": "pytorch_model-00043-of-00061.bin", "model.layers.43.self_attn.rotary_emb.inv_freq": "pytorch_model-00043-of-00061.bin", "model.layers.44.input_layernorm.weight": "pytorch_model-00044-of-00061.bin", "model.layers.44.post_attention_layernorm.weight": "pytorch_model-00044-of-00061.bin", "model.layers.44.self_attn.q_proj.weight": "pytorch_model-00044-of-00061.bin", "model.layers.44.self_attn.k_proj.weight": "pytorch_model-00044-of-00061.bin", "model.layers.44.self_attn.v_proj.weight": "pytorch_model-00044-of-00061.bin", "model.layers.44.self_attn.o_proj.weight": "pytorch_model-00044-of-00061.bin", "model.layers.44.mlp.gate_proj.weight": "pytorch_model-00044-of-00061.bin", "model.layers.44.mlp.down_proj.weight": "pytorch_model-00044-of-00061.bin", "model.layers.44.mlp.up_proj.weight": "pytorch_model-00044-of-00061.bin", "model.layers.44.self_attn.rotary_emb.inv_freq": "pytorch_model-00044-of-00061.bin", "model.layers.45.input_layernorm.weight": "pytorch_model-00045-of-00061.bin", "model.layers.45.post_attention_layernorm.weight": "pytorch_model-00045-of-00061.bin", "model.layers.45.self_attn.q_proj.weight": "pytorch_model-00045-of-00061.bin", "model.layers.45.self_attn.k_proj.weight": "pytorch_model-00045-of-00061.bin", "model.layers.45.self_attn.v_proj.weight": "pytorch_model-00045-of-00061.bin", "model.layers.45.self_attn.o_proj.weight": "pytorch_model-00045-of-00061.bin", "model.layers.45.mlp.gate_proj.weight": "pytorch_model-00045-of-00061.bin", "model.layers.45.mlp.down_proj.weight": "pytorch_model-00045-of-00061.bin", "model.layers.45.mlp.up_proj.weight": "pytorch_model-00045-of-00061.bin", "model.layers.45.self_attn.rotary_emb.inv_freq": "pytorch_model-00045-of-00061.bin", "model.layers.46.input_layernorm.weight": "pytorch_model-00046-of-00061.bin", "model.layers.46.post_attention_layernorm.weight": "pytorch_model-00046-of-00061.bin", "model.layers.46.self_attn.q_proj.weight": "pytorch_model-00046-of-00061.bin", "model.layers.46.self_attn.k_proj.weight": "pytorch_model-00046-of-00061.bin", "model.layers.46.self_attn.v_proj.weight": "pytorch_model-00046-of-00061.bin", "model.layers.46.self_attn.o_proj.weight": "pytorch_model-00046-of-00061.bin", "model.layers.46.mlp.gate_proj.weight": "pytorch_model-00046-of-00061.bin", "model.layers.46.mlp.down_proj.weight": "pytorch_model-00046-of-00061.bin", "model.layers.46.mlp.up_proj.weight": "pytorch_model-00046-of-00061.bin", "model.layers.46.self_attn.rotary_emb.inv_freq": "pytorch_model-00046-of-00061.bin", "model.layers.47.input_layernorm.weight": "pytorch_model-00047-of-00061.bin", "model.layers.47.post_attention_layernorm.weight": "pytorch_model-00047-of-00061.bin", "model.layers.47.self_attn.q_proj.weight": "pytorch_model-00047-of-00061.bin", "model.layers.47.self_attn.k_proj.weight": "pytorch_model-00047-of-00061.bin", "model.layers.47.self_attn.v_proj.weight": "pytorch_model-00047-of-00061.bin", "model.layers.47.self_attn.o_proj.weight": "pytorch_model-00047-of-00061.bin", "model.layers.47.mlp.gate_proj.weight": "pytorch_model-00047-of-00061.bin", "model.layers.47.mlp.down_proj.weight": "pytorch_model-00047-of-00061.bin", "model.layers.47.mlp.up_proj.weight": "pytorch_model-00047-of-00061.bin", "model.layers.47.self_attn.rotary_emb.inv_freq": "pytorch_model-00047-of-00061.bin", "model.layers.48.input_layernorm.weight": "pytorch_model-00048-of-00061.bin", "model.layers.48.post_attention_layernorm.weight": "pytorch_model-00048-of-00061.bin", "model.layers.48.self_attn.q_proj.weight": "pytorch_model-00048-of-00061.bin", "model.layers.48.self_attn.k_proj.weight": "pytorch_model-00048-of-00061.bin", "model.layers.48.self_attn.v_proj.weight": "pytorch_model-00048-of-00061.bin", "model.layers.48.self_attn.o_proj.weight": "pytorch_model-00048-of-00061.bin", "model.layers.48.mlp.gate_proj.weight": "pytorch_model-00048-of-00061.bin", "model.layers.48.mlp.down_proj.weight": "pytorch_model-00048-of-00061.bin", "model.layers.48.mlp.up_proj.weight": "pytorch_model-00048-of-00061.bin", "model.layers.48.self_attn.rotary_emb.inv_freq": "pytorch_model-00048-of-00061.bin", "model.layers.49.input_layernorm.weight": "pytorch_model-00049-of-00061.bin", "model.layers.49.post_attention_layernorm.weight": "pytorch_model-00049-of-00061.bin", "model.layers.49.self_attn.q_proj.weight": "pytorch_model-00049-of-00061.bin", "model.layers.49.self_attn.k_proj.weight": "pytorch_model-00049-of-00061.bin", "model.layers.49.self_attn.v_proj.weight": "pytorch_model-00049-of-00061.bin", "model.layers.49.self_attn.o_proj.weight": "pytorch_model-00049-of-00061.bin", "model.layers.49.mlp.gate_proj.weight": "pytorch_model-00049-of-00061.bin", "model.layers.49.mlp.down_proj.weight": "pytorch_model-00049-of-00061.bin", "model.layers.49.mlp.up_proj.weight": "pytorch_model-00049-of-00061.bin", "model.layers.49.self_attn.rotary_emb.inv_freq": "pytorch_model-00049-of-00061.bin", "model.layers.50.input_layernorm.weight": "pytorch_model-00050-of-00061.bin", "model.layers.50.post_attention_layernorm.weight": "pytorch_model-00050-of-00061.bin", "model.layers.50.self_attn.q_proj.weight": "pytorch_model-00050-of-00061.bin", "model.layers.50.self_attn.k_proj.weight": "pytorch_model-00050-of-00061.bin", "model.layers.50.self_attn.v_proj.weight": "pytorch_model-00050-of-00061.bin", "model.layers.50.self_attn.o_proj.weight": "pytorch_model-00050-of-00061.bin", "model.layers.50.mlp.gate_proj.weight": "pytorch_model-00050-of-00061.bin", "model.layers.50.mlp.down_proj.weight": "pytorch_model-00050-of-00061.bin", "model.layers.50.mlp.up_proj.weight": "pytorch_model-00050-of-00061.bin", "model.layers.50.self_attn.rotary_emb.inv_freq": "pytorch_model-00050-of-00061.bin", "model.layers.51.input_layernorm.weight": "pytorch_model-00051-of-00061.bin", "model.layers.51.post_attention_layernorm.weight": "pytorch_model-00051-of-00061.bin", "model.layers.51.self_attn.q_proj.weight": "pytorch_model-00051-of-00061.bin", "model.layers.51.self_attn.k_proj.weight": "pytorch_model-00051-of-00061.bin", "model.layers.51.self_attn.v_proj.weight": "pytorch_model-00051-of-00061.bin", "model.layers.51.self_attn.o_proj.weight": "pytorch_model-00051-of-00061.bin", "model.layers.51.mlp.gate_proj.weight": "pytorch_model-00051-of-00061.bin", "model.layers.51.mlp.down_proj.weight": "pytorch_model-00051-of-00061.bin", "model.layers.51.mlp.up_proj.weight": "pytorch_model-00051-of-00061.bin", "model.layers.51.self_attn.rotary_emb.inv_freq": "pytorch_model-00051-of-00061.bin", "model.layers.52.input_layernorm.weight": "pytorch_model-00052-of-00061.bin", "model.layers.52.post_attention_layernorm.weight": "pytorch_model-00052-of-00061.bin", "model.layers.52.self_attn.q_proj.weight": "pytorch_model-00052-of-00061.bin", "model.layers.52.self_attn.k_proj.weight": "pytorch_model-00052-of-00061.bin", "model.layers.52.self_attn.v_proj.weight": "pytorch_model-00052-of-00061.bin", "model.layers.52.self_attn.o_proj.weight": "pytorch_model-00052-of-00061.bin", "model.layers.52.mlp.gate_proj.weight": "pytorch_model-00052-of-00061.bin", "model.layers.52.mlp.down_proj.weight": "pytorch_model-00052-of-00061.bin", "model.layers.52.mlp.up_proj.weight": "pytorch_model-00052-of-00061.bin", "model.layers.52.self_attn.rotary_emb.inv_freq": "pytorch_model-00052-of-00061.bin", "model.layers.53.input_layernorm.weight": "pytorch_model-00053-of-00061.bin", "model.layers.53.post_attention_layernorm.weight": "pytorch_model-00053-of-00061.bin", "model.layers.53.self_attn.q_proj.weight": "pytorch_model-00053-of-00061.bin", "model.layers.53.self_attn.k_proj.weight": "pytorch_model-00053-of-00061.bin", "model.layers.53.self_attn.v_proj.weight": "pytorch_model-00053-of-00061.bin", "model.layers.53.self_attn.o_proj.weight": "pytorch_model-00053-of-00061.bin", "model.layers.53.mlp.gate_proj.weight": "pytorch_model-00053-of-00061.bin", "model.layers.53.mlp.down_proj.weight": "pytorch_model-00053-of-00061.bin", "model.layers.53.mlp.up_proj.weight": "pytorch_model-00053-of-00061.bin", "model.layers.53.self_attn.rotary_emb.inv_freq": "pytorch_model-00053-of-00061.bin", "model.layers.54.input_layernorm.weight": "pytorch_model-00054-of-00061.bin", "model.layers.54.post_attention_layernorm.weight": "pytorch_model-00054-of-00061.bin", "model.layers.54.self_attn.q_proj.weight": "pytorch_model-00054-of-00061.bin", "model.layers.54.self_attn.k_proj.weight": "pytorch_model-00054-of-00061.bin", "model.layers.54.self_attn.v_proj.weight": "pytorch_model-00054-of-00061.bin", "model.layers.54.self_attn.o_proj.weight": "pytorch_model-00054-of-00061.bin", "model.layers.54.mlp.gate_proj.weight": "pytorch_model-00054-of-00061.bin", "model.layers.54.mlp.down_proj.weight": "pytorch_model-00054-of-00061.bin", "model.layers.54.mlp.up_proj.weight": "pytorch_model-00054-of-00061.bin", "model.layers.54.self_attn.rotary_emb.inv_freq": "pytorch_model-00054-of-00061.bin", "model.layers.55.input_layernorm.weight": "pytorch_model-00055-of-00061.bin", "model.layers.55.post_attention_layernorm.weight": "pytorch_model-00055-of-00061.bin", "model.layers.55.self_attn.q_proj.weight": "pytorch_model-00055-of-00061.bin", "model.layers.55.self_attn.k_proj.weight": "pytorch_model-00055-of-00061.bin", "model.layers.55.self_attn.v_proj.weight": "pytorch_model-00055-of-00061.bin", "model.layers.55.self_attn.o_proj.weight": "pytorch_model-00055-of-00061.bin", "model.layers.55.mlp.gate_proj.weight": "pytorch_model-00055-of-00061.bin", "model.layers.55.mlp.down_proj.weight": "pytorch_model-00055-of-00061.bin", "model.layers.55.mlp.up_proj.weight": "pytorch_model-00055-of-00061.bin", "model.layers.55.self_attn.rotary_emb.inv_freq": "pytorch_model-00055-of-00061.bin", "model.layers.56.input_layernorm.weight": "pytorch_model-00056-of-00061.bin", "model.layers.56.post_attention_layernorm.weight": "pytorch_model-00056-of-00061.bin", "model.layers.56.self_attn.q_proj.weight": "pytorch_model-00056-of-00061.bin", "model.layers.56.self_attn.k_proj.weight": "pytorch_model-00056-of-00061.bin", "model.layers.56.self_attn.v_proj.weight": "pytorch_model-00056-of-00061.bin", "model.layers.56.self_attn.o_proj.weight": "pytorch_model-00056-of-00061.bin", "model.layers.56.mlp.gate_proj.weight": "pytorch_model-00056-of-00061.bin", "model.layers.56.mlp.down_proj.weight": "pytorch_model-00056-of-00061.bin", "model.layers.56.mlp.up_proj.weight": "pytorch_model-00056-of-00061.bin", "model.layers.56.self_attn.rotary_emb.inv_freq": "pytorch_model-00056-of-00061.bin", "model.layers.57.input_layernorm.weight": "pytorch_model-00057-of-00061.bin", "model.layers.57.post_attention_layernorm.weight": "pytorch_model-00057-of-00061.bin", "model.layers.57.self_attn.q_proj.weight": "pytorch_model-00057-of-00061.bin", "model.layers.57.self_attn.k_proj.weight": "pytorch_model-00057-of-00061.bin", "model.layers.57.self_attn.v_proj.weight": "pytorch_model-00057-of-00061.bin", "model.layers.57.self_attn.o_proj.weight": "pytorch_model-00057-of-00061.bin", "model.layers.57.mlp.gate_proj.weight": "pytorch_model-00057-of-00061.bin", "model.layers.57.mlp.down_proj.weight": "pytorch_model-00057-of-00061.bin", "model.layers.57.mlp.up_proj.weight": "pytorch_model-00057-of-00061.bin", "model.layers.57.self_attn.rotary_emb.inv_freq": "pytorch_model-00057-of-00061.bin", "model.layers.58.input_layernorm.weight": "pytorch_model-00058-of-00061.bin", "model.layers.58.post_attention_layernorm.weight": "pytorch_model-00058-of-00061.bin", "model.layers.58.self_attn.q_proj.weight": "pytorch_model-00058-of-00061.bin", "model.layers.58.self_attn.k_proj.weight": "pytorch_model-00058-of-00061.bin", "model.layers.58.self_attn.v_proj.weight": "pytorch_model-00058-of-00061.bin", "model.layers.58.self_attn.o_proj.weight": "pytorch_model-00058-of-00061.bin", "model.layers.58.mlp.gate_proj.weight": "pytorch_model-00058-of-00061.bin", "model.layers.58.mlp.down_proj.weight": "pytorch_model-00058-of-00061.bin", "model.layers.58.mlp.up_proj.weight": "pytorch_model-00058-of-00061.bin", "model.layers.58.self_attn.rotary_emb.inv_freq": "pytorch_model-00058-of-00061.bin", "model.layers.59.input_layernorm.weight": "pytorch_model-00059-of-00061.bin", "model.layers.59.post_attention_layernorm.weight": "pytorch_model-00059-of-00061.bin", "model.layers.59.self_attn.q_proj.weight": "pytorch_model-00059-of-00061.bin", "model.layers.59.self_attn.k_proj.weight": "pytorch_model-00059-of-00061.bin", "model.layers.59.self_attn.v_proj.weight": "pytorch_model-00059-of-00061.bin", "model.layers.59.self_attn.o_proj.weight": "pytorch_model-00059-of-00061.bin", "model.layers.59.mlp.gate_proj.weight": "pytorch_model-00059-of-00061.bin", "model.layers.59.mlp.down_proj.weight": "pytorch_model-00059-of-00061.bin", "model.layers.59.mlp.up_proj.weight": "pytorch_model-00059-of-00061.bin", "model.layers.59.self_attn.rotary_emb.inv_freq": "pytorch_model-00059-of-00061.bin", "model.norm.weight": "pytorch_model-00060-of-00061.bin", "model.embed_tokens.weight": "pytorch_model-00060-of-00061.bin", "lm_head.weight": "pytorch_model-00060-of-00061.bin"}, "metadata": {"total_size": 65057894912}} \ No newline at end of file +{"weight_map": {"model.layers.0.input_layernorm.weight": "pytorch_model-00001-of-00061.bin", "model.layers.0.post_attention_layernorm.weight": "pytorch_model-00001-of-00061.bin", "model.layers.0.self_attn.q_proj.weight": "pytorch_model-00001-of-00061.bin", "model.layers.0.self_attn.k_proj.weight": "pytorch_model-00001-of-00061.bin", "model.layers.0.self_attn.v_proj.weight": "pytorch_model-00001-of-00061.bin", "model.layers.0.self_attn.o_proj.weight": "pytorch_model-00001-of-00061.bin", "model.layers.0.mlp.gate_proj.weight": "pytorch_model-00001-of-00061.bin", "model.layers.0.mlp.down_proj.weight": "pytorch_model-00001-of-00061.bin", "model.layers.0.mlp.up_proj.weight": "pytorch_model-00001-of-00061.bin", "model.layers.0.self_attn.rotary_emb.inv_freq": "pytorch_model-00001-of-00061.bin", "model.layers.1.input_layernorm.weight": "pytorch_model-00002-of-00061.bin", "model.layers.1.post_attention_layernorm.weight": "pytorch_model-00002-of-00061.bin", "model.layers.1.self_attn.q_proj.weight": "pytorch_model-00002-of-00061.bin", "model.layers.1.self_attn.k_proj.weight": "pytorch_model-00002-of-00061.bin", "model.layers.1.self_attn.v_proj.weight": "pytorch_model-00002-of-00061.bin", "model.layers.1.self_attn.o_proj.weight": "pytorch_model-00002-of-00061.bin", "model.layers.1.mlp.gate_proj.weight": "pytorch_model-00002-of-00061.bin", "model.layers.1.mlp.down_proj.weight": "pytorch_model-00002-of-00061.bin", "model.layers.1.mlp.up_proj.weight": "pytorch_model-00002-of-00061.bin", "model.layers.1.self_attn.rotary_emb.inv_freq": "pytorch_model-00002-of-00061.bin", "model.layers.2.input_layernorm.weight": "pytorch_model-00003-of-00061.bin", "model.layers.2.post_attention_layernorm.weight": "pytorch_model-00003-of-00061.bin", "model.layers.2.self_attn.q_proj.weight": "pytorch_model-00003-of-00061.bin", "model.layers.2.self_attn.k_proj.weight": "pytorch_model-00003-of-00061.bin", "model.layers.2.self_attn.v_proj.weight": "pytorch_model-00003-of-00061.bin", "model.layers.2.self_attn.o_proj.weight": "pytorch_model-00003-of-00061.bin", "model.layers.2.mlp.gate_proj.weight": "pytorch_model-00003-of-00061.bin", "model.layers.2.mlp.down_proj.weight": "pytorch_model-00003-of-00061.bin", "model.layers.2.mlp.up_proj.weight": "pytorch_model-00003-of-00061.bin", "model.layers.2.self_attn.rotary_emb.inv_freq": "pytorch_model-00003-of-00061.bin", "model.layers.3.input_layernorm.weight": "pytorch_model-00004-of-00061.bin", "model.layers.3.post_attention_layernorm.weight": "pytorch_model-00004-of-00061.bin", "model.layers.3.self_attn.q_proj.weight": "pytorch_model-00004-of-00061.bin", "model.layers.3.self_attn.k_proj.weight": "pytorch_model-00004-of-00061.bin", "model.layers.3.self_attn.v_proj.weight": "pytorch_model-00004-of-00061.bin", "model.layers.3.self_attn.o_proj.weight": "pytorch_model-00004-of-00061.bin", "model.layers.3.mlp.gate_proj.weight": "pytorch_model-00004-of-00061.bin", "model.layers.3.mlp.down_proj.weight": "pytorch_model-00004-of-00061.bin", "model.layers.3.mlp.up_proj.weight": "pytorch_model-00004-of-00061.bin", "model.layers.3.self_attn.rotary_emb.inv_freq": "pytorch_model-00004-of-00061.bin", "model.layers.4.input_layernorm.weight": "pytorch_model-00005-of-00061.bin", "model.layers.4.post_attention_layernorm.weight": "pytorch_model-00005-of-00061.bin", "model.layers.4.self_attn.q_proj.weight": "pytorch_model-00005-of-00061.bin", "model.layers.4.self_attn.k_proj.weight": "pytorch_model-00005-of-00061.bin", "model.layers.4.self_attn.v_proj.weight": "pytorch_model-00005-of-00061.bin", "model.layers.4.self_attn.o_proj.weight": "pytorch_model-00005-of-00061.bin", "model.layers.4.mlp.gate_proj.weight": "pytorch_model-00005-of-00061.bin", "model.layers.4.mlp.down_proj.weight": "pytorch_model-00005-of-00061.bin", "model.layers.4.mlp.up_proj.weight": "pytorch_model-00005-of-00061.bin", "model.layers.4.self_attn.rotary_emb.inv_freq": "pytorch_model-00005-of-00061.bin", "model.layers.5.input_layernorm.weight": "pytorch_model-00006-of-00061.bin", "model.layers.5.post_attention_layernorm.weight": "pytorch_model-00006-of-00061.bin", "model.layers.5.self_attn.q_proj.weight": "pytorch_model-00006-of-00061.bin", "model.layers.5.self_attn.k_proj.weight": "pytorch_model-00006-of-00061.bin", "model.layers.5.self_attn.v_proj.weight": "pytorch_model-00006-of-00061.bin", "model.layers.5.self_attn.o_proj.weight": "pytorch_model-00006-of-00061.bin", "model.layers.5.mlp.gate_proj.weight": "pytorch_model-00006-of-00061.bin", "model.layers.5.mlp.down_proj.weight": "pytorch_model-00006-of-00061.bin", "model.layers.5.mlp.up_proj.weight": "pytorch_model-00006-of-00061.bin", "model.layers.5.self_attn.rotary_emb.inv_freq": "pytorch_model-00006-of-00061.bin", "model.layers.6.input_layernorm.weight": "pytorch_model-00007-of-00061.bin", "model.layers.6.post_attention_layernorm.weight": "pytorch_model-00007-of-00061.bin", "model.layers.6.self_attn.q_proj.weight": "pytorch_model-00007-of-00061.bin", "model.layers.6.self_attn.k_proj.weight": "pytorch_model-00007-of-00061.bin", "model.layers.6.self_attn.v_proj.weight": "pytorch_model-00007-of-00061.bin", "model.layers.6.self_attn.o_proj.weight": "pytorch_model-00007-of-00061.bin", "model.layers.6.mlp.gate_proj.weight": "pytorch_model-00007-of-00061.bin", "model.layers.6.mlp.down_proj.weight": "pytorch_model-00007-of-00061.bin", "model.layers.6.mlp.up_proj.weight": "pytorch_model-00007-of-00061.bin", "model.layers.6.self_attn.rotary_emb.inv_freq": "pytorch_model-00007-of-00061.bin", "model.layers.7.input_layernorm.weight": "pytorch_model-00008-of-00061.bin", "model.layers.7.post_attention_layernorm.weight": "pytorch_model-00008-of-00061.bin", "model.layers.7.self_attn.q_proj.weight": "pytorch_model-00008-of-00061.bin", "model.layers.7.self_attn.k_proj.weight": "pytorch_model-00008-of-00061.bin", "model.layers.7.self_attn.v_proj.weight": "pytorch_model-00008-of-00061.bin", "model.layers.7.self_attn.o_proj.weight": "pytorch_model-00008-of-00061.bin", "model.layers.7.mlp.gate_proj.weight": "pytorch_model-00008-of-00061.bin", "model.layers.7.mlp.down_proj.weight": "pytorch_model-00008-of-00061.bin", "model.layers.7.mlp.up_proj.weight": "pytorch_model-00008-of-00061.bin", "model.layers.7.self_attn.rotary_emb.inv_freq": "pytorch_model-00008-of-00061.bin", "model.layers.8.input_layernorm.weight": "pytorch_model-00009-of-00061.bin", "model.layers.8.post_attention_layernorm.weight": "pytorch_model-00009-of-00061.bin", "model.layers.8.self_attn.q_proj.weight": "pytorch_model-00009-of-00061.bin", "model.layers.8.self_attn.k_proj.weight": "pytorch_model-00009-of-00061.bin", "model.layers.8.self_attn.v_proj.weight": "pytorch_model-00009-of-00061.bin", "model.layers.8.self_attn.o_proj.weight": "pytorch_model-00009-of-00061.bin", "model.layers.8.mlp.gate_proj.weight": "pytorch_model-00009-of-00061.bin", "model.layers.8.mlp.down_proj.weight": "pytorch_model-00009-of-00061.bin", "model.layers.8.mlp.up_proj.weight": "pytorch_model-00009-of-00061.bin", "model.layers.8.self_attn.rotary_emb.inv_freq": "pytorch_model-00009-of-00061.bin", "model.layers.9.input_layernorm.weight": "pytorch_model-00010-of-00061.bin", "model.layers.9.post_attention_layernorm.weight": "pytorch_model-00010-of-00061.bin", "model.layers.9.self_attn.q_proj.weight": "pytorch_model-00010-of-00061.bin", "model.layers.9.self_attn.k_proj.weight": "pytorch_model-00010-of-00061.bin", "model.layers.9.self_attn.v_proj.weight": "pytorch_model-00010-of-00061.bin", "model.layers.9.self_attn.o_proj.weight": "pytorch_model-00010-of-00061.bin", "model.layers.9.mlp.gate_proj.weight": "pytorch_model-00010-of-00061.bin", "model.layers.9.mlp.down_proj.weight": "pytorch_model-00010-of-00061.bin", "model.layers.9.mlp.up_proj.weight": "pytorch_model-00010-of-00061.bin", "model.layers.9.self_attn.rotary_emb.inv_freq": "pytorch_model-00010-of-00061.bin", "model.layers.10.input_layernorm.weight": "pytorch_model-00011-of-00061.bin", "model.layers.10.post_attention_layernorm.weight": "pytorch_model-00011-of-00061.bin", "model.layers.10.self_attn.q_proj.weight": "pytorch_model-00011-of-00061.bin", "model.layers.10.self_attn.k_proj.weight": "pytorch_model-00011-of-00061.bin", "model.layers.10.self_attn.v_proj.weight": "pytorch_model-00011-of-00061.bin", "model.layers.10.self_attn.o_proj.weight": "pytorch_model-00011-of-00061.bin", "model.layers.10.mlp.gate_proj.weight": "pytorch_model-00011-of-00061.bin", "model.layers.10.mlp.down_proj.weight": "pytorch_model-00011-of-00061.bin", "model.layers.10.mlp.up_proj.weight": "pytorch_model-00011-of-00061.bin", "model.layers.10.self_attn.rotary_emb.inv_freq": "pytorch_model-00011-of-00061.bin", "model.layers.11.input_layernorm.weight": "pytorch_model-00012-of-00061.bin", "model.layers.11.post_attention_layernorm.weight": "pytorch_model-00012-of-00061.bin", "model.layers.11.self_attn.q_proj.weight": "pytorch_model-00012-of-00061.bin", "model.layers.11.self_attn.k_proj.weight": "pytorch_model-00012-of-00061.bin", "model.layers.11.self_attn.v_proj.weight": "pytorch_model-00012-of-00061.bin", "model.layers.11.self_attn.o_proj.weight": "pytorch_model-00012-of-00061.bin", "model.layers.11.mlp.gate_proj.weight": "pytorch_model-00012-of-00061.bin", "model.layers.11.mlp.down_proj.weight": "pytorch_model-00012-of-00061.bin", "model.layers.11.mlp.up_proj.weight": "pytorch_model-00012-of-00061.bin", "model.layers.11.self_attn.rotary_emb.inv_freq": "pytorch_model-00012-of-00061.bin", "model.layers.12.input_layernorm.weight": "pytorch_model-00013-of-00061.bin", "model.layers.12.post_attention_layernorm.weight": "pytorch_model-00013-of-00061.bin", "model.layers.12.self_attn.q_proj.weight": "pytorch_model-00013-of-00061.bin", "model.layers.12.self_attn.k_proj.weight": "pytorch_model-00013-of-00061.bin", "model.layers.12.self_attn.v_proj.weight": "pytorch_model-00013-of-00061.bin", "model.layers.12.self_attn.o_proj.weight": "pytorch_model-00013-of-00061.bin", "model.layers.12.mlp.gate_proj.weight": "pytorch_model-00013-of-00061.bin", "model.layers.12.mlp.down_proj.weight": "pytorch_model-00013-of-00061.bin", "model.layers.12.mlp.up_proj.weight": "pytorch_model-00013-of-00061.bin", "model.layers.12.self_attn.rotary_emb.inv_freq": "pytorch_model-00013-of-00061.bin", "model.layers.13.input_layernorm.weight": "pytorch_model-00014-of-00061.bin", "model.layers.13.post_attention_layernorm.weight": "pytorch_model-00014-of-00061.bin", "model.layers.13.self_attn.q_proj.weight": "pytorch_model-00014-of-00061.bin", "model.layers.13.self_attn.k_proj.weight": "pytorch_model-00014-of-00061.bin", "model.layers.13.self_attn.v_proj.weight": "pytorch_model-00014-of-00061.bin", "model.layers.13.self_attn.o_proj.weight": "pytorch_model-00014-of-00061.bin", "model.layers.13.mlp.gate_proj.weight": "pytorch_model-00014-of-00061.bin", "model.layers.13.mlp.down_proj.weight": "pytorch_model-00014-of-00061.bin", "model.layers.13.mlp.up_proj.weight": "pytorch_model-00014-of-00061.bin", "model.layers.13.self_attn.rotary_emb.inv_freq": "pytorch_model-00014-of-00061.bin", "model.layers.14.input_layernorm.weight": "pytorch_model-00015-of-00061.bin", "model.layers.14.post_attention_layernorm.weight": "pytorch_model-00015-of-00061.bin", "model.layers.14.self_attn.q_proj.weight": "pytorch_model-00015-of-00061.bin", "model.layers.14.self_attn.k_proj.weight": "pytorch_model-00015-of-00061.bin", "model.layers.14.self_attn.v_proj.weight": "pytorch_model-00015-of-00061.bin", "model.layers.14.self_attn.o_proj.weight": "pytorch_model-00015-of-00061.bin", "model.layers.14.mlp.gate_proj.weight": "pytorch_model-00015-of-00061.bin", "model.layers.14.mlp.down_proj.weight": "pytorch_model-00015-of-00061.bin", "model.layers.14.mlp.up_proj.weight": "pytorch_model-00015-of-00061.bin", "model.layers.14.self_attn.rotary_emb.inv_freq": "pytorch_model-00015-of-00061.bin", "model.layers.15.input_layernorm.weight": "pytorch_model-00016-of-00061.bin", "model.layers.15.post_attention_layernorm.weight": "pytorch_model-00016-of-00061.bin", "model.layers.15.self_attn.q_proj.weight": "pytorch_model-00016-of-00061.bin", "model.layers.15.self_attn.k_proj.weight": "pytorch_model-00016-of-00061.bin", "model.layers.15.self_attn.v_proj.weight": "pytorch_model-00016-of-00061.bin", "model.layers.15.self_attn.o_proj.weight": "pytorch_model-00016-of-00061.bin", "model.layers.15.mlp.gate_proj.weight": "pytorch_model-00016-of-00061.bin", "model.layers.15.mlp.down_proj.weight": "pytorch_model-00016-of-00061.bin", "model.layers.15.mlp.up_proj.weight": "pytorch_model-00016-of-00061.bin", "model.layers.15.self_attn.rotary_emb.inv_freq": "pytorch_model-00016-of-00061.bin", "model.layers.16.input_layernorm.weight": "pytorch_model-00017-of-00061.bin", "model.layers.16.post_attention_layernorm.weight": "pytorch_model-00017-of-00061.bin", "model.layers.16.self_attn.q_proj.weight": "pytorch_model-00017-of-00061.bin", "model.layers.16.self_attn.k_proj.weight": "pytorch_model-00017-of-00061.bin", "model.layers.16.self_attn.v_proj.weight": "pytorch_model-00017-of-00061.bin", "model.layers.16.self_attn.o_proj.weight": "pytorch_model-00017-of-00061.bin", "model.layers.16.mlp.gate_proj.weight": "pytorch_model-00017-of-00061.bin", "model.layers.16.mlp.down_proj.weight": "pytorch_model-00017-of-00061.bin", "model.layers.16.mlp.up_proj.weight": "pytorch_model-00017-of-00061.bin", "model.layers.16.self_attn.rotary_emb.inv_freq": "pytorch_model-00017-of-00061.bin", "model.layers.17.input_layernorm.weight": "pytorch_model-00018-of-00061.bin", "model.layers.17.post_attention_layernorm.weight": "pytorch_model-00018-of-00061.bin", "model.layers.17.self_attn.q_proj.weight": "pytorch_model-00018-of-00061.bin", "model.layers.17.self_attn.k_proj.weight": "pytorch_model-00018-of-00061.bin", "model.layers.17.self_attn.v_proj.weight": "pytorch_model-00018-of-00061.bin", "model.layers.17.self_attn.o_proj.weight": "pytorch_model-00018-of-00061.bin", "model.layers.17.mlp.gate_proj.weight": "pytorch_model-00018-of-00061.bin", "model.layers.17.mlp.down_proj.weight": "pytorch_model-00018-of-00061.bin", "model.layers.17.mlp.up_proj.weight": "pytorch_model-00018-of-00061.bin", "model.layers.17.self_attn.rotary_emb.inv_freq": "pytorch_model-00018-of-00061.bin", "model.layers.18.input_layernorm.weight": "pytorch_model-00019-of-00061.bin", "model.layers.18.post_attention_layernorm.weight": "pytorch_model-00019-of-00061.bin", "model.layers.18.self_attn.q_proj.weight": "pytorch_model-00019-of-00061.bin", "model.layers.18.self_attn.k_proj.weight": "pytorch_model-00019-of-00061.bin", "model.layers.18.self_attn.v_proj.weight": "pytorch_model-00019-of-00061.bin", "model.layers.18.self_attn.o_proj.weight": "pytorch_model-00019-of-00061.bin", "model.layers.18.mlp.gate_proj.weight": "pytorch_model-00019-of-00061.bin", "model.layers.18.mlp.down_proj.weight": "pytorch_model-00019-of-00061.bin", "model.layers.18.mlp.up_proj.weight": "pytorch_model-00019-of-00061.bin", "model.layers.18.self_attn.rotary_emb.inv_freq": "pytorch_model-00019-of-00061.bin", "model.layers.19.input_layernorm.weight": "pytorch_model-00020-of-00061.bin", "model.layers.19.post_attention_layernorm.weight": "pytorch_model-00020-of-00061.bin", "model.layers.19.self_attn.q_proj.weight": "pytorch_model-00020-of-00061.bin", "model.layers.19.self_attn.k_proj.weight": "pytorch_model-00020-of-00061.bin", "model.layers.19.self_attn.v_proj.weight": "pytorch_model-00020-of-00061.bin", "model.layers.19.self_attn.o_proj.weight": "pytorch_model-00020-of-00061.bin", "model.layers.19.mlp.gate_proj.weight": "pytorch_model-00020-of-00061.bin", "model.layers.19.mlp.down_proj.weight": "pytorch_model-00020-of-00061.bin", "model.layers.19.mlp.up_proj.weight": "pytorch_model-00020-of-00061.bin", "model.layers.19.self_attn.rotary_emb.inv_freq": "pytorch_model-00020-of-00061.bin", "model.layers.20.input_layernorm.weight": "pytorch_model-00021-of-00061.bin", "model.layers.20.post_attention_layernorm.weight": "pytorch_model-00021-of-00061.bin", "model.layers.20.self_attn.q_proj.weight": "pytorch_model-00021-of-00061.bin", "model.layers.20.self_attn.k_proj.weight": "pytorch_model-00021-of-00061.bin", "model.layers.20.self_attn.v_proj.weight": "pytorch_model-00021-of-00061.bin", "model.layers.20.self_attn.o_proj.weight": "pytorch_model-00021-of-00061.bin", "model.layers.20.mlp.gate_proj.weight": "pytorch_model-00021-of-00061.bin", "model.layers.20.mlp.down_proj.weight": "pytorch_model-00021-of-00061.bin", "model.layers.20.mlp.up_proj.weight": "pytorch_model-00021-of-00061.bin", "model.layers.20.self_attn.rotary_emb.inv_freq": "pytorch_model-00021-of-00061.bin", "model.layers.21.input_layernorm.weight": "pytorch_model-00022-of-00061.bin", "model.layers.21.post_attention_layernorm.weight": "pytorch_model-00022-of-00061.bin", "model.layers.21.self_attn.q_proj.weight": "pytorch_model-00022-of-00061.bin", "model.layers.21.self_attn.k_proj.weight": "pytorch_model-00022-of-00061.bin", "model.layers.21.self_attn.v_proj.weight": "pytorch_model-00022-of-00061.bin", "model.layers.21.self_attn.o_proj.weight": "pytorch_model-00022-of-00061.bin", "model.layers.21.mlp.gate_proj.weight": "pytorch_model-00022-of-00061.bin", "model.layers.21.mlp.down_proj.weight": "pytorch_model-00022-of-00061.bin", "model.layers.21.mlp.up_proj.weight": "pytorch_model-00022-of-00061.bin", "model.layers.21.self_attn.rotary_emb.inv_freq": "pytorch_model-00022-of-00061.bin", "model.layers.22.input_layernorm.weight": "pytorch_model-00023-of-00061.bin", "model.layers.22.post_attention_layernorm.weight": "pytorch_model-00023-of-00061.bin", "model.layers.22.self_attn.q_proj.weight": "pytorch_model-00023-of-00061.bin", "model.layers.22.self_attn.k_proj.weight": "pytorch_model-00023-of-00061.bin", "model.layers.22.self_attn.v_proj.weight": "pytorch_model-00023-of-00061.bin", "model.layers.22.self_attn.o_proj.weight": "pytorch_model-00023-of-00061.bin", "model.layers.22.mlp.gate_proj.weight": "pytorch_model-00023-of-00061.bin", "model.layers.22.mlp.down_proj.weight": "pytorch_model-00023-of-00061.bin", "model.layers.22.mlp.up_proj.weight": "pytorch_model-00023-of-00061.bin", "model.layers.22.self_attn.rotary_emb.inv_freq": "pytorch_model-00023-of-00061.bin", "model.layers.23.input_layernorm.weight": "pytorch_model-00024-of-00061.bin", "model.layers.23.post_attention_layernorm.weight": "pytorch_model-00024-of-00061.bin", "model.layers.23.self_attn.q_proj.weight": "pytorch_model-00024-of-00061.bin", "model.layers.23.self_attn.k_proj.weight": "pytorch_model-00024-of-00061.bin", "model.layers.23.self_attn.v_proj.weight": "pytorch_model-00024-of-00061.bin", "model.layers.23.self_attn.o_proj.weight": "pytorch_model-00024-of-00061.bin", "model.layers.23.mlp.gate_proj.weight": "pytorch_model-00024-of-00061.bin", "model.layers.23.mlp.down_proj.weight": "pytorch_model-00024-of-00061.bin", "model.layers.23.mlp.up_proj.weight": "pytorch_model-00024-of-00061.bin", "model.layers.23.self_attn.rotary_emb.inv_freq": "pytorch_model-00024-of-00061.bin", "model.layers.24.input_layernorm.weight": "pytorch_model-00025-of-00061.bin", "model.layers.24.post_attention_layernorm.weight": "pytorch_model-00025-of-00061.bin", "model.layers.24.self_attn.q_proj.weight": "pytorch_model-00025-of-00061.bin", "model.layers.24.self_attn.k_proj.weight": "pytorch_model-00025-of-00061.bin", "model.layers.24.self_attn.v_proj.weight": "pytorch_model-00025-of-00061.bin", "model.layers.24.self_attn.o_proj.weight": "pytorch_model-00025-of-00061.bin", "model.layers.24.mlp.gate_proj.weight": "pytorch_model-00025-of-00061.bin", "model.layers.24.mlp.down_proj.weight": "pytorch_model-00025-of-00061.bin", "model.layers.24.mlp.up_proj.weight": "pytorch_model-00025-of-00061.bin", "model.layers.24.self_attn.rotary_emb.inv_freq": "pytorch_model-00025-of-00061.bin", "model.layers.25.input_layernorm.weight": "pytorch_model-00026-of-00061.bin", "model.layers.25.post_attention_layernorm.weight": "pytorch_model-00026-of-00061.bin", "model.layers.25.self_attn.q_proj.weight": "pytorch_model-00026-of-00061.bin", "model.layers.25.self_attn.k_proj.weight": "pytorch_model-00026-of-00061.bin", "model.layers.25.self_attn.v_proj.weight": "pytorch_model-00026-of-00061.bin", "model.layers.25.self_attn.o_proj.weight": "pytorch_model-00026-of-00061.bin", "model.layers.25.mlp.gate_proj.weight": "pytorch_model-00026-of-00061.bin", "model.layers.25.mlp.down_proj.weight": "pytorch_model-00026-of-00061.bin", "model.layers.25.mlp.up_proj.weight": "pytorch_model-00026-of-00061.bin", "model.layers.25.self_attn.rotary_emb.inv_freq": "pytorch_model-00026-of-00061.bin", "model.layers.26.input_layernorm.weight": "pytorch_model-00027-of-00061.bin", "model.layers.26.post_attention_layernorm.weight": "pytorch_model-00027-of-00061.bin", "model.layers.26.self_attn.q_proj.weight": "pytorch_model-00027-of-00061.bin", "model.layers.26.self_attn.k_proj.weight": "pytorch_model-00027-of-00061.bin", "model.layers.26.self_attn.v_proj.weight": "pytorch_model-00027-of-00061.bin", "model.layers.26.self_attn.o_proj.weight": "pytorch_model-00027-of-00061.bin", "model.layers.26.mlp.gate_proj.weight": "pytorch_model-00027-of-00061.bin", "model.layers.26.mlp.down_proj.weight": "pytorch_model-00027-of-00061.bin", "model.layers.26.mlp.up_proj.weight": "pytorch_model-00027-of-00061.bin", "model.layers.26.self_attn.rotary_emb.inv_freq": "pytorch_model-00027-of-00061.bin", "model.layers.27.input_layernorm.weight": "pytorch_model-00028-of-00061.bin", "model.layers.27.post_attention_layernorm.weight": "pytorch_model-00028-of-00061.bin", "model.layers.27.self_attn.q_proj.weight": "pytorch_model-00028-of-00061.bin", "model.layers.27.self_attn.k_proj.weight": "pytorch_model-00028-of-00061.bin", "model.layers.27.self_attn.v_proj.weight": "pytorch_model-00028-of-00061.bin", "model.layers.27.self_attn.o_proj.weight": "pytorch_model-00028-of-00061.bin", "model.layers.27.mlp.gate_proj.weight": "pytorch_model-00028-of-00061.bin", "model.layers.27.mlp.down_proj.weight": "pytorch_model-00028-of-00061.bin", "model.layers.27.mlp.up_proj.weight": "pytorch_model-00028-of-00061.bin", "model.layers.27.self_attn.rotary_emb.inv_freq": "pytorch_model-00028-of-00061.bin", "model.layers.28.input_layernorm.weight": "pytorch_model-00029-of-00061.bin", "model.layers.28.post_attention_layernorm.weight": "pytorch_model-00029-of-00061.bin", "model.layers.28.self_attn.q_proj.weight": "pytorch_model-00029-of-00061.bin", "model.layers.28.self_attn.k_proj.weight": "pytorch_model-00029-of-00061.bin", "model.layers.28.self_attn.v_proj.weight": "pytorch_model-00029-of-00061.bin", "model.layers.28.self_attn.o_proj.weight": "pytorch_model-00029-of-00061.bin", "model.layers.28.mlp.gate_proj.weight": "pytorch_model-00029-of-00061.bin", "model.layers.28.mlp.down_proj.weight": "pytorch_model-00029-of-00061.bin", "model.layers.28.mlp.up_proj.weight": "pytorch_model-00029-of-00061.bin", "model.layers.28.self_attn.rotary_emb.inv_freq": "pytorch_model-00029-of-00061.bin", "model.layers.29.input_layernorm.weight": "pytorch_model-00030-of-00061.bin", "model.layers.29.post_attention_layernorm.weight": "pytorch_model-00030-of-00061.bin", "model.layers.29.self_attn.q_proj.weight": "pytorch_model-00030-of-00061.bin", "model.layers.29.self_attn.k_proj.weight": "pytorch_model-00030-of-00061.bin", "model.layers.29.self_attn.v_proj.weight": "pytorch_model-00030-of-00061.bin", "model.layers.29.self_attn.o_proj.weight": "pytorch_model-00030-of-00061.bin", "model.layers.29.mlp.gate_proj.weight": "pytorch_model-00030-of-00061.bin", "model.layers.29.mlp.down_proj.weight": "pytorch_model-00030-of-00061.bin", "model.layers.29.mlp.up_proj.weight": "pytorch_model-00030-of-00061.bin", "model.layers.29.self_attn.rotary_emb.inv_freq": "pytorch_model-00030-of-00061.bin", "model.layers.30.input_layernorm.weight": "pytorch_model-00031-of-00061.bin", "model.layers.30.post_attention_layernorm.weight": "pytorch_model-00031-of-00061.bin", "model.layers.30.self_attn.q_proj.weight": "pytorch_model-00031-of-00061.bin", "model.layers.30.self_attn.k_proj.weight": "pytorch_model-00031-of-00061.bin", "model.layers.30.self_attn.v_proj.weight": "pytorch_model-00031-of-00061.bin", "model.layers.30.self_attn.o_proj.weight": "pytorch_model-00031-of-00061.bin", "model.layers.30.mlp.gate_proj.weight": "pytorch_model-00031-of-00061.bin", "model.layers.30.mlp.down_proj.weight": "pytorch_model-00031-of-00061.bin", "model.layers.30.mlp.up_proj.weight": "pytorch_model-00031-of-00061.bin", "model.layers.30.self_attn.rotary_emb.inv_freq": "pytorch_model-00031-of-00061.bin", "model.layers.31.input_layernorm.weight": "pytorch_model-00032-of-00061.bin", "model.layers.31.post_attention_layernorm.weight": "pytorch_model-00032-of-00061.bin", "model.layers.31.self_attn.q_proj.weight": "pytorch_model-00032-of-00061.bin", "model.layers.31.self_attn.k_proj.weight": "pytorch_model-00032-of-00061.bin", "model.layers.31.self_attn.v_proj.weight": "pytorch_model-00032-of-00061.bin", "model.layers.31.self_attn.o_proj.weight": "pytorch_model-00032-of-00061.bin", "model.layers.31.mlp.gate_proj.weight": "pytorch_model-00032-of-00061.bin", "model.layers.31.mlp.down_proj.weight": "pytorch_model-00032-of-00061.bin", "model.layers.31.mlp.up_proj.weight": "pytorch_model-00032-of-00061.bin", "model.layers.31.self_attn.rotary_emb.inv_freq": "pytorch_model-00032-of-00061.bin", "model.layers.32.input_layernorm.weight": "pytorch_model-00033-of-00061.bin", "model.layers.32.post_attention_layernorm.weight": "pytorch_model-00033-of-00061.bin", "model.layers.32.self_attn.q_proj.weight": "pytorch_model-00033-of-00061.bin", "model.layers.32.self_attn.k_proj.weight": "pytorch_model-00033-of-00061.bin", "model.layers.32.self_attn.v_proj.weight": "pytorch_model-00033-of-00061.bin", "model.layers.32.self_attn.o_proj.weight": "pytorch_model-00033-of-00061.bin", "model.layers.32.mlp.gate_proj.weight": "pytorch_model-00033-of-00061.bin", "model.layers.32.mlp.down_proj.weight": "pytorch_model-00033-of-00061.bin", "model.layers.32.mlp.up_proj.weight": "pytorch_model-00033-of-00061.bin", "model.layers.32.self_attn.rotary_emb.inv_freq": "pytorch_model-00033-of-00061.bin", "model.layers.33.input_layernorm.weight": "pytorch_model-00034-of-00061.bin", "model.layers.33.post_attention_layernorm.weight": "pytorch_model-00034-of-00061.bin", "model.layers.33.self_attn.q_proj.weight": "pytorch_model-00034-of-00061.bin", "model.layers.33.self_attn.k_proj.weight": "pytorch_model-00034-of-00061.bin", "model.layers.33.self_attn.v_proj.weight": "pytorch_model-00034-of-00061.bin", "model.layers.33.self_attn.o_proj.weight": "pytorch_model-00034-of-00061.bin", "model.layers.33.mlp.gate_proj.weight": "pytorch_model-00034-of-00061.bin", "model.layers.33.mlp.down_proj.weight": "pytorch_model-00034-of-00061.bin", "model.layers.33.mlp.up_proj.weight": "pytorch_model-00034-of-00061.bin", "model.layers.33.self_attn.rotary_emb.inv_freq": "pytorch_model-00034-of-00061.bin", "model.layers.34.input_layernorm.weight": "pytorch_model-00035-of-00061.bin", "model.layers.34.post_attention_layernorm.weight": "pytorch_model-00035-of-00061.bin", "model.layers.34.self_attn.q_proj.weight": "pytorch_model-00035-of-00061.bin", "model.layers.34.self_attn.k_proj.weight": "pytorch_model-00035-of-00061.bin", "model.layers.34.self_attn.v_proj.weight": "pytorch_model-00035-of-00061.bin", "model.layers.34.self_attn.o_proj.weight": "pytorch_model-00035-of-00061.bin", "model.layers.34.mlp.gate_proj.weight": "pytorch_model-00035-of-00061.bin", "model.layers.34.mlp.down_proj.weight": "pytorch_model-00035-of-00061.bin", "model.layers.34.mlp.up_proj.weight": "pytorch_model-00035-of-00061.bin", "model.layers.34.self_attn.rotary_emb.inv_freq": "pytorch_model-00035-of-00061.bin", "model.layers.35.input_layernorm.weight": "pytorch_model-00036-of-00061.bin", "model.layers.35.post_attention_layernorm.weight": "pytorch_model-00036-of-00061.bin", "model.layers.35.self_attn.q_proj.weight": "pytorch_model-00036-of-00061.bin", "model.layers.35.self_attn.k_proj.weight": "pytorch_model-00036-of-00061.bin", "model.layers.35.self_attn.v_proj.weight": "pytorch_model-00036-of-00061.bin", "model.layers.35.self_attn.o_proj.weight": "pytorch_model-00036-of-00061.bin", "model.layers.35.mlp.gate_proj.weight": "pytorch_model-00036-of-00061.bin", "model.layers.35.mlp.down_proj.weight": "pytorch_model-00036-of-00061.bin", "model.layers.35.mlp.up_proj.weight": "pytorch_model-00036-of-00061.bin", "model.layers.35.self_attn.rotary_emb.inv_freq": "pytorch_model-00036-of-00061.bin", "model.layers.36.input_layernorm.weight": "pytorch_model-00037-of-00061.bin", "model.layers.36.post_attention_layernorm.weight": "pytorch_model-00037-of-00061.bin", "model.layers.36.self_attn.q_proj.weight": "pytorch_model-00037-of-00061.bin", "model.layers.36.self_attn.k_proj.weight": "pytorch_model-00037-of-00061.bin", "model.layers.36.self_attn.v_proj.weight": "pytorch_model-00037-of-00061.bin", "model.layers.36.self_attn.o_proj.weight": "pytorch_model-00037-of-00061.bin", "model.layers.36.mlp.gate_proj.weight": "pytorch_model-00037-of-00061.bin", "model.layers.36.mlp.down_proj.weight": "pytorch_model-00037-of-00061.bin", "model.layers.36.mlp.up_proj.weight": "pytorch_model-00037-of-00061.bin", "model.layers.36.self_attn.rotary_emb.inv_freq": "pytorch_model-00037-of-00061.bin", "model.layers.37.input_layernorm.weight": "pytorch_model-00038-of-00061.bin", "model.layers.37.post_attention_layernorm.weight": "pytorch_model-00038-of-00061.bin", "model.layers.37.self_attn.q_proj.weight": "pytorch_model-00038-of-00061.bin", "model.layers.37.self_attn.k_proj.weight": "pytorch_model-00038-of-00061.bin", "model.layers.37.self_attn.v_proj.weight": "pytorch_model-00038-of-00061.bin", "model.layers.37.self_attn.o_proj.weight": "pytorch_model-00038-of-00061.bin", "model.layers.37.mlp.gate_proj.weight": "pytorch_model-00038-of-00061.bin", "model.layers.37.mlp.down_proj.weight": "pytorch_model-00038-of-00061.bin", "model.layers.37.mlp.up_proj.weight": "pytorch_model-00038-of-00061.bin", "model.layers.37.self_attn.rotary_emb.inv_freq": "pytorch_model-00038-of-00061.bin", "model.layers.38.input_layernorm.weight": "pytorch_model-00039-of-00061.bin", "model.layers.38.post_attention_layernorm.weight": "pytorch_model-00039-of-00061.bin", "model.layers.38.self_attn.q_proj.weight": "pytorch_model-00039-of-00061.bin", "model.layers.38.self_attn.k_proj.weight": "pytorch_model-00039-of-00061.bin", "model.layers.38.self_attn.v_proj.weight": "pytorch_model-00039-of-00061.bin", "model.layers.38.self_attn.o_proj.weight": "pytorch_model-00039-of-00061.bin", "model.layers.38.mlp.gate_proj.weight": "pytorch_model-00039-of-00061.bin", "model.layers.38.mlp.down_proj.weight": "pytorch_model-00039-of-00061.bin", "model.layers.38.mlp.up_proj.weight": "pytorch_model-00039-of-00061.bin", "model.layers.38.self_attn.rotary_emb.inv_freq": "pytorch_model-00039-of-00061.bin", "model.layers.39.input_layernorm.weight": "pytorch_model-00040-of-00061.bin", "model.layers.39.post_attention_layernorm.weight": "pytorch_model-00040-of-00061.bin", "model.layers.39.self_attn.q_proj.weight": "pytorch_model-00040-of-00061.bin", "model.layers.39.self_attn.k_proj.weight": "pytorch_model-00040-of-00061.bin", "model.layers.39.self_attn.v_proj.weight": "pytorch_model-00040-of-00061.bin", "model.layers.39.self_attn.o_proj.weight": "pytorch_model-00040-of-00061.bin", "model.layers.39.mlp.gate_proj.weight": "pytorch_model-00040-of-00061.bin", "model.layers.39.mlp.down_proj.weight": "pytorch_model-00040-of-00061.bin", "model.layers.39.mlp.up_proj.weight": "pytorch_model-00040-of-00061.bin", "model.layers.39.self_attn.rotary_emb.inv_freq": "pytorch_model-00040-of-00061.bin", "model.layers.40.input_layernorm.weight": "pytorch_model-00041-of-00061.bin", "model.layers.40.post_attention_layernorm.weight": "pytorch_model-00041-of-00061.bin", "model.layers.40.self_attn.q_proj.weight": "pytorch_model-00041-of-00061.bin", "model.layers.40.self_attn.k_proj.weight": "pytorch_model-00041-of-00061.bin", "model.layers.40.self_attn.v_proj.weight": "pytorch_model-00041-of-00061.bin", "model.layers.40.self_attn.o_proj.weight": "pytorch_model-00041-of-00061.bin", "model.layers.40.mlp.gate_proj.weight": "pytorch_model-00041-of-00061.bin", "model.layers.40.mlp.down_proj.weight": "pytorch_model-00041-of-00061.bin", "model.layers.40.mlp.up_proj.weight": "pytorch_model-00041-of-00061.bin", "model.layers.40.self_attn.rotary_emb.inv_freq": "pytorch_model-00041-of-00061.bin", "model.layers.41.input_layernorm.weight": "pytorch_model-00042-of-00061.bin", "model.layers.41.post_attention_layernorm.weight": "pytorch_model-00042-of-00061.bin", "model.layers.41.self_attn.q_proj.weight": "pytorch_model-00042-of-00061.bin", "model.layers.41.self_attn.k_proj.weight": "pytorch_model-00042-of-00061.bin", "model.layers.41.self_attn.v_proj.weight": "pytorch_model-00042-of-00061.bin", "model.layers.41.self_attn.o_proj.weight": "pytorch_model-00042-of-00061.bin", "model.layers.41.mlp.gate_proj.weight": "pytorch_model-00042-of-00061.bin", "model.layers.41.mlp.down_proj.weight": "pytorch_model-00042-of-00061.bin", "model.layers.41.mlp.up_proj.weight": "pytorch_model-00042-of-00061.bin", "model.layers.41.self_attn.rotary_emb.inv_freq": "pytorch_model-00042-of-00061.bin", "model.layers.42.input_layernorm.weight": "pytorch_model-00043-of-00061.bin", "model.layers.42.post_attention_layernorm.weight": "pytorch_model-00043-of-00061.bin", "model.layers.42.self_attn.q_proj.weight": "pytorch_model-00043-of-00061.bin", "model.layers.42.self_attn.k_proj.weight": "pytorch_model-00043-of-00061.bin", "model.layers.42.self_attn.v_proj.weight": "pytorch_model-00043-of-00061.bin", "model.layers.42.self_attn.o_proj.weight": "pytorch_model-00043-of-00061.bin", "model.layers.42.mlp.gate_proj.weight": "pytorch_model-00043-of-00061.bin", "model.layers.42.mlp.down_proj.weight": "pytorch_model-00043-of-00061.bin", "model.layers.42.mlp.up_proj.weight": "pytorch_model-00043-of-00061.bin", "model.layers.42.self_attn.rotary_emb.inv_freq": "pytorch_model-00043-of-00061.bin", "model.layers.43.input_layernorm.weight": "pytorch_model-00044-of-00061.bin", "model.layers.43.post_attention_layernorm.weight": "pytorch_model-00044-of-00061.bin", "model.layers.43.self_attn.q_proj.weight": "pytorch_model-00044-of-00061.bin", "model.layers.43.self_attn.k_proj.weight": "pytorch_model-00044-of-00061.bin", "model.layers.43.self_attn.v_proj.weight": "pytorch_model-00044-of-00061.bin", "model.layers.43.self_attn.o_proj.weight": "pytorch_model-00044-of-00061.bin", "model.layers.43.mlp.gate_proj.weight": "pytorch_model-00044-of-00061.bin", "model.layers.43.mlp.down_proj.weight": "pytorch_model-00044-of-00061.bin", "model.layers.43.mlp.up_proj.weight": "pytorch_model-00044-of-00061.bin", "model.layers.43.self_attn.rotary_emb.inv_freq": "pytorch_model-00044-of-00061.bin", "model.layers.44.input_layernorm.weight": "pytorch_model-00045-of-00061.bin", "model.layers.44.post_attention_layernorm.weight": "pytorch_model-00045-of-00061.bin", "model.layers.44.self_attn.q_proj.weight": "pytorch_model-00045-of-00061.bin", "model.layers.44.self_attn.k_proj.weight": "pytorch_model-00045-of-00061.bin", "model.layers.44.self_attn.v_proj.weight": "pytorch_model-00045-of-00061.bin", "model.layers.44.self_attn.o_proj.weight": "pytorch_model-00045-of-00061.bin", "model.layers.44.mlp.gate_proj.weight": "pytorch_model-00045-of-00061.bin", "model.layers.44.mlp.down_proj.weight": "pytorch_model-00045-of-00061.bin", "model.layers.44.mlp.up_proj.weight": "pytorch_model-00045-of-00061.bin", "model.layers.44.self_attn.rotary_emb.inv_freq": "pytorch_model-00045-of-00061.bin", "model.layers.45.input_layernorm.weight": "pytorch_model-00046-of-00061.bin", "model.layers.45.post_attention_layernorm.weight": "pytorch_model-00046-of-00061.bin", "model.layers.45.self_attn.q_proj.weight": "pytorch_model-00046-of-00061.bin", "model.layers.45.self_attn.k_proj.weight": "pytorch_model-00046-of-00061.bin", "model.layers.45.self_attn.v_proj.weight": "pytorch_model-00046-of-00061.bin", "model.layers.45.self_attn.o_proj.weight": "pytorch_model-00046-of-00061.bin", "model.layers.45.mlp.gate_proj.weight": "pytorch_model-00046-of-00061.bin", "model.layers.45.mlp.down_proj.weight": "pytorch_model-00046-of-00061.bin", "model.layers.45.mlp.up_proj.weight": "pytorch_model-00046-of-00061.bin", "model.layers.45.self_attn.rotary_emb.inv_freq": "pytorch_model-00046-of-00061.bin", "model.layers.46.input_layernorm.weight": "pytorch_model-00047-of-00061.bin", "model.layers.46.post_attention_layernorm.weight": "pytorch_model-00047-of-00061.bin", "model.layers.46.self_attn.q_proj.weight": "pytorch_model-00047-of-00061.bin", "model.layers.46.self_attn.k_proj.weight": "pytorch_model-00047-of-00061.bin", "model.layers.46.self_attn.v_proj.weight": "pytorch_model-00047-of-00061.bin", "model.layers.46.self_attn.o_proj.weight": "pytorch_model-00047-of-00061.bin", "model.layers.46.mlp.gate_proj.weight": "pytorch_model-00047-of-00061.bin", "model.layers.46.mlp.down_proj.weight": "pytorch_model-00047-of-00061.bin", "model.layers.46.mlp.up_proj.weight": "pytorch_model-00047-of-00061.bin", "model.layers.46.self_attn.rotary_emb.inv_freq": "pytorch_model-00047-of-00061.bin", "model.layers.47.input_layernorm.weight": "pytorch_model-00048-of-00061.bin", "model.layers.47.post_attention_layernorm.weight": "pytorch_model-00048-of-00061.bin", "model.layers.47.self_attn.q_proj.weight": "pytorch_model-00048-of-00061.bin", "model.layers.47.self_attn.k_proj.weight": "pytorch_model-00048-of-00061.bin", "model.layers.47.self_attn.v_proj.weight": "pytorch_model-00048-of-00061.bin", "model.layers.47.self_attn.o_proj.weight": "pytorch_model-00048-of-00061.bin", "model.layers.47.mlp.gate_proj.weight": "pytorch_model-00048-of-00061.bin", "model.layers.47.mlp.down_proj.weight": "pytorch_model-00048-of-00061.bin", "model.layers.47.mlp.up_proj.weight": "pytorch_model-00048-of-00061.bin", "model.layers.47.self_attn.rotary_emb.inv_freq": "pytorch_model-00048-of-00061.bin", "model.layers.48.input_layernorm.weight": "pytorch_model-00049-of-00061.bin", "model.layers.48.post_attention_layernorm.weight": "pytorch_model-00049-of-00061.bin", "model.layers.48.self_attn.q_proj.weight": "pytorch_model-00049-of-00061.bin", "model.layers.48.self_attn.k_proj.weight": "pytorch_model-00049-of-00061.bin", "model.layers.48.self_attn.v_proj.weight": "pytorch_model-00049-of-00061.bin", "model.layers.48.self_attn.o_proj.weight": "pytorch_model-00049-of-00061.bin", "model.layers.48.mlp.gate_proj.weight": "pytorch_model-00049-of-00061.bin", "model.layers.48.mlp.down_proj.weight": "pytorch_model-00049-of-00061.bin", "model.layers.48.mlp.up_proj.weight": "pytorch_model-00049-of-00061.bin", "model.layers.48.self_attn.rotary_emb.inv_freq": "pytorch_model-00049-of-00061.bin", "model.layers.49.input_layernorm.weight": "pytorch_model-00050-of-00061.bin", "model.layers.49.post_attention_layernorm.weight": "pytorch_model-00050-of-00061.bin", "model.layers.49.self_attn.q_proj.weight": "pytorch_model-00050-of-00061.bin", "model.layers.49.self_attn.k_proj.weight": "pytorch_model-00050-of-00061.bin", "model.layers.49.self_attn.v_proj.weight": "pytorch_model-00050-of-00061.bin", "model.layers.49.self_attn.o_proj.weight": "pytorch_model-00050-of-00061.bin", "model.layers.49.mlp.gate_proj.weight": "pytorch_model-00050-of-00061.bin", "model.layers.49.mlp.down_proj.weight": "pytorch_model-00050-of-00061.bin", "model.layers.49.mlp.up_proj.weight": "pytorch_model-00050-of-00061.bin", "model.layers.49.self_attn.rotary_emb.inv_freq": "pytorch_model-00050-of-00061.bin", "model.layers.50.input_layernorm.weight": "pytorch_model-00051-of-00061.bin", "model.layers.50.post_attention_layernorm.weight": "pytorch_model-00051-of-00061.bin", "model.layers.50.self_attn.q_proj.weight": "pytorch_model-00051-of-00061.bin", "model.layers.50.self_attn.k_proj.weight": "pytorch_model-00051-of-00061.bin", "model.layers.50.self_attn.v_proj.weight": "pytorch_model-00051-of-00061.bin", "model.layers.50.self_attn.o_proj.weight": "pytorch_model-00051-of-00061.bin", "model.layers.50.mlp.gate_proj.weight": "pytorch_model-00051-of-00061.bin", "model.layers.50.mlp.down_proj.weight": "pytorch_model-00051-of-00061.bin", "model.layers.50.mlp.up_proj.weight": "pytorch_model-00051-of-00061.bin", "model.layers.50.self_attn.rotary_emb.inv_freq": "pytorch_model-00051-of-00061.bin", "model.layers.51.input_layernorm.weight": "pytorch_model-00052-of-00061.bin", "model.layers.51.post_attention_layernorm.weight": "pytorch_model-00052-of-00061.bin", "model.layers.51.self_attn.q_proj.weight": "pytorch_model-00052-of-00061.bin", "model.layers.51.self_attn.k_proj.weight": "pytorch_model-00052-of-00061.bin", "model.layers.51.self_attn.v_proj.weight": "pytorch_model-00052-of-00061.bin", "model.layers.51.self_attn.o_proj.weight": "pytorch_model-00052-of-00061.bin", "model.layers.51.mlp.gate_proj.weight": "pytorch_model-00052-of-00061.bin", "model.layers.51.mlp.down_proj.weight": "pytorch_model-00052-of-00061.bin", "model.layers.51.mlp.up_proj.weight": "pytorch_model-00052-of-00061.bin", "model.layers.51.self_attn.rotary_emb.inv_freq": "pytorch_model-00052-of-00061.bin", "model.layers.52.input_layernorm.weight": "pytorch_model-00053-of-00061.bin", "model.layers.52.post_attention_layernorm.weight": "pytorch_model-00053-of-00061.bin", "model.layers.52.self_attn.q_proj.weight": "pytorch_model-00053-of-00061.bin", "model.layers.52.self_attn.k_proj.weight": "pytorch_model-00053-of-00061.bin", "model.layers.52.self_attn.v_proj.weight": "pytorch_model-00053-of-00061.bin", "model.layers.52.self_attn.o_proj.weight": "pytorch_model-00053-of-00061.bin", "model.layers.52.mlp.gate_proj.weight": "pytorch_model-00053-of-00061.bin", "model.layers.52.mlp.down_proj.weight": "pytorch_model-00053-of-00061.bin", "model.layers.52.mlp.up_proj.weight": "pytorch_model-00053-of-00061.bin", "model.layers.52.self_attn.rotary_emb.inv_freq": "pytorch_model-00053-of-00061.bin", "model.layers.53.input_layernorm.weight": "pytorch_model-00054-of-00061.bin", "model.layers.53.post_attention_layernorm.weight": "pytorch_model-00054-of-00061.bin", "model.layers.53.self_attn.q_proj.weight": "pytorch_model-00054-of-00061.bin", "model.layers.53.self_attn.k_proj.weight": "pytorch_model-00054-of-00061.bin", "model.layers.53.self_attn.v_proj.weight": "pytorch_model-00054-of-00061.bin", "model.layers.53.self_attn.o_proj.weight": "pytorch_model-00054-of-00061.bin", "model.layers.53.mlp.gate_proj.weight": "pytorch_model-00054-of-00061.bin", "model.layers.53.mlp.down_proj.weight": "pytorch_model-00054-of-00061.bin", "model.layers.53.mlp.up_proj.weight": "pytorch_model-00054-of-00061.bin", "model.layers.53.self_attn.rotary_emb.inv_freq": "pytorch_model-00054-of-00061.bin", "model.layers.54.input_layernorm.weight": "pytorch_model-00055-of-00061.bin", "model.layers.54.post_attention_layernorm.weight": "pytorch_model-00055-of-00061.bin", "model.layers.54.self_attn.q_proj.weight": "pytorch_model-00055-of-00061.bin", "model.layers.54.self_attn.k_proj.weight": "pytorch_model-00055-of-00061.bin", "model.layers.54.self_attn.v_proj.weight": "pytorch_model-00055-of-00061.bin", "model.layers.54.self_attn.o_proj.weight": "pytorch_model-00055-of-00061.bin", "model.layers.54.mlp.gate_proj.weight": "pytorch_model-00055-of-00061.bin", "model.layers.54.mlp.down_proj.weight": "pytorch_model-00055-of-00061.bin", "model.layers.54.mlp.up_proj.weight": "pytorch_model-00055-of-00061.bin", "model.layers.54.self_attn.rotary_emb.inv_freq": "pytorch_model-00055-of-00061.bin", "model.layers.55.input_layernorm.weight": "pytorch_model-00056-of-00061.bin", "model.layers.55.post_attention_layernorm.weight": "pytorch_model-00056-of-00061.bin", "model.layers.55.self_attn.q_proj.weight": "pytorch_model-00056-of-00061.bin", "model.layers.55.self_attn.k_proj.weight": "pytorch_model-00056-of-00061.bin", "model.layers.55.self_attn.v_proj.weight": "pytorch_model-00056-of-00061.bin", "model.layers.55.self_attn.o_proj.weight": "pytorch_model-00056-of-00061.bin", "model.layers.55.mlp.gate_proj.weight": "pytorch_model-00056-of-00061.bin", "model.layers.55.mlp.down_proj.weight": "pytorch_model-00056-of-00061.bin", "model.layers.55.mlp.up_proj.weight": "pytorch_model-00056-of-00061.bin", "model.layers.55.self_attn.rotary_emb.inv_freq": "pytorch_model-00056-of-00061.bin", "model.layers.56.input_layernorm.weight": "pytorch_model-00057-of-00061.bin", "model.layers.56.post_attention_layernorm.weight": "pytorch_model-00057-of-00061.bin", "model.layers.56.self_attn.q_proj.weight": "pytorch_model-00057-of-00061.bin", "model.layers.56.self_attn.k_proj.weight": "pytorch_model-00057-of-00061.bin", "model.layers.56.self_attn.v_proj.weight": "pytorch_model-00057-of-00061.bin", "model.layers.56.self_attn.o_proj.weight": "pytorch_model-00057-of-00061.bin", "model.layers.56.mlp.gate_proj.weight": "pytorch_model-00057-of-00061.bin", "model.layers.56.mlp.down_proj.weight": "pytorch_model-00057-of-00061.bin", "model.layers.56.mlp.up_proj.weight": "pytorch_model-00057-of-00061.bin", "model.layers.56.self_attn.rotary_emb.inv_freq": "pytorch_model-00057-of-00061.bin", "model.layers.57.input_layernorm.weight": "pytorch_model-00058-of-00061.bin", "model.layers.57.post_attention_layernorm.weight": "pytorch_model-00058-of-00061.bin", "model.layers.57.self_attn.q_proj.weight": "pytorch_model-00058-of-00061.bin", "model.layers.57.self_attn.k_proj.weight": "pytorch_model-00058-of-00061.bin", "model.layers.57.self_attn.v_proj.weight": "pytorch_model-00058-of-00061.bin", "model.layers.57.self_attn.o_proj.weight": "pytorch_model-00058-of-00061.bin", "model.layers.57.mlp.gate_proj.weight": "pytorch_model-00058-of-00061.bin", "model.layers.57.mlp.down_proj.weight": "pytorch_model-00058-of-00061.bin", "model.layers.57.mlp.up_proj.weight": "pytorch_model-00058-of-00061.bin", "model.layers.57.self_attn.rotary_emb.inv_freq": "pytorch_model-00058-of-00061.bin", "model.layers.58.input_layernorm.weight": "pytorch_model-00059-of-00061.bin", "model.layers.58.post_attention_layernorm.weight": "pytorch_model-00059-of-00061.bin", "model.layers.58.self_attn.q_proj.weight": "pytorch_model-00059-of-00061.bin", "model.layers.58.self_attn.k_proj.weight": "pytorch_model-00059-of-00061.bin", "model.layers.58.self_attn.v_proj.weight": "pytorch_model-00059-of-00061.bin", "model.layers.58.self_attn.o_proj.weight": "pytorch_model-00059-of-00061.bin", "model.layers.58.mlp.gate_proj.weight": "pytorch_model-00059-of-00061.bin", "model.layers.58.mlp.down_proj.weight": "pytorch_model-00059-of-00061.bin", "model.layers.58.mlp.up_proj.weight": "pytorch_model-00059-of-00061.bin", "model.layers.58.self_attn.rotary_emb.inv_freq": "pytorch_model-00059-of-00061.bin", "model.layers.59.input_layernorm.weight": "pytorch_model-00060-of-00061.bin", "model.layers.59.post_attention_layernorm.weight": "pytorch_model-00060-of-00061.bin", "model.layers.59.self_attn.q_proj.weight": "pytorch_model-00060-of-00061.bin", "model.layers.59.self_attn.k_proj.weight": "pytorch_model-00060-of-00061.bin", "model.layers.59.self_attn.v_proj.weight": "pytorch_model-00060-of-00061.bin", "model.layers.59.self_attn.o_proj.weight": "pytorch_model-00060-of-00061.bin", "model.layers.59.mlp.gate_proj.weight": "pytorch_model-00060-of-00061.bin", "model.layers.59.mlp.down_proj.weight": "pytorch_model-00060-of-00061.bin", "model.layers.59.mlp.up_proj.weight": "pytorch_model-00060-of-00061.bin", "model.layers.59.self_attn.rotary_emb.inv_freq": "pytorch_model-00060-of-00061.bin", "model.norm.weight": "pytorch_model-00061-of-00061.bin", "model.embed_tokens.weight": "pytorch_model-00061-of-00061.bin", "lm_head.weight": "pytorch_model-00061-of-00061.bin"}, "metadata": {"total_size": 65057894912}} \ No newline at end of file diff --git a/tokenizer_config.json b/tokenizer_config.json index 1aabb918c81c56bcb61ba76e6f93dc3ff601ee1a..a54b01aa3699f19e1aea416fc337f910f60c6839 100644 --- a/tokenizer_config.json +++ b/tokenizer_config.json @@ -1 +1 @@ -{"bos_token": "", "eos_token": "", "model_max_length": 1000000000000000019884624838656, "tokenizer_class": "LLaMATokenizer", "unk_token": ""} \ No newline at end of file +{"bos_token": "", "eos_token": "", "model_max_length": 1000000000000000019884624838656, "tokenizer_class": "LlamaTokenizer", "unk_token": ""} \ No newline at end of file