diff --git a/model-00001-of-00051.safetensors b/model-00001-of-00051.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..54869948cf8a548ba821fa8eff7cb763b750f1d6 --- /dev/null +++ b/model-00001-of-00051.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:cab0ce8b1402c68211cc18dd5e33b96c5ce8a33eb602ae1c261317fbf0fe5835 +size 4932577008 diff --git a/model-00002-of-00051.safetensors b/model-00002-of-00051.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..d4bc97066123b4facfa877a29fc14429f545daed --- /dev/null +++ b/model-00002-of-00051.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:844aece8b4f324e3526025e86813cb6b2ce939f0cbac9732c49da0cf195315e9 +size 4831938528 diff --git a/model-00003-of-00051.safetensors b/model-00003-of-00051.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..1aa659c0af391b663dd418b846b223a62dfeb506 --- /dev/null +++ b/model-00003-of-00051.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:4d502d68ec3753f8a2101e3ad97d10ba10b8c6a3d3696f7168e33639d6b7edda +size 4882269816 diff --git a/model-00004-of-00051.safetensors b/model-00004-of-00051.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..2ce5163e962d8bbe89c331444261723a1cdf0e9d --- /dev/null +++ b/model-00004-of-00051.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:c4573c03a714b0a0a89a1cf080726b7736de335128bcf12a785b398b68e1b3af +size 4831889136 diff --git a/model-00005-of-00051.safetensors b/model-00005-of-00051.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..9aa3e61b8b3c0ded4fe5066701ebb664ca65295d --- /dev/null +++ b/model-00005-of-00051.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:fd2858380ffc7ebcd8a161e328be897ec3468ead27cbcc26a926b2983f1165fa +size 4831938520 diff --git a/model-00006-of-00051.safetensors b/model-00006-of-00051.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..abba43ecd83098c71c39ca90ab2ca92eafff46f3 --- /dev/null +++ b/model-00006-of-00051.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:985c66c0f66decb20428784a5f5213f1572a19742fd8964013f680705cbd95f7 +size 4831938528 diff --git a/model-00007-of-00051.safetensors b/model-00007-of-00051.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..7d7d9eb6b94c995cfba021b88d8b1a8d576d6f42 --- /dev/null +++ b/model-00007-of-00051.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:e183ed7e2fee41e91d92f0223752994a1e4269ea255aabceca1d6c6c47fa03a0 +size 4882269832 diff --git a/model-00008-of-00051.safetensors b/model-00008-of-00051.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..13bc0778641a056c62bdab680337f9fb7b417ade --- /dev/null +++ b/model-00008-of-00051.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:0b5585f63974f210f0dcb95b1bad6053957b1324adee67bc754282f0e4e961ce +size 4831889152 diff --git a/model-00009-of-00051.safetensors b/model-00009-of-00051.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..715f015499f2790b11e28309a949bcbda14296f0 --- /dev/null +++ b/model-00009-of-00051.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:24ac96edc17c48b910a51b6ae5e93c59004c8aae01b84acc6b64bc941aab27af +size 4831938536 diff --git a/model-00010-of-00051.safetensors b/model-00010-of-00051.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..fda1b09ba093be88497062febff12a4520d29f8e --- /dev/null +++ b/model-00010-of-00051.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:bb03fd90350bff28d91e1017e31e89a753ace52ad26f6f9972e6875fe7f92693 +size 4831938544 diff --git a/model-00011-of-00051.safetensors b/model-00011-of-00051.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..74498cf8dbbffecec35416b95b84be283a43a608 --- /dev/null +++ b/model-00011-of-00051.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:4aa89d854291ef71f30cfc95046aee600ad87b1c3b2d975de0a32b63eea9ad55 +size 4882269832 diff --git a/model-00012-of-00051.safetensors b/model-00012-of-00051.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..f99bd2ca3966580d459124e642fd25984de3eacf --- /dev/null +++ b/model-00012-of-00051.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:d02f911e42eb34d24519c6b4dbc79fc3e8aa2f349b4c2e87d0aebb01ba6989a9 +size 4831889152 diff --git a/model-00013-of-00051.safetensors b/model-00013-of-00051.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..84950ca053099002c6d134ddcb8d6cb17cc53a21 --- /dev/null +++ b/model-00013-of-00051.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:f79abd028620bfeacd5368946c1cbceb42a37a4f164be69019f37a1c2ea08ee3 +size 4831938536 diff --git a/model-00014-of-00051.safetensors b/model-00014-of-00051.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..d9d0f07f811e6a4d1c9648631c82ffbeac61ee06 --- /dev/null +++ b/model-00014-of-00051.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:1cfa4fe0bec754dc6e11d3028516c61e7e5e088031d2c491786e27b6c015b95e +size 4831938544 diff --git a/model-00015-of-00051.safetensors b/model-00015-of-00051.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..69bcda1351a32e0984016c1f66bf28f4c3f43e9c --- /dev/null +++ b/model-00015-of-00051.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:d2707415c8862b863dced79ac20c1db5d7a34f67ada8d665ac81cf6814448517 +size 4882269832 diff --git a/model-00016-of-00051.safetensors b/model-00016-of-00051.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..60297092252e4e220f478828f92a331b5fd8b373 --- /dev/null +++ b/model-00016-of-00051.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:3b462a1b0192a465750618b28cf7e63fc425f6ce1ac288eb10d244618e8d02e0 +size 4831889152 diff --git a/model-00017-of-00051.safetensors b/model-00017-of-00051.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..6c4996213fdda798603446fc238f41f64de8415e --- /dev/null +++ b/model-00017-of-00051.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:e22613b5ca0b5ca2d086a5c4306a680f6fae3a8e78d1de8b8a16db930e9ee43d +size 4831938536 diff --git a/model-00018-of-00051.safetensors b/model-00018-of-00051.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..597be73de06814ccbc6249116e77b90e7618176d --- /dev/null +++ b/model-00018-of-00051.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:42312ffe1555f235179936aa40a094f7ad9c6d9fcf6ea5f113504a2275848f6e +size 4831938544 diff --git a/model-00019-of-00051.safetensors b/model-00019-of-00051.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..e70ac880e4f195d639f643dec2e104d718f4cc9d --- /dev/null +++ b/model-00019-of-00051.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:18d672582755f600823d5e6dcc98ae1f19589330f35c0c7a653b8a2927e07f9a +size 4882269832 diff --git a/model-00020-of-00051.safetensors b/model-00020-of-00051.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..03a995926cefba6f9bcdf720e66dfe1da9438647 --- /dev/null +++ b/model-00020-of-00051.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:c9517682195b8ea550381399700b70c755206fc7fedb92adebd64a481cac61c5 +size 4831889152 diff --git a/model-00021-of-00051.safetensors b/model-00021-of-00051.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..7f1f3b8abaac152b970c43cdb5427e491de501bd --- /dev/null +++ b/model-00021-of-00051.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:3533643f664503ffb540b7fab7b40a33bdd02e0c335deeb4e47ae63a4b2b7e68 +size 4831938536 diff --git a/model-00022-of-00051.safetensors b/model-00022-of-00051.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..f741dd8e90d7baaff6d0d64cc7cc635f33e3c8fe --- /dev/null +++ b/model-00022-of-00051.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:b60d9e934ac41fcaec9e23f82058a1d000322754f60a823ff7e1b9b59cd98b9f +size 4831938544 diff --git a/model-00023-of-00051.safetensors b/model-00023-of-00051.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..c68bdaf51366ebe14e6276042119825c97c57b95 --- /dev/null +++ b/model-00023-of-00051.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:793a60189a0e0841a7896a6b6db36c8e9b1b17f77d1b444ded53941e79109b54 +size 4882269832 diff --git a/model-00024-of-00051.safetensors b/model-00024-of-00051.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..0d15cf4866c615d45d9c1a2d8b7e6820b8e4fc5b --- /dev/null +++ b/model-00024-of-00051.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:3168f5e7c44ccc3ff925ae601479093166df2862a2ed705f39b91992c7feb0e2 +size 4831889152 diff --git a/model-00025-of-00051.safetensors b/model-00025-of-00051.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..fdcb1957741f540d78ab39e00ad2252fe9abb480 --- /dev/null +++ b/model-00025-of-00051.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:c3541a51fe251150150aa231b63aa1aa10e4e0e360acdb9d3b0d27b2b25cb797 +size 4831938536 diff --git a/model-00026-of-00051.safetensors b/model-00026-of-00051.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..acb6917f0bb9ed924c337edcd27782fb1392e804 --- /dev/null +++ b/model-00026-of-00051.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:c110c4044184f764771c8060fccaf15c24ade127c9ec0421af6a98e6f310bfda +size 4831938544 diff --git a/model-00027-of-00051.safetensors b/model-00027-of-00051.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..b0391c38f915dc08088f9df2b7c794c73338759f --- /dev/null +++ b/model-00027-of-00051.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:4b01ab4d5327b682ff82e9de5f78dea60958b615f4c69aa0035b4876f3174281 +size 4882269832 diff --git a/model-00028-of-00051.safetensors b/model-00028-of-00051.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..765bb8f2245751b083d0b0390af8827a54db49cd --- /dev/null +++ b/model-00028-of-00051.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:77e6ce17c648cc18848d10057bc8bd68e60db7a16d055467edfcc320e10ebc08 +size 4831889152 diff --git a/model-00029-of-00051.safetensors b/model-00029-of-00051.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..7d181dc36273bd5f42d559c72af5bbbefa8a85fa --- /dev/null +++ b/model-00029-of-00051.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:53d59c22fa826e0aac2c27fe93c21df288fa912d48aa7f46aaa4164d1ccaec8f +size 4831938536 diff --git a/model-00030-of-00051.safetensors b/model-00030-of-00051.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..e459c2c19ae8ff8b2ab5d80a692d671155abf01d --- /dev/null +++ b/model-00030-of-00051.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:eb39d6c7b2c82051a1d1e37efca6bc5752d7e5ee57d4f8eabfcd89eaf275fa86 +size 4831938544 diff --git a/model-00031-of-00051.safetensors b/model-00031-of-00051.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..42381f7df7ae75b9b2988fe105e711ca14c7a13c --- /dev/null +++ b/model-00031-of-00051.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:6e58c1de0386fd43a3b304e223971ef5b59da79ec1dadd00ca059b126730a47e +size 4882269832 diff --git a/model-00032-of-00051.safetensors b/model-00032-of-00051.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..58f266aa7c699dbd701dba85189bba7e887c2a24 --- /dev/null +++ b/model-00032-of-00051.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:070d8b6833da4454813fc857d780fbcab537b3edb0b18b78ff01b3d798c05228 +size 4831889152 diff --git a/model-00033-of-00051.safetensors b/model-00033-of-00051.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..6234ae256ab2b20b4778c334bf69a49c718ec8d6 --- /dev/null +++ b/model-00033-of-00051.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:a77e07a9c85f0ae8290336054621362fe71d8033025d514f6fe8d4c7c176217d +size 4831938536 diff --git a/model-00034-of-00051.safetensors b/model-00034-of-00051.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..d25a1cfee5d5a1ac7999772c189e463d10a89b50 --- /dev/null +++ b/model-00034-of-00051.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:fddea9b6042e8ed67acd9be086fd4b813777a3a19f7f4f06e0436a4517616888 +size 4831938544 diff --git a/model-00035-of-00051.safetensors b/model-00035-of-00051.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..0ba66f54ece7253e49f2b5fa77ab41037099aa67 --- /dev/null +++ b/model-00035-of-00051.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:9d57421df67aaa1a2b6d15d8f477ae32cd56780321468c1a1e767b902b828b4f +size 4882269832 diff --git a/model-00036-of-00051.safetensors b/model-00036-of-00051.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..6711c8399e8b9626e06acddaf5112774255ff141 --- /dev/null +++ b/model-00036-of-00051.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:5daaac424c0a21e512801baae23f2793c1b7f8991247641c9921f33221a092d5 +size 4831889152 diff --git a/model-00037-of-00051.safetensors b/model-00037-of-00051.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..b5307f44cc758415b0abef0863c26ed3eea66756 --- /dev/null +++ b/model-00037-of-00051.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:fe6af8056fb7efb75c996fcd13e6882026d54d101c16e2b794ffb4417da942f2 +size 4831938536 diff --git a/model-00038-of-00051.safetensors b/model-00038-of-00051.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..c635d3df51fca2ce95c52df9d1d645182efcfaa5 --- /dev/null +++ b/model-00038-of-00051.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:28a25340df4299426bbe6772dc9be4c42e072728fa9f8d49084d9ec5be8b25f7 +size 4831938544 diff --git a/model-00039-of-00051.safetensors b/model-00039-of-00051.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..7d95707bebcaf4eb78b37268bff26465eb06f5fc --- /dev/null +++ b/model-00039-of-00051.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:45bd63625fbe97f028ec3c86c69a57a0951a932486ffd3e45513b4053d9075b2 +size 4882269832 diff --git a/model-00040-of-00051.safetensors b/model-00040-of-00051.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..65209975e668e67c631e44f4dfc9ee500bf52bd4 --- /dev/null +++ b/model-00040-of-00051.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:05e2c83afeee6068648d7b07a6fa2e53648259501ab305fdad886ea95d969c68 +size 4831889152 diff --git a/model-00041-of-00051.safetensors b/model-00041-of-00051.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..b7582e94b5782e67f1eecde2f58dfcd347dc3baf --- /dev/null +++ b/model-00041-of-00051.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:223eac33cad7cc1c1b940cc33ca0ef9fc4bf6d93a5d5e39f4536f433fc580ff0 +size 4831938536 diff --git a/model-00042-of-00051.safetensors b/model-00042-of-00051.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..a565a6c0e63aa8e022ef04538eef86186238b8c1 --- /dev/null +++ b/model-00042-of-00051.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:2e76a11067a52e329446705928b743b9d22097d3e9ea113224398e23662cb472 +size 4831938544 diff --git a/model-00043-of-00051.safetensors b/model-00043-of-00051.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..d4083e72bbbc65db2f352b2a8ea7087b7420234e --- /dev/null +++ b/model-00043-of-00051.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:27457a59dab6b7dd22721145516402f846b42b7eed3672dfe2bc3907303e4ccd +size 4882269832 diff --git a/model-00044-of-00051.safetensors b/model-00044-of-00051.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..5380386caa21ba3f85d4e75aa46fdd1d6ef7da47 --- /dev/null +++ b/model-00044-of-00051.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:bd9943c3ccaf790afda2ffca14e96b39cbfc42f2b5b1dd6773a55725159eb007 +size 4831889152 diff --git a/model-00045-of-00051.safetensors b/model-00045-of-00051.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..1151e71534af6642f337271f1f4bc14620a94f9d --- /dev/null +++ b/model-00045-of-00051.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:aeb7fa2d1e2e30085350babc5962b011afeaf50348ed2713c298febf3708f1f1 +size 4831938536 diff --git a/model-00046-of-00051.safetensors b/model-00046-of-00051.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..3d52ffbd93747af718906fda93ae30b971c79e21 --- /dev/null +++ b/model-00046-of-00051.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:02347fdd33347cbaa2a8b5ce7e6a66a3860dd5b7b52272e9eb38440d1ace8268 +size 4831938544 diff --git a/model-00047-of-00051.safetensors b/model-00047-of-00051.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..14f71a5c34578263cf739fbab058120f6610ad79 --- /dev/null +++ b/model-00047-of-00051.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:8edc4c8d7d173e1f93f0366f2e1d43cf7211dca3693561102ccf6ba130d77f5d +size 4882269832 diff --git a/model-00048-of-00051.safetensors b/model-00048-of-00051.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..9f5b4be73fb1327f0137e921a275d16977d81c9f --- /dev/null +++ b/model-00048-of-00051.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:6420439422a3d0d7f02e64e2d211272284679b043a31945ba549ac0a631625d4 +size 4831889152 diff --git a/model-00049-of-00051.safetensors b/model-00049-of-00051.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..f5b0cea1cb8754ff4806482f43290c885250ffc7 --- /dev/null +++ b/model-00049-of-00051.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:885619a959925d9057447f466513c3238b4cd4448dbfaa7033e4fb69ba5d7f94 +size 4831938536 diff --git a/model-00050-of-00051.safetensors b/model-00050-of-00051.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..b403270c4777008f947429be852c3cfee1c562c7 --- /dev/null +++ b/model-00050-of-00051.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:d402a338ce50d5c7d712bb8b41274f2e1d3c61a0b8652d34f5093e4dfabab7cb +size 4831938544 diff --git a/model-00051-of-00051.safetensors b/model-00051-of-00051.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..6109190ff233ab1dd13baff3bc5bea270e5c6d7b --- /dev/null +++ b/model-00051-of-00051.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:f95a948a3b94918ead8ad3630b8f0ece3d00f104a69a834e6fdd85e8b81d8339 +size 2919334688 diff --git a/model.safetensors.index.json b/model.safetensors.index.json new file mode 100644 index 0000000000000000000000000000000000000000..a91a2e2b0a69176ff845fee59c30b563aa8b6854 --- /dev/null +++ b/model.safetensors.index.json @@ -0,0 +1,802 @@ +{ + "metadata": { + "total_size": 245220188160 + }, + "weight_map": { + "lm_head.weight": "model-00051-of-00051.safetensors", + "model.embed_tokens.weight": "model-00001-of-00051.safetensors", + "model.layers.0.input_layernorm.weight": "model-00001-of-00051.safetensors", + "model.layers.0.mlp.down_proj.weight": "model-00001-of-00051.safetensors", + "model.layers.0.mlp.gate_proj.weight": "model-00001-of-00051.safetensors", + "model.layers.0.mlp.up_proj.weight": "model-00001-of-00051.safetensors", + "model.layers.0.post_attention_layernorm.weight": "model-00001-of-00051.safetensors", + "model.layers.0.self_attn.k_proj.weight": "model-00001-of-00051.safetensors", + "model.layers.0.self_attn.o_proj.weight": "model-00001-of-00051.safetensors", + "model.layers.0.self_attn.q_proj.weight": "model-00001-of-00051.safetensors", + "model.layers.0.self_attn.v_proj.weight": "model-00001-of-00051.safetensors", + "model.layers.1.input_layernorm.weight": "model-00002-of-00051.safetensors", + "model.layers.1.mlp.down_proj.weight": "model-00002-of-00051.safetensors", + "model.layers.1.mlp.gate_proj.weight": "model-00001-of-00051.safetensors", + "model.layers.1.mlp.up_proj.weight": "model-00002-of-00051.safetensors", + "model.layers.1.post_attention_layernorm.weight": "model-00002-of-00051.safetensors", + "model.layers.1.self_attn.k_proj.weight": "model-00001-of-00051.safetensors", + "model.layers.1.self_attn.o_proj.weight": "model-00001-of-00051.safetensors", + "model.layers.1.self_attn.q_proj.weight": "model-00001-of-00051.safetensors", + "model.layers.1.self_attn.v_proj.weight": "model-00001-of-00051.safetensors", + "model.layers.10.input_layernorm.weight": "model-00007-of-00051.safetensors", + "model.layers.10.mlp.down_proj.weight": "model-00007-of-00051.safetensors", + "model.layers.10.mlp.gate_proj.weight": "model-00007-of-00051.safetensors", + "model.layers.10.mlp.up_proj.weight": "model-00007-of-00051.safetensors", + "model.layers.10.post_attention_layernorm.weight": "model-00007-of-00051.safetensors", + "model.layers.10.self_attn.k_proj.weight": "model-00006-of-00051.safetensors", + "model.layers.10.self_attn.o_proj.weight": "model-00006-of-00051.safetensors", + "model.layers.10.self_attn.q_proj.weight": "model-00006-of-00051.safetensors", + "model.layers.10.self_attn.v_proj.weight": "model-00006-of-00051.safetensors", + "model.layers.11.input_layernorm.weight": "model-00007-of-00051.safetensors", + "model.layers.11.mlp.down_proj.weight": "model-00007-of-00051.safetensors", + "model.layers.11.mlp.gate_proj.weight": "model-00007-of-00051.safetensors", + "model.layers.11.mlp.up_proj.weight": "model-00007-of-00051.safetensors", + "model.layers.11.post_attention_layernorm.weight": "model-00007-of-00051.safetensors", + "model.layers.11.self_attn.k_proj.weight": "model-00007-of-00051.safetensors", + "model.layers.11.self_attn.o_proj.weight": "model-00007-of-00051.safetensors", + "model.layers.11.self_attn.q_proj.weight": "model-00007-of-00051.safetensors", + "model.layers.11.self_attn.v_proj.weight": "model-00007-of-00051.safetensors", + "model.layers.12.input_layernorm.weight": "model-00008-of-00051.safetensors", + "model.layers.12.mlp.down_proj.weight": "model-00008-of-00051.safetensors", + "model.layers.12.mlp.gate_proj.weight": "model-00008-of-00051.safetensors", + "model.layers.12.mlp.up_proj.weight": "model-00008-of-00051.safetensors", + "model.layers.12.post_attention_layernorm.weight": "model-00008-of-00051.safetensors", + "model.layers.12.self_attn.k_proj.weight": "model-00008-of-00051.safetensors", + "model.layers.12.self_attn.o_proj.weight": "model-00008-of-00051.safetensors", + "model.layers.12.self_attn.q_proj.weight": "model-00008-of-00051.safetensors", + "model.layers.12.self_attn.v_proj.weight": "model-00008-of-00051.safetensors", + "model.layers.13.input_layernorm.weight": "model-00009-of-00051.safetensors", + "model.layers.13.mlp.down_proj.weight": "model-00009-of-00051.safetensors", + "model.layers.13.mlp.gate_proj.weight": "model-00008-of-00051.safetensors", + "model.layers.13.mlp.up_proj.weight": "model-00008-of-00051.safetensors", + "model.layers.13.post_attention_layernorm.weight": "model-00009-of-00051.safetensors", + "model.layers.13.self_attn.k_proj.weight": "model-00008-of-00051.safetensors", + "model.layers.13.self_attn.o_proj.weight": "model-00008-of-00051.safetensors", + "model.layers.13.self_attn.q_proj.weight": "model-00008-of-00051.safetensors", + "model.layers.13.self_attn.v_proj.weight": "model-00008-of-00051.safetensors", + "model.layers.14.input_layernorm.weight": "model-00009-of-00051.safetensors", + "model.layers.14.mlp.down_proj.weight": "model-00009-of-00051.safetensors", + "model.layers.14.mlp.gate_proj.weight": "model-00009-of-00051.safetensors", + "model.layers.14.mlp.up_proj.weight": "model-00009-of-00051.safetensors", + "model.layers.14.post_attention_layernorm.weight": "model-00009-of-00051.safetensors", + "model.layers.14.self_attn.k_proj.weight": "model-00009-of-00051.safetensors", + "model.layers.14.self_attn.o_proj.weight": "model-00009-of-00051.safetensors", + "model.layers.14.self_attn.q_proj.weight": "model-00009-of-00051.safetensors", + "model.layers.14.self_attn.v_proj.weight": "model-00009-of-00051.safetensors", + "model.layers.15.input_layernorm.weight": "model-00010-of-00051.safetensors", + "model.layers.15.mlp.down_proj.weight": "model-00010-of-00051.safetensors", + "model.layers.15.mlp.gate_proj.weight": "model-00009-of-00051.safetensors", + "model.layers.15.mlp.up_proj.weight": "model-00010-of-00051.safetensors", + "model.layers.15.post_attention_layernorm.weight": "model-00010-of-00051.safetensors", + "model.layers.15.self_attn.k_proj.weight": "model-00009-of-00051.safetensors", + "model.layers.15.self_attn.o_proj.weight": "model-00009-of-00051.safetensors", + "model.layers.15.self_attn.q_proj.weight": "model-00009-of-00051.safetensors", + "model.layers.15.self_attn.v_proj.weight": "model-00009-of-00051.safetensors", + "model.layers.16.input_layernorm.weight": "model-00010-of-00051.safetensors", + "model.layers.16.mlp.down_proj.weight": "model-00010-of-00051.safetensors", + "model.layers.16.mlp.gate_proj.weight": "model-00010-of-00051.safetensors", + "model.layers.16.mlp.up_proj.weight": "model-00010-of-00051.safetensors", + "model.layers.16.post_attention_layernorm.weight": "model-00010-of-00051.safetensors", + "model.layers.16.self_attn.k_proj.weight": "model-00010-of-00051.safetensors", + "model.layers.16.self_attn.o_proj.weight": "model-00010-of-00051.safetensors", + "model.layers.16.self_attn.q_proj.weight": "model-00010-of-00051.safetensors", + "model.layers.16.self_attn.v_proj.weight": "model-00010-of-00051.safetensors", + "model.layers.17.input_layernorm.weight": "model-00011-of-00051.safetensors", + "model.layers.17.mlp.down_proj.weight": "model-00011-of-00051.safetensors", + "model.layers.17.mlp.gate_proj.weight": "model-00011-of-00051.safetensors", + "model.layers.17.mlp.up_proj.weight": "model-00011-of-00051.safetensors", + "model.layers.17.post_attention_layernorm.weight": "model-00011-of-00051.safetensors", + "model.layers.17.self_attn.k_proj.weight": "model-00010-of-00051.safetensors", + "model.layers.17.self_attn.o_proj.weight": "model-00010-of-00051.safetensors", + "model.layers.17.self_attn.q_proj.weight": "model-00010-of-00051.safetensors", + "model.layers.17.self_attn.v_proj.weight": "model-00010-of-00051.safetensors", + "model.layers.18.input_layernorm.weight": "model-00011-of-00051.safetensors", + "model.layers.18.mlp.down_proj.weight": "model-00011-of-00051.safetensors", + "model.layers.18.mlp.gate_proj.weight": "model-00011-of-00051.safetensors", + "model.layers.18.mlp.up_proj.weight": "model-00011-of-00051.safetensors", + "model.layers.18.post_attention_layernorm.weight": "model-00011-of-00051.safetensors", + "model.layers.18.self_attn.k_proj.weight": "model-00011-of-00051.safetensors", + "model.layers.18.self_attn.o_proj.weight": "model-00011-of-00051.safetensors", + "model.layers.18.self_attn.q_proj.weight": "model-00011-of-00051.safetensors", + "model.layers.18.self_attn.v_proj.weight": "model-00011-of-00051.safetensors", + "model.layers.19.input_layernorm.weight": "model-00012-of-00051.safetensors", + "model.layers.19.mlp.down_proj.weight": "model-00012-of-00051.safetensors", + "model.layers.19.mlp.gate_proj.weight": "model-00012-of-00051.safetensors", + "model.layers.19.mlp.up_proj.weight": "model-00012-of-00051.safetensors", + "model.layers.19.post_attention_layernorm.weight": "model-00012-of-00051.safetensors", + "model.layers.19.self_attn.k_proj.weight": "model-00012-of-00051.safetensors", + "model.layers.19.self_attn.o_proj.weight": "model-00012-of-00051.safetensors", + "model.layers.19.self_attn.q_proj.weight": "model-00012-of-00051.safetensors", + "model.layers.19.self_attn.v_proj.weight": "model-00012-of-00051.safetensors", + "model.layers.2.input_layernorm.weight": "model-00002-of-00051.safetensors", + "model.layers.2.mlp.down_proj.weight": "model-00002-of-00051.safetensors", + "model.layers.2.mlp.gate_proj.weight": "model-00002-of-00051.safetensors", + "model.layers.2.mlp.up_proj.weight": "model-00002-of-00051.safetensors", + "model.layers.2.post_attention_layernorm.weight": "model-00002-of-00051.safetensors", + "model.layers.2.self_attn.k_proj.weight": "model-00002-of-00051.safetensors", + "model.layers.2.self_attn.o_proj.weight": "model-00002-of-00051.safetensors", + "model.layers.2.self_attn.q_proj.weight": "model-00002-of-00051.safetensors", + "model.layers.2.self_attn.v_proj.weight": "model-00002-of-00051.safetensors", + "model.layers.20.input_layernorm.weight": "model-00013-of-00051.safetensors", + "model.layers.20.mlp.down_proj.weight": "model-00013-of-00051.safetensors", + "model.layers.20.mlp.gate_proj.weight": "model-00012-of-00051.safetensors", + "model.layers.20.mlp.up_proj.weight": "model-00012-of-00051.safetensors", + "model.layers.20.post_attention_layernorm.weight": "model-00013-of-00051.safetensors", + "model.layers.20.self_attn.k_proj.weight": "model-00012-of-00051.safetensors", + "model.layers.20.self_attn.o_proj.weight": "model-00012-of-00051.safetensors", + "model.layers.20.self_attn.q_proj.weight": "model-00012-of-00051.safetensors", + "model.layers.20.self_attn.v_proj.weight": "model-00012-of-00051.safetensors", + "model.layers.21.input_layernorm.weight": "model-00013-of-00051.safetensors", + "model.layers.21.mlp.down_proj.weight": "model-00013-of-00051.safetensors", + "model.layers.21.mlp.gate_proj.weight": "model-00013-of-00051.safetensors", + "model.layers.21.mlp.up_proj.weight": "model-00013-of-00051.safetensors", + "model.layers.21.post_attention_layernorm.weight": "model-00013-of-00051.safetensors", + "model.layers.21.self_attn.k_proj.weight": "model-00013-of-00051.safetensors", + "model.layers.21.self_attn.o_proj.weight": "model-00013-of-00051.safetensors", + "model.layers.21.self_attn.q_proj.weight": "model-00013-of-00051.safetensors", + "model.layers.21.self_attn.v_proj.weight": "model-00013-of-00051.safetensors", + "model.layers.22.input_layernorm.weight": "model-00014-of-00051.safetensors", + "model.layers.22.mlp.down_proj.weight": "model-00014-of-00051.safetensors", + "model.layers.22.mlp.gate_proj.weight": "model-00013-of-00051.safetensors", + "model.layers.22.mlp.up_proj.weight": "model-00014-of-00051.safetensors", + "model.layers.22.post_attention_layernorm.weight": "model-00014-of-00051.safetensors", + "model.layers.22.self_attn.k_proj.weight": "model-00013-of-00051.safetensors", + "model.layers.22.self_attn.o_proj.weight": "model-00013-of-00051.safetensors", + "model.layers.22.self_attn.q_proj.weight": "model-00013-of-00051.safetensors", + "model.layers.22.self_attn.v_proj.weight": "model-00013-of-00051.safetensors", + "model.layers.23.input_layernorm.weight": "model-00014-of-00051.safetensors", + "model.layers.23.mlp.down_proj.weight": "model-00014-of-00051.safetensors", + "model.layers.23.mlp.gate_proj.weight": "model-00014-of-00051.safetensors", + "model.layers.23.mlp.up_proj.weight": "model-00014-of-00051.safetensors", + "model.layers.23.post_attention_layernorm.weight": "model-00014-of-00051.safetensors", + "model.layers.23.self_attn.k_proj.weight": "model-00014-of-00051.safetensors", + "model.layers.23.self_attn.o_proj.weight": "model-00014-of-00051.safetensors", + "model.layers.23.self_attn.q_proj.weight": "model-00014-of-00051.safetensors", + "model.layers.23.self_attn.v_proj.weight": "model-00014-of-00051.safetensors", + "model.layers.24.input_layernorm.weight": "model-00015-of-00051.safetensors", + "model.layers.24.mlp.down_proj.weight": "model-00015-of-00051.safetensors", + "model.layers.24.mlp.gate_proj.weight": "model-00015-of-00051.safetensors", + "model.layers.24.mlp.up_proj.weight": "model-00015-of-00051.safetensors", + "model.layers.24.post_attention_layernorm.weight": "model-00015-of-00051.safetensors", + "model.layers.24.self_attn.k_proj.weight": "model-00014-of-00051.safetensors", + "model.layers.24.self_attn.o_proj.weight": "model-00014-of-00051.safetensors", + "model.layers.24.self_attn.q_proj.weight": "model-00014-of-00051.safetensors", + "model.layers.24.self_attn.v_proj.weight": "model-00014-of-00051.safetensors", + "model.layers.25.input_layernorm.weight": "model-00015-of-00051.safetensors", + "model.layers.25.mlp.down_proj.weight": "model-00015-of-00051.safetensors", + "model.layers.25.mlp.gate_proj.weight": "model-00015-of-00051.safetensors", + "model.layers.25.mlp.up_proj.weight": "model-00015-of-00051.safetensors", + "model.layers.25.post_attention_layernorm.weight": "model-00015-of-00051.safetensors", + "model.layers.25.self_attn.k_proj.weight": "model-00015-of-00051.safetensors", + "model.layers.25.self_attn.o_proj.weight": "model-00015-of-00051.safetensors", + "model.layers.25.self_attn.q_proj.weight": "model-00015-of-00051.safetensors", + "model.layers.25.self_attn.v_proj.weight": "model-00015-of-00051.safetensors", + "model.layers.26.input_layernorm.weight": "model-00016-of-00051.safetensors", + "model.layers.26.mlp.down_proj.weight": "model-00016-of-00051.safetensors", + "model.layers.26.mlp.gate_proj.weight": "model-00016-of-00051.safetensors", + "model.layers.26.mlp.up_proj.weight": "model-00016-of-00051.safetensors", + "model.layers.26.post_attention_layernorm.weight": "model-00016-of-00051.safetensors", + "model.layers.26.self_attn.k_proj.weight": "model-00016-of-00051.safetensors", + "model.layers.26.self_attn.o_proj.weight": "model-00016-of-00051.safetensors", + "model.layers.26.self_attn.q_proj.weight": "model-00016-of-00051.safetensors", + "model.layers.26.self_attn.v_proj.weight": "model-00016-of-00051.safetensors", + "model.layers.27.input_layernorm.weight": "model-00017-of-00051.safetensors", + "model.layers.27.mlp.down_proj.weight": "model-00017-of-00051.safetensors", + "model.layers.27.mlp.gate_proj.weight": "model-00016-of-00051.safetensors", + "model.layers.27.mlp.up_proj.weight": "model-00016-of-00051.safetensors", + "model.layers.27.post_attention_layernorm.weight": "model-00017-of-00051.safetensors", + "model.layers.27.self_attn.k_proj.weight": "model-00016-of-00051.safetensors", + "model.layers.27.self_attn.o_proj.weight": "model-00016-of-00051.safetensors", + "model.layers.27.self_attn.q_proj.weight": "model-00016-of-00051.safetensors", + "model.layers.27.self_attn.v_proj.weight": "model-00016-of-00051.safetensors", + "model.layers.28.input_layernorm.weight": "model-00017-of-00051.safetensors", + "model.layers.28.mlp.down_proj.weight": "model-00017-of-00051.safetensors", + "model.layers.28.mlp.gate_proj.weight": "model-00017-of-00051.safetensors", + "model.layers.28.mlp.up_proj.weight": "model-00017-of-00051.safetensors", + "model.layers.28.post_attention_layernorm.weight": "model-00017-of-00051.safetensors", + "model.layers.28.self_attn.k_proj.weight": "model-00017-of-00051.safetensors", + "model.layers.28.self_attn.o_proj.weight": "model-00017-of-00051.safetensors", + "model.layers.28.self_attn.q_proj.weight": "model-00017-of-00051.safetensors", + "model.layers.28.self_attn.v_proj.weight": "model-00017-of-00051.safetensors", + "model.layers.29.input_layernorm.weight": "model-00018-of-00051.safetensors", + "model.layers.29.mlp.down_proj.weight": "model-00018-of-00051.safetensors", + "model.layers.29.mlp.gate_proj.weight": "model-00017-of-00051.safetensors", + "model.layers.29.mlp.up_proj.weight": "model-00018-of-00051.safetensors", + "model.layers.29.post_attention_layernorm.weight": "model-00018-of-00051.safetensors", + "model.layers.29.self_attn.k_proj.weight": "model-00017-of-00051.safetensors", + "model.layers.29.self_attn.o_proj.weight": "model-00017-of-00051.safetensors", + "model.layers.29.self_attn.q_proj.weight": "model-00017-of-00051.safetensors", + "model.layers.29.self_attn.v_proj.weight": "model-00017-of-00051.safetensors", + "model.layers.3.input_layernorm.weight": "model-00003-of-00051.safetensors", + "model.layers.3.mlp.down_proj.weight": "model-00003-of-00051.safetensors", + "model.layers.3.mlp.gate_proj.weight": "model-00003-of-00051.safetensors", + "model.layers.3.mlp.up_proj.weight": "model-00003-of-00051.safetensors", + "model.layers.3.post_attention_layernorm.weight": "model-00003-of-00051.safetensors", + "model.layers.3.self_attn.k_proj.weight": "model-00002-of-00051.safetensors", + "model.layers.3.self_attn.o_proj.weight": "model-00002-of-00051.safetensors", + "model.layers.3.self_attn.q_proj.weight": "model-00002-of-00051.safetensors", + "model.layers.3.self_attn.v_proj.weight": "model-00002-of-00051.safetensors", + "model.layers.30.input_layernorm.weight": "model-00018-of-00051.safetensors", + "model.layers.30.mlp.down_proj.weight": "model-00018-of-00051.safetensors", + "model.layers.30.mlp.gate_proj.weight": "model-00018-of-00051.safetensors", + "model.layers.30.mlp.up_proj.weight": "model-00018-of-00051.safetensors", + "model.layers.30.post_attention_layernorm.weight": "model-00018-of-00051.safetensors", + "model.layers.30.self_attn.k_proj.weight": "model-00018-of-00051.safetensors", + "model.layers.30.self_attn.o_proj.weight": "model-00018-of-00051.safetensors", + "model.layers.30.self_attn.q_proj.weight": "model-00018-of-00051.safetensors", + "model.layers.30.self_attn.v_proj.weight": "model-00018-of-00051.safetensors", + "model.layers.31.input_layernorm.weight": "model-00019-of-00051.safetensors", + "model.layers.31.mlp.down_proj.weight": "model-00019-of-00051.safetensors", + "model.layers.31.mlp.gate_proj.weight": "model-00019-of-00051.safetensors", + "model.layers.31.mlp.up_proj.weight": "model-00019-of-00051.safetensors", + "model.layers.31.post_attention_layernorm.weight": "model-00019-of-00051.safetensors", + "model.layers.31.self_attn.k_proj.weight": "model-00018-of-00051.safetensors", + "model.layers.31.self_attn.o_proj.weight": "model-00018-of-00051.safetensors", + "model.layers.31.self_attn.q_proj.weight": "model-00018-of-00051.safetensors", + "model.layers.31.self_attn.v_proj.weight": "model-00018-of-00051.safetensors", + "model.layers.32.input_layernorm.weight": "model-00019-of-00051.safetensors", + "model.layers.32.mlp.down_proj.weight": "model-00019-of-00051.safetensors", + "model.layers.32.mlp.gate_proj.weight": "model-00019-of-00051.safetensors", + "model.layers.32.mlp.up_proj.weight": "model-00019-of-00051.safetensors", + "model.layers.32.post_attention_layernorm.weight": "model-00019-of-00051.safetensors", + "model.layers.32.self_attn.k_proj.weight": "model-00019-of-00051.safetensors", + "model.layers.32.self_attn.o_proj.weight": "model-00019-of-00051.safetensors", + "model.layers.32.self_attn.q_proj.weight": "model-00019-of-00051.safetensors", + "model.layers.32.self_attn.v_proj.weight": "model-00019-of-00051.safetensors", + "model.layers.33.input_layernorm.weight": "model-00020-of-00051.safetensors", + "model.layers.33.mlp.down_proj.weight": "model-00020-of-00051.safetensors", + "model.layers.33.mlp.gate_proj.weight": "model-00020-of-00051.safetensors", + "model.layers.33.mlp.up_proj.weight": "model-00020-of-00051.safetensors", + "model.layers.33.post_attention_layernorm.weight": "model-00020-of-00051.safetensors", + "model.layers.33.self_attn.k_proj.weight": "model-00020-of-00051.safetensors", + "model.layers.33.self_attn.o_proj.weight": "model-00020-of-00051.safetensors", + "model.layers.33.self_attn.q_proj.weight": "model-00020-of-00051.safetensors", + "model.layers.33.self_attn.v_proj.weight": "model-00020-of-00051.safetensors", + "model.layers.34.input_layernorm.weight": "model-00021-of-00051.safetensors", + "model.layers.34.mlp.down_proj.weight": "model-00021-of-00051.safetensors", + "model.layers.34.mlp.gate_proj.weight": "model-00020-of-00051.safetensors", + "model.layers.34.mlp.up_proj.weight": "model-00020-of-00051.safetensors", + "model.layers.34.post_attention_layernorm.weight": "model-00021-of-00051.safetensors", + "model.layers.34.self_attn.k_proj.weight": "model-00020-of-00051.safetensors", + "model.layers.34.self_attn.o_proj.weight": "model-00020-of-00051.safetensors", + "model.layers.34.self_attn.q_proj.weight": "model-00020-of-00051.safetensors", + "model.layers.34.self_attn.v_proj.weight": "model-00020-of-00051.safetensors", + "model.layers.35.input_layernorm.weight": "model-00021-of-00051.safetensors", + "model.layers.35.mlp.down_proj.weight": "model-00021-of-00051.safetensors", + "model.layers.35.mlp.gate_proj.weight": "model-00021-of-00051.safetensors", + "model.layers.35.mlp.up_proj.weight": "model-00021-of-00051.safetensors", + "model.layers.35.post_attention_layernorm.weight": "model-00021-of-00051.safetensors", + "model.layers.35.self_attn.k_proj.weight": "model-00021-of-00051.safetensors", + "model.layers.35.self_attn.o_proj.weight": "model-00021-of-00051.safetensors", + "model.layers.35.self_attn.q_proj.weight": "model-00021-of-00051.safetensors", + "model.layers.35.self_attn.v_proj.weight": "model-00021-of-00051.safetensors", + "model.layers.36.input_layernorm.weight": "model-00022-of-00051.safetensors", + "model.layers.36.mlp.down_proj.weight": "model-00022-of-00051.safetensors", + "model.layers.36.mlp.gate_proj.weight": "model-00021-of-00051.safetensors", + "model.layers.36.mlp.up_proj.weight": "model-00022-of-00051.safetensors", + "model.layers.36.post_attention_layernorm.weight": "model-00022-of-00051.safetensors", + "model.layers.36.self_attn.k_proj.weight": "model-00021-of-00051.safetensors", + "model.layers.36.self_attn.o_proj.weight": "model-00021-of-00051.safetensors", + "model.layers.36.self_attn.q_proj.weight": "model-00021-of-00051.safetensors", + "model.layers.36.self_attn.v_proj.weight": "model-00021-of-00051.safetensors", + "model.layers.37.input_layernorm.weight": "model-00022-of-00051.safetensors", + "model.layers.37.mlp.down_proj.weight": "model-00022-of-00051.safetensors", + "model.layers.37.mlp.gate_proj.weight": "model-00022-of-00051.safetensors", + "model.layers.37.mlp.up_proj.weight": "model-00022-of-00051.safetensors", + "model.layers.37.post_attention_layernorm.weight": "model-00022-of-00051.safetensors", + "model.layers.37.self_attn.k_proj.weight": "model-00022-of-00051.safetensors", + "model.layers.37.self_attn.o_proj.weight": "model-00022-of-00051.safetensors", + "model.layers.37.self_attn.q_proj.weight": "model-00022-of-00051.safetensors", + "model.layers.37.self_attn.v_proj.weight": "model-00022-of-00051.safetensors", + "model.layers.38.input_layernorm.weight": "model-00023-of-00051.safetensors", + "model.layers.38.mlp.down_proj.weight": "model-00023-of-00051.safetensors", + "model.layers.38.mlp.gate_proj.weight": "model-00023-of-00051.safetensors", + "model.layers.38.mlp.up_proj.weight": "model-00023-of-00051.safetensors", + "model.layers.38.post_attention_layernorm.weight": "model-00023-of-00051.safetensors", + "model.layers.38.self_attn.k_proj.weight": "model-00022-of-00051.safetensors", + "model.layers.38.self_attn.o_proj.weight": "model-00022-of-00051.safetensors", + "model.layers.38.self_attn.q_proj.weight": "model-00022-of-00051.safetensors", + "model.layers.38.self_attn.v_proj.weight": "model-00022-of-00051.safetensors", + "model.layers.39.input_layernorm.weight": "model-00023-of-00051.safetensors", + "model.layers.39.mlp.down_proj.weight": "model-00023-of-00051.safetensors", + "model.layers.39.mlp.gate_proj.weight": "model-00023-of-00051.safetensors", + "model.layers.39.mlp.up_proj.weight": "model-00023-of-00051.safetensors", + "model.layers.39.post_attention_layernorm.weight": "model-00023-of-00051.safetensors", + "model.layers.39.self_attn.k_proj.weight": "model-00023-of-00051.safetensors", + "model.layers.39.self_attn.o_proj.weight": "model-00023-of-00051.safetensors", + "model.layers.39.self_attn.q_proj.weight": "model-00023-of-00051.safetensors", + "model.layers.39.self_attn.v_proj.weight": "model-00023-of-00051.safetensors", + "model.layers.4.input_layernorm.weight": "model-00003-of-00051.safetensors", + "model.layers.4.mlp.down_proj.weight": "model-00003-of-00051.safetensors", + "model.layers.4.mlp.gate_proj.weight": "model-00003-of-00051.safetensors", + "model.layers.4.mlp.up_proj.weight": "model-00003-of-00051.safetensors", + "model.layers.4.post_attention_layernorm.weight": "model-00003-of-00051.safetensors", + "model.layers.4.self_attn.k_proj.weight": "model-00003-of-00051.safetensors", + "model.layers.4.self_attn.o_proj.weight": "model-00003-of-00051.safetensors", + "model.layers.4.self_attn.q_proj.weight": "model-00003-of-00051.safetensors", + "model.layers.4.self_attn.v_proj.weight": "model-00003-of-00051.safetensors", + "model.layers.40.input_layernorm.weight": "model-00024-of-00051.safetensors", + "model.layers.40.mlp.down_proj.weight": "model-00024-of-00051.safetensors", + "model.layers.40.mlp.gate_proj.weight": "model-00024-of-00051.safetensors", + "model.layers.40.mlp.up_proj.weight": "model-00024-of-00051.safetensors", + "model.layers.40.post_attention_layernorm.weight": "model-00024-of-00051.safetensors", + "model.layers.40.self_attn.k_proj.weight": "model-00024-of-00051.safetensors", + "model.layers.40.self_attn.o_proj.weight": "model-00024-of-00051.safetensors", + "model.layers.40.self_attn.q_proj.weight": "model-00024-of-00051.safetensors", + "model.layers.40.self_attn.v_proj.weight": "model-00024-of-00051.safetensors", + "model.layers.41.input_layernorm.weight": "model-00025-of-00051.safetensors", + "model.layers.41.mlp.down_proj.weight": "model-00025-of-00051.safetensors", + "model.layers.41.mlp.gate_proj.weight": "model-00024-of-00051.safetensors", + "model.layers.41.mlp.up_proj.weight": "model-00024-of-00051.safetensors", + "model.layers.41.post_attention_layernorm.weight": "model-00025-of-00051.safetensors", + "model.layers.41.self_attn.k_proj.weight": "model-00024-of-00051.safetensors", + "model.layers.41.self_attn.o_proj.weight": "model-00024-of-00051.safetensors", + "model.layers.41.self_attn.q_proj.weight": "model-00024-of-00051.safetensors", + "model.layers.41.self_attn.v_proj.weight": "model-00024-of-00051.safetensors", + "model.layers.42.input_layernorm.weight": "model-00025-of-00051.safetensors", + "model.layers.42.mlp.down_proj.weight": "model-00025-of-00051.safetensors", + "model.layers.42.mlp.gate_proj.weight": "model-00025-of-00051.safetensors", + "model.layers.42.mlp.up_proj.weight": "model-00025-of-00051.safetensors", + "model.layers.42.post_attention_layernorm.weight": "model-00025-of-00051.safetensors", + "model.layers.42.self_attn.k_proj.weight": "model-00025-of-00051.safetensors", + "model.layers.42.self_attn.o_proj.weight": "model-00025-of-00051.safetensors", + "model.layers.42.self_attn.q_proj.weight": "model-00025-of-00051.safetensors", + "model.layers.42.self_attn.v_proj.weight": "model-00025-of-00051.safetensors", + "model.layers.43.input_layernorm.weight": "model-00026-of-00051.safetensors", + "model.layers.43.mlp.down_proj.weight": "model-00026-of-00051.safetensors", + "model.layers.43.mlp.gate_proj.weight": "model-00025-of-00051.safetensors", + "model.layers.43.mlp.up_proj.weight": "model-00026-of-00051.safetensors", + "model.layers.43.post_attention_layernorm.weight": "model-00026-of-00051.safetensors", + "model.layers.43.self_attn.k_proj.weight": "model-00025-of-00051.safetensors", + "model.layers.43.self_attn.o_proj.weight": "model-00025-of-00051.safetensors", + "model.layers.43.self_attn.q_proj.weight": "model-00025-of-00051.safetensors", + "model.layers.43.self_attn.v_proj.weight": "model-00025-of-00051.safetensors", + "model.layers.44.input_layernorm.weight": "model-00026-of-00051.safetensors", + "model.layers.44.mlp.down_proj.weight": "model-00026-of-00051.safetensors", + "model.layers.44.mlp.gate_proj.weight": "model-00026-of-00051.safetensors", + "model.layers.44.mlp.up_proj.weight": "model-00026-of-00051.safetensors", + "model.layers.44.post_attention_layernorm.weight": "model-00026-of-00051.safetensors", + "model.layers.44.self_attn.k_proj.weight": "model-00026-of-00051.safetensors", + "model.layers.44.self_attn.o_proj.weight": "model-00026-of-00051.safetensors", + "model.layers.44.self_attn.q_proj.weight": "model-00026-of-00051.safetensors", + "model.layers.44.self_attn.v_proj.weight": "model-00026-of-00051.safetensors", + "model.layers.45.input_layernorm.weight": "model-00027-of-00051.safetensors", + "model.layers.45.mlp.down_proj.weight": "model-00027-of-00051.safetensors", + "model.layers.45.mlp.gate_proj.weight": "model-00027-of-00051.safetensors", + "model.layers.45.mlp.up_proj.weight": "model-00027-of-00051.safetensors", + "model.layers.45.post_attention_layernorm.weight": "model-00027-of-00051.safetensors", + "model.layers.45.self_attn.k_proj.weight": "model-00026-of-00051.safetensors", + "model.layers.45.self_attn.o_proj.weight": "model-00026-of-00051.safetensors", + "model.layers.45.self_attn.q_proj.weight": "model-00026-of-00051.safetensors", + "model.layers.45.self_attn.v_proj.weight": "model-00026-of-00051.safetensors", + "model.layers.46.input_layernorm.weight": "model-00027-of-00051.safetensors", + "model.layers.46.mlp.down_proj.weight": "model-00027-of-00051.safetensors", + "model.layers.46.mlp.gate_proj.weight": "model-00027-of-00051.safetensors", + "model.layers.46.mlp.up_proj.weight": "model-00027-of-00051.safetensors", + "model.layers.46.post_attention_layernorm.weight": "model-00027-of-00051.safetensors", + "model.layers.46.self_attn.k_proj.weight": "model-00027-of-00051.safetensors", + "model.layers.46.self_attn.o_proj.weight": "model-00027-of-00051.safetensors", + "model.layers.46.self_attn.q_proj.weight": "model-00027-of-00051.safetensors", + "model.layers.46.self_attn.v_proj.weight": "model-00027-of-00051.safetensors", + "model.layers.47.input_layernorm.weight": "model-00028-of-00051.safetensors", + "model.layers.47.mlp.down_proj.weight": "model-00028-of-00051.safetensors", + "model.layers.47.mlp.gate_proj.weight": "model-00028-of-00051.safetensors", + "model.layers.47.mlp.up_proj.weight": "model-00028-of-00051.safetensors", + "model.layers.47.post_attention_layernorm.weight": "model-00028-of-00051.safetensors", + "model.layers.47.self_attn.k_proj.weight": "model-00028-of-00051.safetensors", + "model.layers.47.self_attn.o_proj.weight": "model-00028-of-00051.safetensors", + "model.layers.47.self_attn.q_proj.weight": "model-00028-of-00051.safetensors", + "model.layers.47.self_attn.v_proj.weight": "model-00028-of-00051.safetensors", + "model.layers.48.input_layernorm.weight": "model-00029-of-00051.safetensors", + "model.layers.48.mlp.down_proj.weight": "model-00029-of-00051.safetensors", + "model.layers.48.mlp.gate_proj.weight": "model-00028-of-00051.safetensors", + "model.layers.48.mlp.up_proj.weight": "model-00028-of-00051.safetensors", + "model.layers.48.post_attention_layernorm.weight": "model-00029-of-00051.safetensors", + "model.layers.48.self_attn.k_proj.weight": "model-00028-of-00051.safetensors", + "model.layers.48.self_attn.o_proj.weight": "model-00028-of-00051.safetensors", + "model.layers.48.self_attn.q_proj.weight": "model-00028-of-00051.safetensors", + "model.layers.48.self_attn.v_proj.weight": "model-00028-of-00051.safetensors", + "model.layers.49.input_layernorm.weight": "model-00029-of-00051.safetensors", + "model.layers.49.mlp.down_proj.weight": "model-00029-of-00051.safetensors", + "model.layers.49.mlp.gate_proj.weight": "model-00029-of-00051.safetensors", + "model.layers.49.mlp.up_proj.weight": "model-00029-of-00051.safetensors", + "model.layers.49.post_attention_layernorm.weight": "model-00029-of-00051.safetensors", + "model.layers.49.self_attn.k_proj.weight": "model-00029-of-00051.safetensors", + "model.layers.49.self_attn.o_proj.weight": "model-00029-of-00051.safetensors", + "model.layers.49.self_attn.q_proj.weight": "model-00029-of-00051.safetensors", + "model.layers.49.self_attn.v_proj.weight": "model-00029-of-00051.safetensors", + "model.layers.5.input_layernorm.weight": "model-00004-of-00051.safetensors", + "model.layers.5.mlp.down_proj.weight": "model-00004-of-00051.safetensors", + "model.layers.5.mlp.gate_proj.weight": "model-00004-of-00051.safetensors", + "model.layers.5.mlp.up_proj.weight": "model-00004-of-00051.safetensors", + "model.layers.5.post_attention_layernorm.weight": "model-00004-of-00051.safetensors", + "model.layers.5.self_attn.k_proj.weight": "model-00004-of-00051.safetensors", + "model.layers.5.self_attn.o_proj.weight": "model-00004-of-00051.safetensors", + "model.layers.5.self_attn.q_proj.weight": "model-00004-of-00051.safetensors", + "model.layers.5.self_attn.v_proj.weight": "model-00004-of-00051.safetensors", + "model.layers.50.input_layernorm.weight": "model-00030-of-00051.safetensors", + "model.layers.50.mlp.down_proj.weight": "model-00030-of-00051.safetensors", + "model.layers.50.mlp.gate_proj.weight": "model-00029-of-00051.safetensors", + "model.layers.50.mlp.up_proj.weight": "model-00030-of-00051.safetensors", + "model.layers.50.post_attention_layernorm.weight": "model-00030-of-00051.safetensors", + "model.layers.50.self_attn.k_proj.weight": "model-00029-of-00051.safetensors", + "model.layers.50.self_attn.o_proj.weight": "model-00029-of-00051.safetensors", + "model.layers.50.self_attn.q_proj.weight": "model-00029-of-00051.safetensors", + "model.layers.50.self_attn.v_proj.weight": "model-00029-of-00051.safetensors", + "model.layers.51.input_layernorm.weight": "model-00030-of-00051.safetensors", + "model.layers.51.mlp.down_proj.weight": "model-00030-of-00051.safetensors", + "model.layers.51.mlp.gate_proj.weight": "model-00030-of-00051.safetensors", + "model.layers.51.mlp.up_proj.weight": "model-00030-of-00051.safetensors", + "model.layers.51.post_attention_layernorm.weight": "model-00030-of-00051.safetensors", + "model.layers.51.self_attn.k_proj.weight": "model-00030-of-00051.safetensors", + "model.layers.51.self_attn.o_proj.weight": "model-00030-of-00051.safetensors", + "model.layers.51.self_attn.q_proj.weight": "model-00030-of-00051.safetensors", + "model.layers.51.self_attn.v_proj.weight": "model-00030-of-00051.safetensors", + "model.layers.52.input_layernorm.weight": "model-00031-of-00051.safetensors", + "model.layers.52.mlp.down_proj.weight": "model-00031-of-00051.safetensors", + "model.layers.52.mlp.gate_proj.weight": "model-00031-of-00051.safetensors", + "model.layers.52.mlp.up_proj.weight": "model-00031-of-00051.safetensors", + "model.layers.52.post_attention_layernorm.weight": "model-00031-of-00051.safetensors", + "model.layers.52.self_attn.k_proj.weight": "model-00030-of-00051.safetensors", + "model.layers.52.self_attn.o_proj.weight": "model-00030-of-00051.safetensors", + "model.layers.52.self_attn.q_proj.weight": "model-00030-of-00051.safetensors", + "model.layers.52.self_attn.v_proj.weight": "model-00030-of-00051.safetensors", + "model.layers.53.input_layernorm.weight": "model-00031-of-00051.safetensors", + "model.layers.53.mlp.down_proj.weight": "model-00031-of-00051.safetensors", + "model.layers.53.mlp.gate_proj.weight": "model-00031-of-00051.safetensors", + "model.layers.53.mlp.up_proj.weight": "model-00031-of-00051.safetensors", + "model.layers.53.post_attention_layernorm.weight": "model-00031-of-00051.safetensors", + "model.layers.53.self_attn.k_proj.weight": "model-00031-of-00051.safetensors", + "model.layers.53.self_attn.o_proj.weight": "model-00031-of-00051.safetensors", + "model.layers.53.self_attn.q_proj.weight": "model-00031-of-00051.safetensors", + "model.layers.53.self_attn.v_proj.weight": "model-00031-of-00051.safetensors", + "model.layers.54.input_layernorm.weight": "model-00032-of-00051.safetensors", + "model.layers.54.mlp.down_proj.weight": "model-00032-of-00051.safetensors", + "model.layers.54.mlp.gate_proj.weight": "model-00032-of-00051.safetensors", + "model.layers.54.mlp.up_proj.weight": "model-00032-of-00051.safetensors", + "model.layers.54.post_attention_layernorm.weight": "model-00032-of-00051.safetensors", + "model.layers.54.self_attn.k_proj.weight": "model-00032-of-00051.safetensors", + "model.layers.54.self_attn.o_proj.weight": "model-00032-of-00051.safetensors", + "model.layers.54.self_attn.q_proj.weight": "model-00032-of-00051.safetensors", + "model.layers.54.self_attn.v_proj.weight": "model-00032-of-00051.safetensors", + "model.layers.55.input_layernorm.weight": "model-00033-of-00051.safetensors", + "model.layers.55.mlp.down_proj.weight": "model-00033-of-00051.safetensors", + "model.layers.55.mlp.gate_proj.weight": "model-00032-of-00051.safetensors", + "model.layers.55.mlp.up_proj.weight": "model-00032-of-00051.safetensors", + "model.layers.55.post_attention_layernorm.weight": "model-00033-of-00051.safetensors", + "model.layers.55.self_attn.k_proj.weight": "model-00032-of-00051.safetensors", + "model.layers.55.self_attn.o_proj.weight": "model-00032-of-00051.safetensors", + "model.layers.55.self_attn.q_proj.weight": "model-00032-of-00051.safetensors", + "model.layers.55.self_attn.v_proj.weight": "model-00032-of-00051.safetensors", + "model.layers.56.input_layernorm.weight": "model-00033-of-00051.safetensors", + "model.layers.56.mlp.down_proj.weight": "model-00033-of-00051.safetensors", + "model.layers.56.mlp.gate_proj.weight": "model-00033-of-00051.safetensors", + "model.layers.56.mlp.up_proj.weight": "model-00033-of-00051.safetensors", + "model.layers.56.post_attention_layernorm.weight": "model-00033-of-00051.safetensors", + "model.layers.56.self_attn.k_proj.weight": "model-00033-of-00051.safetensors", + "model.layers.56.self_attn.o_proj.weight": "model-00033-of-00051.safetensors", + "model.layers.56.self_attn.q_proj.weight": "model-00033-of-00051.safetensors", + "model.layers.56.self_attn.v_proj.weight": "model-00033-of-00051.safetensors", + "model.layers.57.input_layernorm.weight": "model-00034-of-00051.safetensors", + "model.layers.57.mlp.down_proj.weight": "model-00034-of-00051.safetensors", + "model.layers.57.mlp.gate_proj.weight": "model-00033-of-00051.safetensors", + "model.layers.57.mlp.up_proj.weight": "model-00034-of-00051.safetensors", + "model.layers.57.post_attention_layernorm.weight": "model-00034-of-00051.safetensors", + "model.layers.57.self_attn.k_proj.weight": "model-00033-of-00051.safetensors", + "model.layers.57.self_attn.o_proj.weight": "model-00033-of-00051.safetensors", + "model.layers.57.self_attn.q_proj.weight": "model-00033-of-00051.safetensors", + "model.layers.57.self_attn.v_proj.weight": "model-00033-of-00051.safetensors", + "model.layers.58.input_layernorm.weight": "model-00034-of-00051.safetensors", + "model.layers.58.mlp.down_proj.weight": "model-00034-of-00051.safetensors", + "model.layers.58.mlp.gate_proj.weight": "model-00034-of-00051.safetensors", + "model.layers.58.mlp.up_proj.weight": "model-00034-of-00051.safetensors", + "model.layers.58.post_attention_layernorm.weight": "model-00034-of-00051.safetensors", + "model.layers.58.self_attn.k_proj.weight": "model-00034-of-00051.safetensors", + "model.layers.58.self_attn.o_proj.weight": "model-00034-of-00051.safetensors", + "model.layers.58.self_attn.q_proj.weight": "model-00034-of-00051.safetensors", + "model.layers.58.self_attn.v_proj.weight": "model-00034-of-00051.safetensors", + "model.layers.59.input_layernorm.weight": "model-00035-of-00051.safetensors", + "model.layers.59.mlp.down_proj.weight": "model-00035-of-00051.safetensors", + "model.layers.59.mlp.gate_proj.weight": "model-00035-of-00051.safetensors", + "model.layers.59.mlp.up_proj.weight": "model-00035-of-00051.safetensors", + "model.layers.59.post_attention_layernorm.weight": "model-00035-of-00051.safetensors", + "model.layers.59.self_attn.k_proj.weight": "model-00034-of-00051.safetensors", + "model.layers.59.self_attn.o_proj.weight": "model-00034-of-00051.safetensors", + "model.layers.59.self_attn.q_proj.weight": "model-00034-of-00051.safetensors", + "model.layers.59.self_attn.v_proj.weight": "model-00034-of-00051.safetensors", + "model.layers.6.input_layernorm.weight": "model-00005-of-00051.safetensors", + "model.layers.6.mlp.down_proj.weight": "model-00005-of-00051.safetensors", + "model.layers.6.mlp.gate_proj.weight": "model-00004-of-00051.safetensors", + "model.layers.6.mlp.up_proj.weight": "model-00004-of-00051.safetensors", + "model.layers.6.post_attention_layernorm.weight": "model-00005-of-00051.safetensors", + "model.layers.6.self_attn.k_proj.weight": "model-00004-of-00051.safetensors", + "model.layers.6.self_attn.o_proj.weight": "model-00004-of-00051.safetensors", + "model.layers.6.self_attn.q_proj.weight": "model-00004-of-00051.safetensors", + "model.layers.6.self_attn.v_proj.weight": "model-00004-of-00051.safetensors", + "model.layers.60.input_layernorm.weight": "model-00035-of-00051.safetensors", + "model.layers.60.mlp.down_proj.weight": "model-00035-of-00051.safetensors", + "model.layers.60.mlp.gate_proj.weight": "model-00035-of-00051.safetensors", + "model.layers.60.mlp.up_proj.weight": "model-00035-of-00051.safetensors", + "model.layers.60.post_attention_layernorm.weight": "model-00035-of-00051.safetensors", + "model.layers.60.self_attn.k_proj.weight": "model-00035-of-00051.safetensors", + "model.layers.60.self_attn.o_proj.weight": "model-00035-of-00051.safetensors", + "model.layers.60.self_attn.q_proj.weight": "model-00035-of-00051.safetensors", + "model.layers.60.self_attn.v_proj.weight": "model-00035-of-00051.safetensors", + "model.layers.61.input_layernorm.weight": "model-00036-of-00051.safetensors", + "model.layers.61.mlp.down_proj.weight": "model-00036-of-00051.safetensors", + "model.layers.61.mlp.gate_proj.weight": "model-00036-of-00051.safetensors", + "model.layers.61.mlp.up_proj.weight": "model-00036-of-00051.safetensors", + "model.layers.61.post_attention_layernorm.weight": "model-00036-of-00051.safetensors", + "model.layers.61.self_attn.k_proj.weight": "model-00036-of-00051.safetensors", + "model.layers.61.self_attn.o_proj.weight": "model-00036-of-00051.safetensors", + "model.layers.61.self_attn.q_proj.weight": "model-00036-of-00051.safetensors", + "model.layers.61.self_attn.v_proj.weight": "model-00036-of-00051.safetensors", + "model.layers.62.input_layernorm.weight": "model-00037-of-00051.safetensors", + "model.layers.62.mlp.down_proj.weight": "model-00037-of-00051.safetensors", + "model.layers.62.mlp.gate_proj.weight": "model-00036-of-00051.safetensors", + "model.layers.62.mlp.up_proj.weight": "model-00036-of-00051.safetensors", + "model.layers.62.post_attention_layernorm.weight": "model-00037-of-00051.safetensors", + "model.layers.62.self_attn.k_proj.weight": "model-00036-of-00051.safetensors", + "model.layers.62.self_attn.o_proj.weight": "model-00036-of-00051.safetensors", + "model.layers.62.self_attn.q_proj.weight": "model-00036-of-00051.safetensors", + "model.layers.62.self_attn.v_proj.weight": "model-00036-of-00051.safetensors", + "model.layers.63.input_layernorm.weight": "model-00037-of-00051.safetensors", + "model.layers.63.mlp.down_proj.weight": "model-00037-of-00051.safetensors", + "model.layers.63.mlp.gate_proj.weight": "model-00037-of-00051.safetensors", + "model.layers.63.mlp.up_proj.weight": "model-00037-of-00051.safetensors", + "model.layers.63.post_attention_layernorm.weight": "model-00037-of-00051.safetensors", + "model.layers.63.self_attn.k_proj.weight": "model-00037-of-00051.safetensors", + "model.layers.63.self_attn.o_proj.weight": "model-00037-of-00051.safetensors", + "model.layers.63.self_attn.q_proj.weight": "model-00037-of-00051.safetensors", + "model.layers.63.self_attn.v_proj.weight": "model-00037-of-00051.safetensors", + "model.layers.64.input_layernorm.weight": "model-00038-of-00051.safetensors", + "model.layers.64.mlp.down_proj.weight": "model-00038-of-00051.safetensors", + "model.layers.64.mlp.gate_proj.weight": "model-00037-of-00051.safetensors", + "model.layers.64.mlp.up_proj.weight": "model-00038-of-00051.safetensors", + "model.layers.64.post_attention_layernorm.weight": "model-00038-of-00051.safetensors", + "model.layers.64.self_attn.k_proj.weight": "model-00037-of-00051.safetensors", + "model.layers.64.self_attn.o_proj.weight": "model-00037-of-00051.safetensors", + "model.layers.64.self_attn.q_proj.weight": "model-00037-of-00051.safetensors", + "model.layers.64.self_attn.v_proj.weight": "model-00037-of-00051.safetensors", + "model.layers.65.input_layernorm.weight": "model-00038-of-00051.safetensors", + "model.layers.65.mlp.down_proj.weight": "model-00038-of-00051.safetensors", + "model.layers.65.mlp.gate_proj.weight": "model-00038-of-00051.safetensors", + "model.layers.65.mlp.up_proj.weight": "model-00038-of-00051.safetensors", + "model.layers.65.post_attention_layernorm.weight": "model-00038-of-00051.safetensors", + "model.layers.65.self_attn.k_proj.weight": "model-00038-of-00051.safetensors", + "model.layers.65.self_attn.o_proj.weight": "model-00038-of-00051.safetensors", + "model.layers.65.self_attn.q_proj.weight": "model-00038-of-00051.safetensors", + "model.layers.65.self_attn.v_proj.weight": "model-00038-of-00051.safetensors", + "model.layers.66.input_layernorm.weight": "model-00039-of-00051.safetensors", + "model.layers.66.mlp.down_proj.weight": "model-00039-of-00051.safetensors", + "model.layers.66.mlp.gate_proj.weight": "model-00039-of-00051.safetensors", + "model.layers.66.mlp.up_proj.weight": "model-00039-of-00051.safetensors", + "model.layers.66.post_attention_layernorm.weight": "model-00039-of-00051.safetensors", + "model.layers.66.self_attn.k_proj.weight": "model-00038-of-00051.safetensors", + "model.layers.66.self_attn.o_proj.weight": "model-00038-of-00051.safetensors", + "model.layers.66.self_attn.q_proj.weight": "model-00038-of-00051.safetensors", + "model.layers.66.self_attn.v_proj.weight": "model-00038-of-00051.safetensors", + "model.layers.67.input_layernorm.weight": "model-00039-of-00051.safetensors", + "model.layers.67.mlp.down_proj.weight": "model-00039-of-00051.safetensors", + "model.layers.67.mlp.gate_proj.weight": "model-00039-of-00051.safetensors", + "model.layers.67.mlp.up_proj.weight": "model-00039-of-00051.safetensors", + "model.layers.67.post_attention_layernorm.weight": "model-00039-of-00051.safetensors", + "model.layers.67.self_attn.k_proj.weight": "model-00039-of-00051.safetensors", + "model.layers.67.self_attn.o_proj.weight": "model-00039-of-00051.safetensors", + "model.layers.67.self_attn.q_proj.weight": "model-00039-of-00051.safetensors", + "model.layers.67.self_attn.v_proj.weight": "model-00039-of-00051.safetensors", + "model.layers.68.input_layernorm.weight": "model-00040-of-00051.safetensors", + "model.layers.68.mlp.down_proj.weight": "model-00040-of-00051.safetensors", + "model.layers.68.mlp.gate_proj.weight": "model-00040-of-00051.safetensors", + "model.layers.68.mlp.up_proj.weight": "model-00040-of-00051.safetensors", + "model.layers.68.post_attention_layernorm.weight": "model-00040-of-00051.safetensors", + "model.layers.68.self_attn.k_proj.weight": "model-00040-of-00051.safetensors", + "model.layers.68.self_attn.o_proj.weight": "model-00040-of-00051.safetensors", + "model.layers.68.self_attn.q_proj.weight": "model-00040-of-00051.safetensors", + "model.layers.68.self_attn.v_proj.weight": "model-00040-of-00051.safetensors", + "model.layers.69.input_layernorm.weight": "model-00041-of-00051.safetensors", + "model.layers.69.mlp.down_proj.weight": "model-00041-of-00051.safetensors", + "model.layers.69.mlp.gate_proj.weight": "model-00040-of-00051.safetensors", + "model.layers.69.mlp.up_proj.weight": "model-00040-of-00051.safetensors", + "model.layers.69.post_attention_layernorm.weight": "model-00041-of-00051.safetensors", + "model.layers.69.self_attn.k_proj.weight": "model-00040-of-00051.safetensors", + "model.layers.69.self_attn.o_proj.weight": "model-00040-of-00051.safetensors", + "model.layers.69.self_attn.q_proj.weight": "model-00040-of-00051.safetensors", + "model.layers.69.self_attn.v_proj.weight": "model-00040-of-00051.safetensors", + "model.layers.7.input_layernorm.weight": "model-00005-of-00051.safetensors", + "model.layers.7.mlp.down_proj.weight": "model-00005-of-00051.safetensors", + "model.layers.7.mlp.gate_proj.weight": "model-00005-of-00051.safetensors", + "model.layers.7.mlp.up_proj.weight": "model-00005-of-00051.safetensors", + "model.layers.7.post_attention_layernorm.weight": "model-00005-of-00051.safetensors", + "model.layers.7.self_attn.k_proj.weight": "model-00005-of-00051.safetensors", + "model.layers.7.self_attn.o_proj.weight": "model-00005-of-00051.safetensors", + "model.layers.7.self_attn.q_proj.weight": "model-00005-of-00051.safetensors", + "model.layers.7.self_attn.v_proj.weight": "model-00005-of-00051.safetensors", + "model.layers.70.input_layernorm.weight": "model-00041-of-00051.safetensors", + "model.layers.70.mlp.down_proj.weight": "model-00041-of-00051.safetensors", + "model.layers.70.mlp.gate_proj.weight": "model-00041-of-00051.safetensors", + "model.layers.70.mlp.up_proj.weight": "model-00041-of-00051.safetensors", + "model.layers.70.post_attention_layernorm.weight": "model-00041-of-00051.safetensors", + "model.layers.70.self_attn.k_proj.weight": "model-00041-of-00051.safetensors", + "model.layers.70.self_attn.o_proj.weight": "model-00041-of-00051.safetensors", + "model.layers.70.self_attn.q_proj.weight": "model-00041-of-00051.safetensors", + "model.layers.70.self_attn.v_proj.weight": "model-00041-of-00051.safetensors", + "model.layers.71.input_layernorm.weight": "model-00042-of-00051.safetensors", + "model.layers.71.mlp.down_proj.weight": "model-00042-of-00051.safetensors", + "model.layers.71.mlp.gate_proj.weight": "model-00041-of-00051.safetensors", + "model.layers.71.mlp.up_proj.weight": "model-00042-of-00051.safetensors", + "model.layers.71.post_attention_layernorm.weight": "model-00042-of-00051.safetensors", + "model.layers.71.self_attn.k_proj.weight": "model-00041-of-00051.safetensors", + "model.layers.71.self_attn.o_proj.weight": "model-00041-of-00051.safetensors", + "model.layers.71.self_attn.q_proj.weight": "model-00041-of-00051.safetensors", + "model.layers.71.self_attn.v_proj.weight": "model-00041-of-00051.safetensors", + "model.layers.72.input_layernorm.weight": "model-00042-of-00051.safetensors", + "model.layers.72.mlp.down_proj.weight": "model-00042-of-00051.safetensors", + "model.layers.72.mlp.gate_proj.weight": "model-00042-of-00051.safetensors", + "model.layers.72.mlp.up_proj.weight": "model-00042-of-00051.safetensors", + "model.layers.72.post_attention_layernorm.weight": "model-00042-of-00051.safetensors", + "model.layers.72.self_attn.k_proj.weight": "model-00042-of-00051.safetensors", + "model.layers.72.self_attn.o_proj.weight": "model-00042-of-00051.safetensors", + "model.layers.72.self_attn.q_proj.weight": "model-00042-of-00051.safetensors", + "model.layers.72.self_attn.v_proj.weight": "model-00042-of-00051.safetensors", + "model.layers.73.input_layernorm.weight": "model-00043-of-00051.safetensors", + "model.layers.73.mlp.down_proj.weight": "model-00043-of-00051.safetensors", + "model.layers.73.mlp.gate_proj.weight": "model-00043-of-00051.safetensors", + "model.layers.73.mlp.up_proj.weight": "model-00043-of-00051.safetensors", + "model.layers.73.post_attention_layernorm.weight": "model-00043-of-00051.safetensors", + "model.layers.73.self_attn.k_proj.weight": "model-00042-of-00051.safetensors", + "model.layers.73.self_attn.o_proj.weight": "model-00042-of-00051.safetensors", + "model.layers.73.self_attn.q_proj.weight": "model-00042-of-00051.safetensors", + "model.layers.73.self_attn.v_proj.weight": "model-00042-of-00051.safetensors", + "model.layers.74.input_layernorm.weight": "model-00043-of-00051.safetensors", + "model.layers.74.mlp.down_proj.weight": "model-00043-of-00051.safetensors", + "model.layers.74.mlp.gate_proj.weight": "model-00043-of-00051.safetensors", + "model.layers.74.mlp.up_proj.weight": "model-00043-of-00051.safetensors", + "model.layers.74.post_attention_layernorm.weight": "model-00043-of-00051.safetensors", + "model.layers.74.self_attn.k_proj.weight": "model-00043-of-00051.safetensors", + "model.layers.74.self_attn.o_proj.weight": "model-00043-of-00051.safetensors", + "model.layers.74.self_attn.q_proj.weight": "model-00043-of-00051.safetensors", + "model.layers.74.self_attn.v_proj.weight": "model-00043-of-00051.safetensors", + "model.layers.75.input_layernorm.weight": "model-00044-of-00051.safetensors", + "model.layers.75.mlp.down_proj.weight": "model-00044-of-00051.safetensors", + "model.layers.75.mlp.gate_proj.weight": "model-00044-of-00051.safetensors", + "model.layers.75.mlp.up_proj.weight": "model-00044-of-00051.safetensors", + "model.layers.75.post_attention_layernorm.weight": "model-00044-of-00051.safetensors", + "model.layers.75.self_attn.k_proj.weight": "model-00044-of-00051.safetensors", + "model.layers.75.self_attn.o_proj.weight": "model-00044-of-00051.safetensors", + "model.layers.75.self_attn.q_proj.weight": "model-00044-of-00051.safetensors", + "model.layers.75.self_attn.v_proj.weight": "model-00044-of-00051.safetensors", + "model.layers.76.input_layernorm.weight": "model-00045-of-00051.safetensors", + "model.layers.76.mlp.down_proj.weight": "model-00045-of-00051.safetensors", + "model.layers.76.mlp.gate_proj.weight": "model-00044-of-00051.safetensors", + "model.layers.76.mlp.up_proj.weight": "model-00044-of-00051.safetensors", + "model.layers.76.post_attention_layernorm.weight": "model-00045-of-00051.safetensors", + "model.layers.76.self_attn.k_proj.weight": "model-00044-of-00051.safetensors", + "model.layers.76.self_attn.o_proj.weight": "model-00044-of-00051.safetensors", + "model.layers.76.self_attn.q_proj.weight": "model-00044-of-00051.safetensors", + "model.layers.76.self_attn.v_proj.weight": "model-00044-of-00051.safetensors", + "model.layers.77.input_layernorm.weight": "model-00045-of-00051.safetensors", + "model.layers.77.mlp.down_proj.weight": "model-00045-of-00051.safetensors", + "model.layers.77.mlp.gate_proj.weight": "model-00045-of-00051.safetensors", + "model.layers.77.mlp.up_proj.weight": "model-00045-of-00051.safetensors", + "model.layers.77.post_attention_layernorm.weight": "model-00045-of-00051.safetensors", + "model.layers.77.self_attn.k_proj.weight": "model-00045-of-00051.safetensors", + "model.layers.77.self_attn.o_proj.weight": "model-00045-of-00051.safetensors", + "model.layers.77.self_attn.q_proj.weight": "model-00045-of-00051.safetensors", + "model.layers.77.self_attn.v_proj.weight": "model-00045-of-00051.safetensors", + "model.layers.78.input_layernorm.weight": "model-00046-of-00051.safetensors", + "model.layers.78.mlp.down_proj.weight": "model-00046-of-00051.safetensors", + "model.layers.78.mlp.gate_proj.weight": "model-00045-of-00051.safetensors", + "model.layers.78.mlp.up_proj.weight": "model-00046-of-00051.safetensors", + "model.layers.78.post_attention_layernorm.weight": "model-00046-of-00051.safetensors", + "model.layers.78.self_attn.k_proj.weight": "model-00045-of-00051.safetensors", + "model.layers.78.self_attn.o_proj.weight": "model-00045-of-00051.safetensors", + "model.layers.78.self_attn.q_proj.weight": "model-00045-of-00051.safetensors", + "model.layers.78.self_attn.v_proj.weight": "model-00045-of-00051.safetensors", + "model.layers.79.input_layernorm.weight": "model-00046-of-00051.safetensors", + "model.layers.79.mlp.down_proj.weight": "model-00046-of-00051.safetensors", + "model.layers.79.mlp.gate_proj.weight": "model-00046-of-00051.safetensors", + "model.layers.79.mlp.up_proj.weight": "model-00046-of-00051.safetensors", + "model.layers.79.post_attention_layernorm.weight": "model-00046-of-00051.safetensors", + "model.layers.79.self_attn.k_proj.weight": "model-00046-of-00051.safetensors", + "model.layers.79.self_attn.o_proj.weight": "model-00046-of-00051.safetensors", + "model.layers.79.self_attn.q_proj.weight": "model-00046-of-00051.safetensors", + "model.layers.79.self_attn.v_proj.weight": "model-00046-of-00051.safetensors", + "model.layers.8.input_layernorm.weight": "model-00006-of-00051.safetensors", + "model.layers.8.mlp.down_proj.weight": "model-00006-of-00051.safetensors", + "model.layers.8.mlp.gate_proj.weight": "model-00005-of-00051.safetensors", + "model.layers.8.mlp.up_proj.weight": "model-00006-of-00051.safetensors", + "model.layers.8.post_attention_layernorm.weight": "model-00006-of-00051.safetensors", + "model.layers.8.self_attn.k_proj.weight": "model-00005-of-00051.safetensors", + "model.layers.8.self_attn.o_proj.weight": "model-00005-of-00051.safetensors", + "model.layers.8.self_attn.q_proj.weight": "model-00005-of-00051.safetensors", + "model.layers.8.self_attn.v_proj.weight": "model-00005-of-00051.safetensors", + "model.layers.80.input_layernorm.weight": "model-00047-of-00051.safetensors", + "model.layers.80.mlp.down_proj.weight": "model-00047-of-00051.safetensors", + "model.layers.80.mlp.gate_proj.weight": "model-00047-of-00051.safetensors", + "model.layers.80.mlp.up_proj.weight": "model-00047-of-00051.safetensors", + "model.layers.80.post_attention_layernorm.weight": "model-00047-of-00051.safetensors", + "model.layers.80.self_attn.k_proj.weight": "model-00046-of-00051.safetensors", + "model.layers.80.self_attn.o_proj.weight": "model-00046-of-00051.safetensors", + "model.layers.80.self_attn.q_proj.weight": "model-00046-of-00051.safetensors", + "model.layers.80.self_attn.v_proj.weight": "model-00046-of-00051.safetensors", + "model.layers.81.input_layernorm.weight": "model-00047-of-00051.safetensors", + "model.layers.81.mlp.down_proj.weight": "model-00047-of-00051.safetensors", + "model.layers.81.mlp.gate_proj.weight": "model-00047-of-00051.safetensors", + "model.layers.81.mlp.up_proj.weight": "model-00047-of-00051.safetensors", + "model.layers.81.post_attention_layernorm.weight": "model-00047-of-00051.safetensors", + "model.layers.81.self_attn.k_proj.weight": "model-00047-of-00051.safetensors", + "model.layers.81.self_attn.o_proj.weight": "model-00047-of-00051.safetensors", + "model.layers.81.self_attn.q_proj.weight": "model-00047-of-00051.safetensors", + "model.layers.81.self_attn.v_proj.weight": "model-00047-of-00051.safetensors", + "model.layers.82.input_layernorm.weight": "model-00048-of-00051.safetensors", + "model.layers.82.mlp.down_proj.weight": "model-00048-of-00051.safetensors", + "model.layers.82.mlp.gate_proj.weight": "model-00048-of-00051.safetensors", + "model.layers.82.mlp.up_proj.weight": "model-00048-of-00051.safetensors", + "model.layers.82.post_attention_layernorm.weight": "model-00048-of-00051.safetensors", + "model.layers.82.self_attn.k_proj.weight": "model-00048-of-00051.safetensors", + "model.layers.82.self_attn.o_proj.weight": "model-00048-of-00051.safetensors", + "model.layers.82.self_attn.q_proj.weight": "model-00048-of-00051.safetensors", + "model.layers.82.self_attn.v_proj.weight": "model-00048-of-00051.safetensors", + "model.layers.83.input_layernorm.weight": "model-00049-of-00051.safetensors", + "model.layers.83.mlp.down_proj.weight": "model-00049-of-00051.safetensors", + "model.layers.83.mlp.gate_proj.weight": "model-00048-of-00051.safetensors", + "model.layers.83.mlp.up_proj.weight": "model-00048-of-00051.safetensors", + "model.layers.83.post_attention_layernorm.weight": "model-00049-of-00051.safetensors", + "model.layers.83.self_attn.k_proj.weight": "model-00048-of-00051.safetensors", + "model.layers.83.self_attn.o_proj.weight": "model-00048-of-00051.safetensors", + "model.layers.83.self_attn.q_proj.weight": "model-00048-of-00051.safetensors", + "model.layers.83.self_attn.v_proj.weight": "model-00048-of-00051.safetensors", + "model.layers.84.input_layernorm.weight": "model-00049-of-00051.safetensors", + "model.layers.84.mlp.down_proj.weight": "model-00049-of-00051.safetensors", + "model.layers.84.mlp.gate_proj.weight": "model-00049-of-00051.safetensors", + "model.layers.84.mlp.up_proj.weight": "model-00049-of-00051.safetensors", + "model.layers.84.post_attention_layernorm.weight": "model-00049-of-00051.safetensors", + "model.layers.84.self_attn.k_proj.weight": "model-00049-of-00051.safetensors", + "model.layers.84.self_attn.o_proj.weight": "model-00049-of-00051.safetensors", + "model.layers.84.self_attn.q_proj.weight": "model-00049-of-00051.safetensors", + "model.layers.84.self_attn.v_proj.weight": "model-00049-of-00051.safetensors", + "model.layers.85.input_layernorm.weight": "model-00050-of-00051.safetensors", + "model.layers.85.mlp.down_proj.weight": "model-00050-of-00051.safetensors", + "model.layers.85.mlp.gate_proj.weight": "model-00049-of-00051.safetensors", + "model.layers.85.mlp.up_proj.weight": "model-00050-of-00051.safetensors", + "model.layers.85.post_attention_layernorm.weight": "model-00050-of-00051.safetensors", + "model.layers.85.self_attn.k_proj.weight": "model-00049-of-00051.safetensors", + "model.layers.85.self_attn.o_proj.weight": "model-00049-of-00051.safetensors", + "model.layers.85.self_attn.q_proj.weight": "model-00049-of-00051.safetensors", + "model.layers.85.self_attn.v_proj.weight": "model-00049-of-00051.safetensors", + "model.layers.86.input_layernorm.weight": "model-00050-of-00051.safetensors", + "model.layers.86.mlp.down_proj.weight": "model-00050-of-00051.safetensors", + "model.layers.86.mlp.gate_proj.weight": "model-00050-of-00051.safetensors", + "model.layers.86.mlp.up_proj.weight": "model-00050-of-00051.safetensors", + "model.layers.86.post_attention_layernorm.weight": "model-00050-of-00051.safetensors", + "model.layers.86.self_attn.k_proj.weight": "model-00050-of-00051.safetensors", + "model.layers.86.self_attn.o_proj.weight": "model-00050-of-00051.safetensors", + "model.layers.86.self_attn.q_proj.weight": "model-00050-of-00051.safetensors", + "model.layers.86.self_attn.v_proj.weight": "model-00050-of-00051.safetensors", + "model.layers.87.input_layernorm.weight": "model-00051-of-00051.safetensors", + "model.layers.87.mlp.down_proj.weight": "model-00051-of-00051.safetensors", + "model.layers.87.mlp.gate_proj.weight": "model-00051-of-00051.safetensors", + "model.layers.87.mlp.up_proj.weight": "model-00051-of-00051.safetensors", + "model.layers.87.post_attention_layernorm.weight": "model-00051-of-00051.safetensors", + "model.layers.87.self_attn.k_proj.weight": "model-00050-of-00051.safetensors", + "model.layers.87.self_attn.o_proj.weight": "model-00050-of-00051.safetensors", + "model.layers.87.self_attn.q_proj.weight": "model-00050-of-00051.safetensors", + "model.layers.87.self_attn.v_proj.weight": "model-00050-of-00051.safetensors", + "model.layers.9.input_layernorm.weight": "model-00006-of-00051.safetensors", + "model.layers.9.mlp.down_proj.weight": "model-00006-of-00051.safetensors", + "model.layers.9.mlp.gate_proj.weight": "model-00006-of-00051.safetensors", + "model.layers.9.mlp.up_proj.weight": "model-00006-of-00051.safetensors", + "model.layers.9.post_attention_layernorm.weight": "model-00006-of-00051.safetensors", + "model.layers.9.self_attn.k_proj.weight": "model-00006-of-00051.safetensors", + "model.layers.9.self_attn.o_proj.weight": "model-00006-of-00051.safetensors", + "model.layers.9.self_attn.q_proj.weight": "model-00006-of-00051.safetensors", + "model.layers.9.self_attn.v_proj.weight": "model-00006-of-00051.safetensors", + "model.norm.weight": "model-00051-of-00051.safetensors" + } +} \ No newline at end of file