diff --git a/model-00001-of-00050.safetensors b/model-00001-of-00050.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..2894daf196322833b0d4cbe1b052ea2e75c66f5b --- /dev/null +++ b/model-00001-of-00050.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:4d5e44cd8ecfe1307f6c8463c130bfd762708453b948e71f35736cb585eafc56 +size 4718709464 diff --git a/model-00002-of-00050.safetensors b/model-00002-of-00050.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..63471c93df2eefe8c2f0a4d687ebea0bb5fa5e7d --- /dev/null +++ b/model-00002-of-00050.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:c3c46b505d7d6d492316609bc74d02558a8c389c5f25f2fe951567052a9be32f +size 4664167352 diff --git a/model-00003-of-00050.safetensors b/model-00003-of-00050.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..30638c0ee74feba929e17035229021c48e9a38b4 --- /dev/null +++ b/model-00003-of-00050.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:4edb0a1c2f503cecb91cb823eb954dcd01295a6f43e8770b13c739e31b9c1543 +size 4999711672 diff --git a/model-00004-of-00050.safetensors b/model-00004-of-00050.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..33e9d8d4ee78fe9765f01beb2926a4b172bfb393 --- /dev/null +++ b/model-00004-of-00050.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:c79f34709389b97d11086596f6771085a0ae43754f7ce4b23a87a0812e2cdb24 +size 4966157024 diff --git a/model-00005-of-00050.safetensors b/model-00005-of-00050.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..929544a4f509785fe944cadc9d252032af002970 --- /dev/null +++ b/model-00005-of-00050.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:feec024cad73e752cab21b2eb06323383dfde0f8b2ecc1fffda5b32aa5976c96 +size 4664134384 diff --git a/model-00006-of-00050.safetensors b/model-00006-of-00050.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..9767e224247bd8643c5bafccffb4c18f5c67dcbd --- /dev/null +++ b/model-00006-of-00050.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:2db6da6d8354ae232b956636435e6e2cb698901c5e0fae113c8171a02f5292d0 +size 4664167384 diff --git a/model-00007-of-00050.safetensors b/model-00007-of-00050.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..715b584c14c3d0056848b1c4b504c82a1eea77da --- /dev/null +++ b/model-00007-of-00050.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:6cb1385b1a748f5cad7874b0e0fc7ddacd52a7a23a4bae2cdeeb5247a0ba8bfc +size 4664167376 diff --git a/model-00008-of-00050.safetensors b/model-00008-of-00050.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..a0eef94ebe122488c3485de4a6531bc93ee45a47 --- /dev/null +++ b/model-00008-of-00050.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:d9c21f68a722cffb10d37575a8e467da1255db6764c62395c29c28ac5b6224b4 +size 4999711704 diff --git a/model-00009-of-00050.safetensors b/model-00009-of-00050.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..a78b701c81b8413ea0dcdf36e83f526a0cba64eb --- /dev/null +++ b/model-00009-of-00050.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:d158bb1e8f9bdd0d8cc1ace361c5c11b8be563c96a540be176bd858f03f565c2 +size 4966157032 diff --git a/model-00010-of-00050.safetensors b/model-00010-of-00050.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..cddebe3c11210efb8cca1c26e90d895aa7b0d1ea --- /dev/null +++ b/model-00010-of-00050.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:ee70b85395bfcfb8f8453a05bf73037fe81e8bfc4c964281cc8c3c4df28bfdb9 +size 4664134384 diff --git a/model-00011-of-00050.safetensors b/model-00011-of-00050.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..a5b07121760b8b26a2588438438963ab07ac4ef1 --- /dev/null +++ b/model-00011-of-00050.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:000d4e3ac64e37c81f9605221b7f50ef23a7ede78bd3fb5f170e86ab6c125bd2 +size 4664167384 diff --git a/model-00012-of-00050.safetensors b/model-00012-of-00050.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..fab6338df17973e606bb823979d61a73888e3600 --- /dev/null +++ b/model-00012-of-00050.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:840a219e3bdd7cf498db8b9a7bd46af1321ee11cfc948cdf77533fe07d258e42 +size 4664167376 diff --git a/model-00013-of-00050.safetensors b/model-00013-of-00050.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..86ace56d1be96bc2c0e0f84b024ad02c67635614 --- /dev/null +++ b/model-00013-of-00050.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:68b3ee33ac6be5d1dc71d4f4492001e458dee0ee345ea69f75506c3ee8662663 +size 4999711704 diff --git a/model-00014-of-00050.safetensors b/model-00014-of-00050.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..517c9b4311c27a25462a7eecfa6bf4fa7a1b9a12 --- /dev/null +++ b/model-00014-of-00050.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:ecec581b223d863734476358cf5aa3f77e35e5da05cce1f2b605a0f47038b122 +size 4966157032 diff --git a/model-00015-of-00050.safetensors b/model-00015-of-00050.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..d201bad1352373f1c3b3d19a91941f48c2f670a1 --- /dev/null +++ b/model-00015-of-00050.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:ef56268a473e0fcc265678b1e05bef108c6a3878bc92c01cd55568180cc20d0b +size 4664134384 diff --git a/model-00016-of-00050.safetensors b/model-00016-of-00050.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..602dbc76854ed87ffce3d399b62c14a1768d4417 --- /dev/null +++ b/model-00016-of-00050.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:d9fb0052fe593fd03b4dc2dfc079874780b41d26c24c9919628f323068b56a8f +size 4664167384 diff --git a/model-00017-of-00050.safetensors b/model-00017-of-00050.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..36998b4c0006a616c5a42188b243d99f32936dca --- /dev/null +++ b/model-00017-of-00050.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:45917640b2d9f4d2e7f368d1f84c2a9afc0aae65034b7d2f8ae5a0dea3510f55 +size 4664167376 diff --git a/model-00018-of-00050.safetensors b/model-00018-of-00050.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..dba562ec84f25494ac3ce0c6ba5e4fe0f1e1b093 --- /dev/null +++ b/model-00018-of-00050.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:f68124873a939ab752adcd0d90a6c3aefadc32af2f4a04f8f5392c47c02fe582 +size 4999711704 diff --git a/model-00019-of-00050.safetensors b/model-00019-of-00050.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..49e308c459229b07c017645b33d22808ffdcc9c6 --- /dev/null +++ b/model-00019-of-00050.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:2eb64814dcf480926bcf76b2a916b7f059c6cf215844ec190b7a1bc951fe2b99 +size 4966157032 diff --git a/model-00020-of-00050.safetensors b/model-00020-of-00050.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..d13af1876cbf3861570c13400c3ae5d49a86ed80 --- /dev/null +++ b/model-00020-of-00050.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:9fda2dff07bdac43d73ea9c778754f87506499ca17f4733c681e27a4082e7189 +size 4664134384 diff --git a/model-00021-of-00050.safetensors b/model-00021-of-00050.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..6eef2094fcca0f80847338f020a26c082bd96813 --- /dev/null +++ b/model-00021-of-00050.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:42b0bd113c9cd620ec08d1c585e51a2f3cf2a2006611ed7ebfb729d7e61b6b10 +size 4664167384 diff --git a/model-00022-of-00050.safetensors b/model-00022-of-00050.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..84643d5b8c1e87d292b96be93bd142d5cc240160 --- /dev/null +++ b/model-00022-of-00050.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:37766d9dd08ae8e2040c7fcb6881218e1d58e891674d0aac62d449c7461dafe6 +size 4664167376 diff --git a/model-00023-of-00050.safetensors b/model-00023-of-00050.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..9f76a1bb6854418500f79449daadec018610a11b --- /dev/null +++ b/model-00023-of-00050.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:1b96c640c7a67bd5603f7a2c7b1d8c52b86dcfca81a171b8a3c5e3eb8d9d49ec +size 4999711704 diff --git a/model-00024-of-00050.safetensors b/model-00024-of-00050.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..3ca41737ab3371cfafb6c15a40395e276cd4ccaf --- /dev/null +++ b/model-00024-of-00050.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:780e27aadbc0a671ccd922d510c4ae7e0d71388daad216ab5d9e086e53f9d053 +size 4966157032 diff --git a/model-00025-of-00050.safetensors b/model-00025-of-00050.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..a411a3d64c7d7d9321b783d808678fbd2ebb0b4b --- /dev/null +++ b/model-00025-of-00050.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:fa36b232509cfaff206f62756701af1dd3a654a1a3f0e6a0a0430a75a4167f01 +size 4664134384 diff --git a/model-00026-of-00050.safetensors b/model-00026-of-00050.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..2446612fd2e2df9a6a605f8e0bd4d447bf189cd1 --- /dev/null +++ b/model-00026-of-00050.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:850886c6318362fc517ffc2bb723319de7d66059425bae637b06e23ca3d9f534 +size 4664167384 diff --git a/model-00027-of-00050.safetensors b/model-00027-of-00050.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..cf51b879c1a6a53e394eb0123c78ce6d01692624 --- /dev/null +++ b/model-00027-of-00050.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:342ac104344f69cb063885f4e722ab6d0c2ac9a5a51c35e8b28139703d2173a9 +size 4664167376 diff --git a/model-00028-of-00050.safetensors b/model-00028-of-00050.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..47117a6e7de282e2274fdf9fd5d24b74f44a51ec --- /dev/null +++ b/model-00028-of-00050.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:bb4787a39fe9b50e8810338aaa3c44794877b6a097648af0b7caefe160c6836a +size 4999711704 diff --git a/model-00029-of-00050.safetensors b/model-00029-of-00050.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..22add2b6c847bb2612e826869e463e114819242d --- /dev/null +++ b/model-00029-of-00050.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:56d56d802cca43435985805a9fc466cbb39df33896c0521a1eb8c081fcf8a72d +size 4966157032 diff --git a/model-00030-of-00050.safetensors b/model-00030-of-00050.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..dbadfa1fe3c2817713686621f8781f7f20811f1f --- /dev/null +++ b/model-00030-of-00050.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:13f10d4008ee61ea47a1f80f1dcfaae35c478086ddea843f8c022163e0391869 +size 4664134384 diff --git a/model-00031-of-00050.safetensors b/model-00031-of-00050.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..daa9406fb90226fe8cd9c16bd285db787294fd88 --- /dev/null +++ b/model-00031-of-00050.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:e200f64200dbe42df0c01accff5d8cb17483a71b1949526f879d00a0d86c642e +size 4664167384 diff --git a/model-00032-of-00050.safetensors b/model-00032-of-00050.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..8a3a808fc5d56f90048801853e14e1afc0acc08c --- /dev/null +++ b/model-00032-of-00050.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:d2aa4174abc6b8409b23cd4a1bf64c3d8776c668cc92db9f0784704020a36970 +size 4664167376 diff --git a/model-00033-of-00050.safetensors b/model-00033-of-00050.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..e27d6ae6c3ea431792bb3a8b8160d3551782c20a --- /dev/null +++ b/model-00033-of-00050.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:83f47e0dca7dc55b1af5eed0dd3baf0e44a2ba07956c97a90a3a1a86e1ba3392 +size 4999711704 diff --git a/model-00034-of-00050.safetensors b/model-00034-of-00050.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..9f055f10371faa3027d3d8f7037faa4dbd6fbd71 --- /dev/null +++ b/model-00034-of-00050.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:450404144b8414bfa0a950bd22ea620ba23611290afb2e4e61ba5b91b8c8b778 +size 4966157032 diff --git a/model-00035-of-00050.safetensors b/model-00035-of-00050.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..dab732daa9f62e2687c0aea5cfc3838d5d8605e4 --- /dev/null +++ b/model-00035-of-00050.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:37414684adf9802e7e2e40c53b37d39fdda91e82956668884bdd2899e8e6fa9c +size 4664134384 diff --git a/model-00036-of-00050.safetensors b/model-00036-of-00050.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..c0bacecad3e1dba4dd21b0ffe49b9e46c2a203ba --- /dev/null +++ b/model-00036-of-00050.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:57dd0648f4c0ab2e555a567fec147bf9291b76d2f869aae6a9c6496ddb1e72a5 +size 4664167392 diff --git a/model-00037-of-00050.safetensors b/model-00037-of-00050.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..f952764db9387aedf7923fbab01860ee3db7f87e --- /dev/null +++ b/model-00037-of-00050.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:95283bc9f394df9b4ab6e4238f171f611d65d376a8183308db65b55250a1eac6 +size 4664167408 diff --git a/model-00038-of-00050.safetensors b/model-00038-of-00050.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..a00b6919cbb4d4d3be13ccb0f2dbad12b70fda84 --- /dev/null +++ b/model-00038-of-00050.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:488c24e4dcd0041a1389b69451d0479d0e9a8e23fbd992f8aee14eb3a9632c24 +size 4999711728 diff --git a/model-00039-of-00050.safetensors b/model-00039-of-00050.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..2bd56b28c5d934a189be78b1816464ed877897ff --- /dev/null +++ b/model-00039-of-00050.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:0d252e969237c96555924b822770154b0327040c0c623e01e71524c5d3263361 +size 4966157056 diff --git a/model-00040-of-00050.safetensors b/model-00040-of-00050.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..db1e1de9fd0ba82c10f3b5c42a6275f9dae7afff --- /dev/null +++ b/model-00040-of-00050.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:d542e8fb6b9044dfd9ea6c92268c75f45ff357b3bd89d43964cb7ff4bbf62de3 +size 4664134408 diff --git a/model-00041-of-00050.safetensors b/model-00041-of-00050.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..050e87f22a14fa6e6174f5d6575360effa626c47 --- /dev/null +++ b/model-00041-of-00050.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:521fd4d47b8b856fee2ceab0374650eb05e67086e1638ddfec16d19879b24d1a +size 4664167408 diff --git a/model-00042-of-00050.safetensors b/model-00042-of-00050.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..1cfcaecf1facf864715e70dd20c375f11df248d9 --- /dev/null +++ b/model-00042-of-00050.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:dbc7a79c6eefb635c1279ca6df8cb65506fe5239bee9d69be22bfa7b65f7220c +size 4664167408 diff --git a/model-00043-of-00050.safetensors b/model-00043-of-00050.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..9d6205c1f1e4c417706c33cff501da0f6c38dbbe --- /dev/null +++ b/model-00043-of-00050.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:0706e9131f8587f6bac215116c8dffd5e3e10835d8c7dfbdea51860b5d1aecdc +size 4999711728 diff --git a/model-00044-of-00050.safetensors b/model-00044-of-00050.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..eb4726007de4d9da14f02ec76029c2d04fd83422 --- /dev/null +++ b/model-00044-of-00050.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:5a5ffabff2db8ae5dd080073ed18f5a80761d655222f52dca9c04628050de24d +size 4966157056 diff --git a/model-00045-of-00050.safetensors b/model-00045-of-00050.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..9b1c67ed19d61be121aaeed8b6b473d08c678792 --- /dev/null +++ b/model-00045-of-00050.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:29c0351cedf9458c840b5ebba38897ed0d7d2f1e11ed0e147ef0091addb87710 +size 4664134408 diff --git a/model-00046-of-00050.safetensors b/model-00046-of-00050.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..f3abdb2ced358e5e9c6391f23ed704e987f41372 --- /dev/null +++ b/model-00046-of-00050.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:0b391b4be7e2aa59f979576fde336d978ffde8df075379bea874c9cfc296dfe4 +size 4664167408 diff --git a/model-00047-of-00050.safetensors b/model-00047-of-00050.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..baf149cd518121e9b480ce04795c89787203116c --- /dev/null +++ b/model-00047-of-00050.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:e7cd2f112557152937e2e8f65ab9b7b0d176c5ff361c3a78f2863631edf54acf +size 4664167408 diff --git a/model-00048-of-00050.safetensors b/model-00048-of-00050.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..25efced3eef48507e87099e0f4d2c9e2a2bece63 --- /dev/null +++ b/model-00048-of-00050.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:d9a7e691ace4d7ece57b1c1c77a4a0dd7b6edc374502f7fdbfa0fd692a6dc3e8 +size 4999711728 diff --git a/model-00049-of-00050.safetensors b/model-00049-of-00050.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..cce2c0ce71e78a588662c2a4e29b69ed9b5447d2 --- /dev/null +++ b/model-00049-of-00050.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:1f9e4208fca900b92676a8a5b6c1c3947af8fbe4e7ed14462df00db07f680594 +size 4966173536 diff --git a/model-00050-of-00050.safetensors b/model-00050-of-00050.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..1a80f341fb9946bd5d43a9d817cc4b3af487bb0f --- /dev/null +++ b/model-00050-of-00050.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:33067f55430e936aae098c13876c64be7f0484a0f6dc05ee28f9e316e2b3de67 +size 524337280 diff --git a/model.safetensors.index.json b/model.safetensors.index.json new file mode 100644 index 0000000000000000000000000000000000000000..671ce20733d02866d2a3b85de6439770ebef22b8 --- /dev/null +++ b/model.safetensors.index.json @@ -0,0 +1,1243 @@ +{ + "metadata": { + "total_size": 235497996288 + }, + "weight_map": { + "lm_head.weight": "model-00050-of-00050.safetensors", + "model.embed_tokens.weight": "model-00001-of-00050.safetensors", + "model.layers.0.input_layernorm.weight": "model-00001-of-00050.safetensors", + "model.layers.0.mlp.down_proj.weight": "model-00001-of-00050.safetensors", + "model.layers.0.mlp.gate_proj.weight": "model-00001-of-00050.safetensors", + "model.layers.0.mlp.up_proj.weight": "model-00001-of-00050.safetensors", + "model.layers.0.post_attention_layernorm.weight": "model-00001-of-00050.safetensors", + "model.layers.0.self_attn.k_proj.weight": "model-00001-of-00050.safetensors", + "model.layers.0.self_attn.o_proj.weight": "model-00001-of-00050.safetensors", + "model.layers.0.self_attn.q_proj.weight": "model-00001-of-00050.safetensors", + "model.layers.0.self_attn.v_proj.weight": "model-00001-of-00050.safetensors", + "model.layers.1.input_layernorm.weight": "model-00001-of-00050.safetensors", + "model.layers.1.mlp.down_proj.weight": "model-00001-of-00050.safetensors", + "model.layers.1.mlp.gate_proj.weight": "model-00001-of-00050.safetensors", + "model.layers.1.mlp.up_proj.weight": "model-00001-of-00050.safetensors", + "model.layers.1.post_attention_layernorm.weight": "model-00001-of-00050.safetensors", + "model.layers.1.self_attn.k_proj.weight": "model-00001-of-00050.safetensors", + "model.layers.1.self_attn.o_proj.weight": "model-00001-of-00050.safetensors", + "model.layers.1.self_attn.q_proj.weight": "model-00001-of-00050.safetensors", + "model.layers.1.self_attn.v_proj.weight": "model-00001-of-00050.safetensors", + "model.layers.10.input_layernorm.weight": "model-00004-of-00050.safetensors", + "model.layers.10.mlp.down_proj.weight": "model-00004-of-00050.safetensors", + "model.layers.10.mlp.gate_proj.weight": "model-00004-of-00050.safetensors", + "model.layers.10.mlp.up_proj.weight": "model-00004-of-00050.safetensors", + "model.layers.10.post_attention_layernorm.weight": "model-00004-of-00050.safetensors", + "model.layers.10.self_attn.k_proj.weight": "model-00004-of-00050.safetensors", + "model.layers.10.self_attn.o_proj.weight": "model-00004-of-00050.safetensors", + "model.layers.10.self_attn.q_proj.weight": "model-00004-of-00050.safetensors", + "model.layers.10.self_attn.v_proj.weight": "model-00004-of-00050.safetensors", + "model.layers.100.input_layernorm.weight": "model-00037-of-00050.safetensors", + "model.layers.100.mlp.down_proj.weight": "model-00037-of-00050.safetensors", + "model.layers.100.mlp.gate_proj.weight": "model-00036-of-00050.safetensors", + "model.layers.100.mlp.up_proj.weight": "model-00037-of-00050.safetensors", + "model.layers.100.post_attention_layernorm.weight": "model-00037-of-00050.safetensors", + "model.layers.100.self_attn.k_proj.weight": "model-00036-of-00050.safetensors", + "model.layers.100.self_attn.o_proj.weight": "model-00036-of-00050.safetensors", + "model.layers.100.self_attn.q_proj.weight": "model-00036-of-00050.safetensors", + "model.layers.100.self_attn.v_proj.weight": "model-00036-of-00050.safetensors", + "model.layers.101.input_layernorm.weight": "model-00037-of-00050.safetensors", + "model.layers.101.mlp.down_proj.weight": "model-00037-of-00050.safetensors", + "model.layers.101.mlp.gate_proj.weight": "model-00037-of-00050.safetensors", + "model.layers.101.mlp.up_proj.weight": "model-00037-of-00050.safetensors", + "model.layers.101.post_attention_layernorm.weight": "model-00037-of-00050.safetensors", + "model.layers.101.self_attn.k_proj.weight": "model-00037-of-00050.safetensors", + "model.layers.101.self_attn.o_proj.weight": "model-00037-of-00050.safetensors", + "model.layers.101.self_attn.q_proj.weight": "model-00037-of-00050.safetensors", + "model.layers.101.self_attn.v_proj.weight": "model-00037-of-00050.safetensors", + "model.layers.102.input_layernorm.weight": "model-00037-of-00050.safetensors", + "model.layers.102.mlp.down_proj.weight": "model-00037-of-00050.safetensors", + "model.layers.102.mlp.gate_proj.weight": "model-00037-of-00050.safetensors", + "model.layers.102.mlp.up_proj.weight": "model-00037-of-00050.safetensors", + "model.layers.102.post_attention_layernorm.weight": "model-00037-of-00050.safetensors", + "model.layers.102.self_attn.k_proj.weight": "model-00037-of-00050.safetensors", + "model.layers.102.self_attn.o_proj.weight": "model-00037-of-00050.safetensors", + "model.layers.102.self_attn.q_proj.weight": "model-00037-of-00050.safetensors", + "model.layers.102.self_attn.v_proj.weight": "model-00037-of-00050.safetensors", + "model.layers.103.input_layernorm.weight": "model-00038-of-00050.safetensors", + "model.layers.103.mlp.down_proj.weight": "model-00038-of-00050.safetensors", + "model.layers.103.mlp.gate_proj.weight": "model-00038-of-00050.safetensors", + "model.layers.103.mlp.up_proj.weight": "model-00038-of-00050.safetensors", + "model.layers.103.post_attention_layernorm.weight": "model-00038-of-00050.safetensors", + "model.layers.103.self_attn.k_proj.weight": "model-00037-of-00050.safetensors", + "model.layers.103.self_attn.o_proj.weight": "model-00037-of-00050.safetensors", + "model.layers.103.self_attn.q_proj.weight": "model-00037-of-00050.safetensors", + "model.layers.103.self_attn.v_proj.weight": "model-00037-of-00050.safetensors", + "model.layers.104.input_layernorm.weight": "model-00038-of-00050.safetensors", + "model.layers.104.mlp.down_proj.weight": "model-00038-of-00050.safetensors", + "model.layers.104.mlp.gate_proj.weight": "model-00038-of-00050.safetensors", + "model.layers.104.mlp.up_proj.weight": "model-00038-of-00050.safetensors", + "model.layers.104.post_attention_layernorm.weight": "model-00038-of-00050.safetensors", + "model.layers.104.self_attn.k_proj.weight": "model-00038-of-00050.safetensors", + "model.layers.104.self_attn.o_proj.weight": "model-00038-of-00050.safetensors", + "model.layers.104.self_attn.q_proj.weight": "model-00038-of-00050.safetensors", + "model.layers.104.self_attn.v_proj.weight": "model-00038-of-00050.safetensors", + "model.layers.105.input_layernorm.weight": "model-00038-of-00050.safetensors", + "model.layers.105.mlp.down_proj.weight": "model-00038-of-00050.safetensors", + "model.layers.105.mlp.gate_proj.weight": "model-00038-of-00050.safetensors", + "model.layers.105.mlp.up_proj.weight": "model-00038-of-00050.safetensors", + "model.layers.105.post_attention_layernorm.weight": "model-00038-of-00050.safetensors", + "model.layers.105.self_attn.k_proj.weight": "model-00038-of-00050.safetensors", + "model.layers.105.self_attn.o_proj.weight": "model-00038-of-00050.safetensors", + "model.layers.105.self_attn.q_proj.weight": "model-00038-of-00050.safetensors", + "model.layers.105.self_attn.v_proj.weight": "model-00038-of-00050.safetensors", + "model.layers.106.input_layernorm.weight": "model-00039-of-00050.safetensors", + "model.layers.106.mlp.down_proj.weight": "model-00039-of-00050.safetensors", + "model.layers.106.mlp.gate_proj.weight": "model-00039-of-00050.safetensors", + "model.layers.106.mlp.up_proj.weight": "model-00039-of-00050.safetensors", + "model.layers.106.post_attention_layernorm.weight": "model-00039-of-00050.safetensors", + "model.layers.106.self_attn.k_proj.weight": "model-00038-of-00050.safetensors", + "model.layers.106.self_attn.o_proj.weight": "model-00039-of-00050.safetensors", + "model.layers.106.self_attn.q_proj.weight": "model-00038-of-00050.safetensors", + "model.layers.106.self_attn.v_proj.weight": "model-00038-of-00050.safetensors", + "model.layers.107.input_layernorm.weight": "model-00039-of-00050.safetensors", + "model.layers.107.mlp.down_proj.weight": "model-00039-of-00050.safetensors", + "model.layers.107.mlp.gate_proj.weight": "model-00039-of-00050.safetensors", + "model.layers.107.mlp.up_proj.weight": "model-00039-of-00050.safetensors", + "model.layers.107.post_attention_layernorm.weight": "model-00039-of-00050.safetensors", + "model.layers.107.self_attn.k_proj.weight": "model-00039-of-00050.safetensors", + "model.layers.107.self_attn.o_proj.weight": "model-00039-of-00050.safetensors", + "model.layers.107.self_attn.q_proj.weight": "model-00039-of-00050.safetensors", + "model.layers.107.self_attn.v_proj.weight": "model-00039-of-00050.safetensors", + "model.layers.108.input_layernorm.weight": "model-00039-of-00050.safetensors", + "model.layers.108.mlp.down_proj.weight": "model-00039-of-00050.safetensors", + "model.layers.108.mlp.gate_proj.weight": "model-00039-of-00050.safetensors", + "model.layers.108.mlp.up_proj.weight": "model-00039-of-00050.safetensors", + "model.layers.108.post_attention_layernorm.weight": "model-00039-of-00050.safetensors", + "model.layers.108.self_attn.k_proj.weight": "model-00039-of-00050.safetensors", + "model.layers.108.self_attn.o_proj.weight": "model-00039-of-00050.safetensors", + "model.layers.108.self_attn.q_proj.weight": "model-00039-of-00050.safetensors", + "model.layers.108.self_attn.v_proj.weight": "model-00039-of-00050.safetensors", + "model.layers.109.input_layernorm.weight": "model-00040-of-00050.safetensors", + "model.layers.109.mlp.down_proj.weight": "model-00040-of-00050.safetensors", + "model.layers.109.mlp.gate_proj.weight": "model-00040-of-00050.safetensors", + "model.layers.109.mlp.up_proj.weight": "model-00040-of-00050.safetensors", + "model.layers.109.post_attention_layernorm.weight": "model-00040-of-00050.safetensors", + "model.layers.109.self_attn.k_proj.weight": "model-00040-of-00050.safetensors", + "model.layers.109.self_attn.o_proj.weight": "model-00040-of-00050.safetensors", + "model.layers.109.self_attn.q_proj.weight": "model-00040-of-00050.safetensors", + "model.layers.109.self_attn.v_proj.weight": "model-00040-of-00050.safetensors", + "model.layers.11.input_layernorm.weight": "model-00005-of-00050.safetensors", + "model.layers.11.mlp.down_proj.weight": "model-00005-of-00050.safetensors", + "model.layers.11.mlp.gate_proj.weight": "model-00005-of-00050.safetensors", + "model.layers.11.mlp.up_proj.weight": "model-00005-of-00050.safetensors", + "model.layers.11.post_attention_layernorm.weight": "model-00005-of-00050.safetensors", + "model.layers.11.self_attn.k_proj.weight": "model-00005-of-00050.safetensors", + "model.layers.11.self_attn.o_proj.weight": "model-00005-of-00050.safetensors", + "model.layers.11.self_attn.q_proj.weight": "model-00005-of-00050.safetensors", + "model.layers.11.self_attn.v_proj.weight": "model-00005-of-00050.safetensors", + "model.layers.110.input_layernorm.weight": "model-00040-of-00050.safetensors", + "model.layers.110.mlp.down_proj.weight": "model-00040-of-00050.safetensors", + "model.layers.110.mlp.gate_proj.weight": "model-00040-of-00050.safetensors", + "model.layers.110.mlp.up_proj.weight": "model-00040-of-00050.safetensors", + "model.layers.110.post_attention_layernorm.weight": "model-00040-of-00050.safetensors", + "model.layers.110.self_attn.k_proj.weight": "model-00040-of-00050.safetensors", + "model.layers.110.self_attn.o_proj.weight": "model-00040-of-00050.safetensors", + "model.layers.110.self_attn.q_proj.weight": "model-00040-of-00050.safetensors", + "model.layers.110.self_attn.v_proj.weight": "model-00040-of-00050.safetensors", + "model.layers.111.input_layernorm.weight": "model-00041-of-00050.safetensors", + "model.layers.111.mlp.down_proj.weight": "model-00041-of-00050.safetensors", + "model.layers.111.mlp.gate_proj.weight": "model-00040-of-00050.safetensors", + "model.layers.111.mlp.up_proj.weight": "model-00040-of-00050.safetensors", + "model.layers.111.post_attention_layernorm.weight": "model-00041-of-00050.safetensors", + "model.layers.111.self_attn.k_proj.weight": "model-00040-of-00050.safetensors", + "model.layers.111.self_attn.o_proj.weight": "model-00040-of-00050.safetensors", + "model.layers.111.self_attn.q_proj.weight": "model-00040-of-00050.safetensors", + "model.layers.111.self_attn.v_proj.weight": "model-00040-of-00050.safetensors", + "model.layers.112.input_layernorm.weight": "model-00041-of-00050.safetensors", + "model.layers.112.mlp.down_proj.weight": "model-00041-of-00050.safetensors", + "model.layers.112.mlp.gate_proj.weight": "model-00041-of-00050.safetensors", + "model.layers.112.mlp.up_proj.weight": "model-00041-of-00050.safetensors", + "model.layers.112.post_attention_layernorm.weight": "model-00041-of-00050.safetensors", + "model.layers.112.self_attn.k_proj.weight": "model-00041-of-00050.safetensors", + "model.layers.112.self_attn.o_proj.weight": "model-00041-of-00050.safetensors", + "model.layers.112.self_attn.q_proj.weight": "model-00041-of-00050.safetensors", + "model.layers.112.self_attn.v_proj.weight": "model-00041-of-00050.safetensors", + "model.layers.113.input_layernorm.weight": "model-00041-of-00050.safetensors", + "model.layers.113.mlp.down_proj.weight": "model-00041-of-00050.safetensors", + "model.layers.113.mlp.gate_proj.weight": "model-00041-of-00050.safetensors", + "model.layers.113.mlp.up_proj.weight": "model-00041-of-00050.safetensors", + "model.layers.113.post_attention_layernorm.weight": "model-00041-of-00050.safetensors", + "model.layers.113.self_attn.k_proj.weight": "model-00041-of-00050.safetensors", + "model.layers.113.self_attn.o_proj.weight": "model-00041-of-00050.safetensors", + "model.layers.113.self_attn.q_proj.weight": "model-00041-of-00050.safetensors", + "model.layers.113.self_attn.v_proj.weight": "model-00041-of-00050.safetensors", + "model.layers.114.input_layernorm.weight": "model-00042-of-00050.safetensors", + "model.layers.114.mlp.down_proj.weight": "model-00042-of-00050.safetensors", + "model.layers.114.mlp.gate_proj.weight": "model-00041-of-00050.safetensors", + "model.layers.114.mlp.up_proj.weight": "model-00042-of-00050.safetensors", + "model.layers.114.post_attention_layernorm.weight": "model-00042-of-00050.safetensors", + "model.layers.114.self_attn.k_proj.weight": "model-00041-of-00050.safetensors", + "model.layers.114.self_attn.o_proj.weight": "model-00041-of-00050.safetensors", + "model.layers.114.self_attn.q_proj.weight": "model-00041-of-00050.safetensors", + "model.layers.114.self_attn.v_proj.weight": "model-00041-of-00050.safetensors", + "model.layers.115.input_layernorm.weight": "model-00042-of-00050.safetensors", + "model.layers.115.mlp.down_proj.weight": "model-00042-of-00050.safetensors", + "model.layers.115.mlp.gate_proj.weight": "model-00042-of-00050.safetensors", + "model.layers.115.mlp.up_proj.weight": "model-00042-of-00050.safetensors", + "model.layers.115.post_attention_layernorm.weight": "model-00042-of-00050.safetensors", + "model.layers.115.self_attn.k_proj.weight": "model-00042-of-00050.safetensors", + "model.layers.115.self_attn.o_proj.weight": "model-00042-of-00050.safetensors", + "model.layers.115.self_attn.q_proj.weight": "model-00042-of-00050.safetensors", + "model.layers.115.self_attn.v_proj.weight": "model-00042-of-00050.safetensors", + "model.layers.116.input_layernorm.weight": "model-00042-of-00050.safetensors", + "model.layers.116.mlp.down_proj.weight": "model-00042-of-00050.safetensors", + "model.layers.116.mlp.gate_proj.weight": "model-00042-of-00050.safetensors", + "model.layers.116.mlp.up_proj.weight": "model-00042-of-00050.safetensors", + "model.layers.116.post_attention_layernorm.weight": "model-00042-of-00050.safetensors", + "model.layers.116.self_attn.k_proj.weight": "model-00042-of-00050.safetensors", + "model.layers.116.self_attn.o_proj.weight": "model-00042-of-00050.safetensors", + "model.layers.116.self_attn.q_proj.weight": "model-00042-of-00050.safetensors", + "model.layers.116.self_attn.v_proj.weight": "model-00042-of-00050.safetensors", + "model.layers.117.input_layernorm.weight": "model-00043-of-00050.safetensors", + "model.layers.117.mlp.down_proj.weight": "model-00043-of-00050.safetensors", + "model.layers.117.mlp.gate_proj.weight": "model-00043-of-00050.safetensors", + "model.layers.117.mlp.up_proj.weight": "model-00043-of-00050.safetensors", + "model.layers.117.post_attention_layernorm.weight": "model-00043-of-00050.safetensors", + "model.layers.117.self_attn.k_proj.weight": "model-00042-of-00050.safetensors", + "model.layers.117.self_attn.o_proj.weight": "model-00042-of-00050.safetensors", + "model.layers.117.self_attn.q_proj.weight": "model-00042-of-00050.safetensors", + "model.layers.117.self_attn.v_proj.weight": "model-00042-of-00050.safetensors", + "model.layers.118.input_layernorm.weight": "model-00043-of-00050.safetensors", + "model.layers.118.mlp.down_proj.weight": "model-00043-of-00050.safetensors", + "model.layers.118.mlp.gate_proj.weight": "model-00043-of-00050.safetensors", + "model.layers.118.mlp.up_proj.weight": "model-00043-of-00050.safetensors", + "model.layers.118.post_attention_layernorm.weight": "model-00043-of-00050.safetensors", + "model.layers.118.self_attn.k_proj.weight": "model-00043-of-00050.safetensors", + "model.layers.118.self_attn.o_proj.weight": "model-00043-of-00050.safetensors", + "model.layers.118.self_attn.q_proj.weight": "model-00043-of-00050.safetensors", + "model.layers.118.self_attn.v_proj.weight": "model-00043-of-00050.safetensors", + "model.layers.119.input_layernorm.weight": "model-00043-of-00050.safetensors", + "model.layers.119.mlp.down_proj.weight": "model-00043-of-00050.safetensors", + "model.layers.119.mlp.gate_proj.weight": "model-00043-of-00050.safetensors", + "model.layers.119.mlp.up_proj.weight": "model-00043-of-00050.safetensors", + "model.layers.119.post_attention_layernorm.weight": "model-00043-of-00050.safetensors", + "model.layers.119.self_attn.k_proj.weight": "model-00043-of-00050.safetensors", + "model.layers.119.self_attn.o_proj.weight": "model-00043-of-00050.safetensors", + "model.layers.119.self_attn.q_proj.weight": "model-00043-of-00050.safetensors", + "model.layers.119.self_attn.v_proj.weight": "model-00043-of-00050.safetensors", + "model.layers.12.input_layernorm.weight": "model-00005-of-00050.safetensors", + "model.layers.12.mlp.down_proj.weight": "model-00005-of-00050.safetensors", + "model.layers.12.mlp.gate_proj.weight": "model-00005-of-00050.safetensors", + "model.layers.12.mlp.up_proj.weight": "model-00005-of-00050.safetensors", + "model.layers.12.post_attention_layernorm.weight": "model-00005-of-00050.safetensors", + "model.layers.12.self_attn.k_proj.weight": "model-00005-of-00050.safetensors", + "model.layers.12.self_attn.o_proj.weight": "model-00005-of-00050.safetensors", + "model.layers.12.self_attn.q_proj.weight": "model-00005-of-00050.safetensors", + "model.layers.12.self_attn.v_proj.weight": "model-00005-of-00050.safetensors", + "model.layers.120.input_layernorm.weight": "model-00044-of-00050.safetensors", + "model.layers.120.mlp.down_proj.weight": "model-00044-of-00050.safetensors", + "model.layers.120.mlp.gate_proj.weight": "model-00044-of-00050.safetensors", + "model.layers.120.mlp.up_proj.weight": "model-00044-of-00050.safetensors", + "model.layers.120.post_attention_layernorm.weight": "model-00044-of-00050.safetensors", + "model.layers.120.self_attn.k_proj.weight": "model-00043-of-00050.safetensors", + "model.layers.120.self_attn.o_proj.weight": "model-00044-of-00050.safetensors", + "model.layers.120.self_attn.q_proj.weight": "model-00043-of-00050.safetensors", + "model.layers.120.self_attn.v_proj.weight": "model-00043-of-00050.safetensors", + "model.layers.121.input_layernorm.weight": "model-00044-of-00050.safetensors", + "model.layers.121.mlp.down_proj.weight": "model-00044-of-00050.safetensors", + "model.layers.121.mlp.gate_proj.weight": "model-00044-of-00050.safetensors", + "model.layers.121.mlp.up_proj.weight": "model-00044-of-00050.safetensors", + "model.layers.121.post_attention_layernorm.weight": "model-00044-of-00050.safetensors", + "model.layers.121.self_attn.k_proj.weight": "model-00044-of-00050.safetensors", + "model.layers.121.self_attn.o_proj.weight": "model-00044-of-00050.safetensors", + "model.layers.121.self_attn.q_proj.weight": "model-00044-of-00050.safetensors", + "model.layers.121.self_attn.v_proj.weight": "model-00044-of-00050.safetensors", + "model.layers.122.input_layernorm.weight": "model-00044-of-00050.safetensors", + "model.layers.122.mlp.down_proj.weight": "model-00044-of-00050.safetensors", + "model.layers.122.mlp.gate_proj.weight": "model-00044-of-00050.safetensors", + "model.layers.122.mlp.up_proj.weight": "model-00044-of-00050.safetensors", + "model.layers.122.post_attention_layernorm.weight": "model-00044-of-00050.safetensors", + "model.layers.122.self_attn.k_proj.weight": "model-00044-of-00050.safetensors", + "model.layers.122.self_attn.o_proj.weight": "model-00044-of-00050.safetensors", + "model.layers.122.self_attn.q_proj.weight": "model-00044-of-00050.safetensors", + "model.layers.122.self_attn.v_proj.weight": "model-00044-of-00050.safetensors", + "model.layers.123.input_layernorm.weight": "model-00045-of-00050.safetensors", + "model.layers.123.mlp.down_proj.weight": "model-00045-of-00050.safetensors", + "model.layers.123.mlp.gate_proj.weight": "model-00045-of-00050.safetensors", + "model.layers.123.mlp.up_proj.weight": "model-00045-of-00050.safetensors", + "model.layers.123.post_attention_layernorm.weight": "model-00045-of-00050.safetensors", + "model.layers.123.self_attn.k_proj.weight": "model-00045-of-00050.safetensors", + "model.layers.123.self_attn.o_proj.weight": "model-00045-of-00050.safetensors", + "model.layers.123.self_attn.q_proj.weight": "model-00045-of-00050.safetensors", + "model.layers.123.self_attn.v_proj.weight": "model-00045-of-00050.safetensors", + "model.layers.124.input_layernorm.weight": "model-00045-of-00050.safetensors", + "model.layers.124.mlp.down_proj.weight": "model-00045-of-00050.safetensors", + "model.layers.124.mlp.gate_proj.weight": "model-00045-of-00050.safetensors", + "model.layers.124.mlp.up_proj.weight": "model-00045-of-00050.safetensors", + "model.layers.124.post_attention_layernorm.weight": "model-00045-of-00050.safetensors", + "model.layers.124.self_attn.k_proj.weight": "model-00045-of-00050.safetensors", + "model.layers.124.self_attn.o_proj.weight": "model-00045-of-00050.safetensors", + "model.layers.124.self_attn.q_proj.weight": "model-00045-of-00050.safetensors", + "model.layers.124.self_attn.v_proj.weight": "model-00045-of-00050.safetensors", + "model.layers.125.input_layernorm.weight": "model-00046-of-00050.safetensors", + "model.layers.125.mlp.down_proj.weight": "model-00046-of-00050.safetensors", + "model.layers.125.mlp.gate_proj.weight": "model-00045-of-00050.safetensors", + "model.layers.125.mlp.up_proj.weight": "model-00045-of-00050.safetensors", + "model.layers.125.post_attention_layernorm.weight": "model-00046-of-00050.safetensors", + "model.layers.125.self_attn.k_proj.weight": "model-00045-of-00050.safetensors", + "model.layers.125.self_attn.o_proj.weight": "model-00045-of-00050.safetensors", + "model.layers.125.self_attn.q_proj.weight": "model-00045-of-00050.safetensors", + "model.layers.125.self_attn.v_proj.weight": "model-00045-of-00050.safetensors", + "model.layers.126.input_layernorm.weight": "model-00046-of-00050.safetensors", + "model.layers.126.mlp.down_proj.weight": "model-00046-of-00050.safetensors", + "model.layers.126.mlp.gate_proj.weight": "model-00046-of-00050.safetensors", + "model.layers.126.mlp.up_proj.weight": "model-00046-of-00050.safetensors", + "model.layers.126.post_attention_layernorm.weight": "model-00046-of-00050.safetensors", + "model.layers.126.self_attn.k_proj.weight": "model-00046-of-00050.safetensors", + "model.layers.126.self_attn.o_proj.weight": "model-00046-of-00050.safetensors", + "model.layers.126.self_attn.q_proj.weight": "model-00046-of-00050.safetensors", + "model.layers.126.self_attn.v_proj.weight": "model-00046-of-00050.safetensors", + "model.layers.127.input_layernorm.weight": "model-00046-of-00050.safetensors", + "model.layers.127.mlp.down_proj.weight": "model-00046-of-00050.safetensors", + "model.layers.127.mlp.gate_proj.weight": "model-00046-of-00050.safetensors", + "model.layers.127.mlp.up_proj.weight": "model-00046-of-00050.safetensors", + "model.layers.127.post_attention_layernorm.weight": "model-00046-of-00050.safetensors", + "model.layers.127.self_attn.k_proj.weight": "model-00046-of-00050.safetensors", + "model.layers.127.self_attn.o_proj.weight": "model-00046-of-00050.safetensors", + "model.layers.127.self_attn.q_proj.weight": "model-00046-of-00050.safetensors", + "model.layers.127.self_attn.v_proj.weight": "model-00046-of-00050.safetensors", + "model.layers.128.input_layernorm.weight": "model-00047-of-00050.safetensors", + "model.layers.128.mlp.down_proj.weight": "model-00047-of-00050.safetensors", + "model.layers.128.mlp.gate_proj.weight": "model-00046-of-00050.safetensors", + "model.layers.128.mlp.up_proj.weight": "model-00047-of-00050.safetensors", + "model.layers.128.post_attention_layernorm.weight": "model-00047-of-00050.safetensors", + "model.layers.128.self_attn.k_proj.weight": "model-00046-of-00050.safetensors", + "model.layers.128.self_attn.o_proj.weight": "model-00046-of-00050.safetensors", + "model.layers.128.self_attn.q_proj.weight": "model-00046-of-00050.safetensors", + "model.layers.128.self_attn.v_proj.weight": "model-00046-of-00050.safetensors", + "model.layers.129.input_layernorm.weight": "model-00047-of-00050.safetensors", + "model.layers.129.mlp.down_proj.weight": "model-00047-of-00050.safetensors", + "model.layers.129.mlp.gate_proj.weight": "model-00047-of-00050.safetensors", + "model.layers.129.mlp.up_proj.weight": "model-00047-of-00050.safetensors", + "model.layers.129.post_attention_layernorm.weight": "model-00047-of-00050.safetensors", + "model.layers.129.self_attn.k_proj.weight": "model-00047-of-00050.safetensors", + "model.layers.129.self_attn.o_proj.weight": "model-00047-of-00050.safetensors", + "model.layers.129.self_attn.q_proj.weight": "model-00047-of-00050.safetensors", + "model.layers.129.self_attn.v_proj.weight": "model-00047-of-00050.safetensors", + "model.layers.13.input_layernorm.weight": "model-00006-of-00050.safetensors", + "model.layers.13.mlp.down_proj.weight": "model-00006-of-00050.safetensors", + "model.layers.13.mlp.gate_proj.weight": "model-00005-of-00050.safetensors", + "model.layers.13.mlp.up_proj.weight": "model-00005-of-00050.safetensors", + "model.layers.13.post_attention_layernorm.weight": "model-00006-of-00050.safetensors", + "model.layers.13.self_attn.k_proj.weight": "model-00005-of-00050.safetensors", + "model.layers.13.self_attn.o_proj.weight": "model-00005-of-00050.safetensors", + "model.layers.13.self_attn.q_proj.weight": "model-00005-of-00050.safetensors", + "model.layers.13.self_attn.v_proj.weight": "model-00005-of-00050.safetensors", + "model.layers.130.input_layernorm.weight": "model-00047-of-00050.safetensors", + "model.layers.130.mlp.down_proj.weight": "model-00047-of-00050.safetensors", + "model.layers.130.mlp.gate_proj.weight": "model-00047-of-00050.safetensors", + "model.layers.130.mlp.up_proj.weight": "model-00047-of-00050.safetensors", + "model.layers.130.post_attention_layernorm.weight": "model-00047-of-00050.safetensors", + "model.layers.130.self_attn.k_proj.weight": "model-00047-of-00050.safetensors", + "model.layers.130.self_attn.o_proj.weight": "model-00047-of-00050.safetensors", + "model.layers.130.self_attn.q_proj.weight": "model-00047-of-00050.safetensors", + "model.layers.130.self_attn.v_proj.weight": "model-00047-of-00050.safetensors", + "model.layers.131.input_layernorm.weight": "model-00048-of-00050.safetensors", + "model.layers.131.mlp.down_proj.weight": "model-00048-of-00050.safetensors", + "model.layers.131.mlp.gate_proj.weight": "model-00048-of-00050.safetensors", + "model.layers.131.mlp.up_proj.weight": "model-00048-of-00050.safetensors", + "model.layers.131.post_attention_layernorm.weight": "model-00048-of-00050.safetensors", + "model.layers.131.self_attn.k_proj.weight": "model-00047-of-00050.safetensors", + "model.layers.131.self_attn.o_proj.weight": "model-00047-of-00050.safetensors", + "model.layers.131.self_attn.q_proj.weight": "model-00047-of-00050.safetensors", + "model.layers.131.self_attn.v_proj.weight": "model-00047-of-00050.safetensors", + "model.layers.132.input_layernorm.weight": "model-00048-of-00050.safetensors", + "model.layers.132.mlp.down_proj.weight": "model-00048-of-00050.safetensors", + "model.layers.132.mlp.gate_proj.weight": "model-00048-of-00050.safetensors", + "model.layers.132.mlp.up_proj.weight": "model-00048-of-00050.safetensors", + "model.layers.132.post_attention_layernorm.weight": "model-00048-of-00050.safetensors", + "model.layers.132.self_attn.k_proj.weight": "model-00048-of-00050.safetensors", + "model.layers.132.self_attn.o_proj.weight": "model-00048-of-00050.safetensors", + "model.layers.132.self_attn.q_proj.weight": "model-00048-of-00050.safetensors", + "model.layers.132.self_attn.v_proj.weight": "model-00048-of-00050.safetensors", + "model.layers.133.input_layernorm.weight": "model-00048-of-00050.safetensors", + "model.layers.133.mlp.down_proj.weight": "model-00048-of-00050.safetensors", + "model.layers.133.mlp.gate_proj.weight": "model-00048-of-00050.safetensors", + "model.layers.133.mlp.up_proj.weight": "model-00048-of-00050.safetensors", + "model.layers.133.post_attention_layernorm.weight": "model-00048-of-00050.safetensors", + "model.layers.133.self_attn.k_proj.weight": "model-00048-of-00050.safetensors", + "model.layers.133.self_attn.o_proj.weight": "model-00048-of-00050.safetensors", + "model.layers.133.self_attn.q_proj.weight": "model-00048-of-00050.safetensors", + "model.layers.133.self_attn.v_proj.weight": "model-00048-of-00050.safetensors", + "model.layers.134.input_layernorm.weight": "model-00049-of-00050.safetensors", + "model.layers.134.mlp.down_proj.weight": "model-00049-of-00050.safetensors", + "model.layers.134.mlp.gate_proj.weight": "model-00049-of-00050.safetensors", + "model.layers.134.mlp.up_proj.weight": "model-00049-of-00050.safetensors", + "model.layers.134.post_attention_layernorm.weight": "model-00049-of-00050.safetensors", + "model.layers.134.self_attn.k_proj.weight": "model-00048-of-00050.safetensors", + "model.layers.134.self_attn.o_proj.weight": "model-00049-of-00050.safetensors", + "model.layers.134.self_attn.q_proj.weight": "model-00048-of-00050.safetensors", + "model.layers.134.self_attn.v_proj.weight": "model-00048-of-00050.safetensors", + "model.layers.135.input_layernorm.weight": "model-00049-of-00050.safetensors", + "model.layers.135.mlp.down_proj.weight": "model-00049-of-00050.safetensors", + "model.layers.135.mlp.gate_proj.weight": "model-00049-of-00050.safetensors", + "model.layers.135.mlp.up_proj.weight": "model-00049-of-00050.safetensors", + "model.layers.135.post_attention_layernorm.weight": "model-00049-of-00050.safetensors", + "model.layers.135.self_attn.k_proj.weight": "model-00049-of-00050.safetensors", + "model.layers.135.self_attn.o_proj.weight": "model-00049-of-00050.safetensors", + "model.layers.135.self_attn.q_proj.weight": "model-00049-of-00050.safetensors", + "model.layers.135.self_attn.v_proj.weight": "model-00049-of-00050.safetensors", + "model.layers.136.input_layernorm.weight": "model-00049-of-00050.safetensors", + "model.layers.136.mlp.down_proj.weight": "model-00049-of-00050.safetensors", + "model.layers.136.mlp.gate_proj.weight": "model-00049-of-00050.safetensors", + "model.layers.136.mlp.up_proj.weight": "model-00049-of-00050.safetensors", + "model.layers.136.post_attention_layernorm.weight": "model-00049-of-00050.safetensors", + "model.layers.136.self_attn.k_proj.weight": "model-00049-of-00050.safetensors", + "model.layers.136.self_attn.o_proj.weight": "model-00049-of-00050.safetensors", + "model.layers.136.self_attn.q_proj.weight": "model-00049-of-00050.safetensors", + "model.layers.136.self_attn.v_proj.weight": "model-00049-of-00050.safetensors", + "model.layers.14.input_layernorm.weight": "model-00006-of-00050.safetensors", + "model.layers.14.mlp.down_proj.weight": "model-00006-of-00050.safetensors", + "model.layers.14.mlp.gate_proj.weight": "model-00006-of-00050.safetensors", + "model.layers.14.mlp.up_proj.weight": "model-00006-of-00050.safetensors", + "model.layers.14.post_attention_layernorm.weight": "model-00006-of-00050.safetensors", + "model.layers.14.self_attn.k_proj.weight": "model-00006-of-00050.safetensors", + "model.layers.14.self_attn.o_proj.weight": "model-00006-of-00050.safetensors", + "model.layers.14.self_attn.q_proj.weight": "model-00006-of-00050.safetensors", + "model.layers.14.self_attn.v_proj.weight": "model-00006-of-00050.safetensors", + "model.layers.15.input_layernorm.weight": "model-00006-of-00050.safetensors", + "model.layers.15.mlp.down_proj.weight": "model-00006-of-00050.safetensors", + "model.layers.15.mlp.gate_proj.weight": "model-00006-of-00050.safetensors", + "model.layers.15.mlp.up_proj.weight": "model-00006-of-00050.safetensors", + "model.layers.15.post_attention_layernorm.weight": "model-00006-of-00050.safetensors", + "model.layers.15.self_attn.k_proj.weight": "model-00006-of-00050.safetensors", + "model.layers.15.self_attn.o_proj.weight": "model-00006-of-00050.safetensors", + "model.layers.15.self_attn.q_proj.weight": "model-00006-of-00050.safetensors", + "model.layers.15.self_attn.v_proj.weight": "model-00006-of-00050.safetensors", + "model.layers.16.input_layernorm.weight": "model-00007-of-00050.safetensors", + "model.layers.16.mlp.down_proj.weight": "model-00007-of-00050.safetensors", + "model.layers.16.mlp.gate_proj.weight": "model-00006-of-00050.safetensors", + "model.layers.16.mlp.up_proj.weight": "model-00007-of-00050.safetensors", + "model.layers.16.post_attention_layernorm.weight": "model-00007-of-00050.safetensors", + "model.layers.16.self_attn.k_proj.weight": "model-00006-of-00050.safetensors", + "model.layers.16.self_attn.o_proj.weight": "model-00006-of-00050.safetensors", + "model.layers.16.self_attn.q_proj.weight": "model-00006-of-00050.safetensors", + "model.layers.16.self_attn.v_proj.weight": "model-00006-of-00050.safetensors", + "model.layers.17.input_layernorm.weight": "model-00007-of-00050.safetensors", + "model.layers.17.mlp.down_proj.weight": "model-00007-of-00050.safetensors", + "model.layers.17.mlp.gate_proj.weight": "model-00007-of-00050.safetensors", + "model.layers.17.mlp.up_proj.weight": "model-00007-of-00050.safetensors", + "model.layers.17.post_attention_layernorm.weight": "model-00007-of-00050.safetensors", + "model.layers.17.self_attn.k_proj.weight": "model-00007-of-00050.safetensors", + "model.layers.17.self_attn.o_proj.weight": "model-00007-of-00050.safetensors", + "model.layers.17.self_attn.q_proj.weight": "model-00007-of-00050.safetensors", + "model.layers.17.self_attn.v_proj.weight": "model-00007-of-00050.safetensors", + "model.layers.18.input_layernorm.weight": "model-00007-of-00050.safetensors", + "model.layers.18.mlp.down_proj.weight": "model-00007-of-00050.safetensors", + "model.layers.18.mlp.gate_proj.weight": "model-00007-of-00050.safetensors", + "model.layers.18.mlp.up_proj.weight": "model-00007-of-00050.safetensors", + "model.layers.18.post_attention_layernorm.weight": "model-00007-of-00050.safetensors", + "model.layers.18.self_attn.k_proj.weight": "model-00007-of-00050.safetensors", + "model.layers.18.self_attn.o_proj.weight": "model-00007-of-00050.safetensors", + "model.layers.18.self_attn.q_proj.weight": "model-00007-of-00050.safetensors", + "model.layers.18.self_attn.v_proj.weight": "model-00007-of-00050.safetensors", + "model.layers.19.input_layernorm.weight": "model-00008-of-00050.safetensors", + "model.layers.19.mlp.down_proj.weight": "model-00008-of-00050.safetensors", + "model.layers.19.mlp.gate_proj.weight": "model-00008-of-00050.safetensors", + "model.layers.19.mlp.up_proj.weight": "model-00008-of-00050.safetensors", + "model.layers.19.post_attention_layernorm.weight": "model-00008-of-00050.safetensors", + "model.layers.19.self_attn.k_proj.weight": "model-00007-of-00050.safetensors", + "model.layers.19.self_attn.o_proj.weight": "model-00007-of-00050.safetensors", + "model.layers.19.self_attn.q_proj.weight": "model-00007-of-00050.safetensors", + "model.layers.19.self_attn.v_proj.weight": "model-00007-of-00050.safetensors", + "model.layers.2.input_layernorm.weight": "model-00002-of-00050.safetensors", + "model.layers.2.mlp.down_proj.weight": "model-00002-of-00050.safetensors", + "model.layers.2.mlp.gate_proj.weight": "model-00001-of-00050.safetensors", + "model.layers.2.mlp.up_proj.weight": "model-00002-of-00050.safetensors", + "model.layers.2.post_attention_layernorm.weight": "model-00002-of-00050.safetensors", + "model.layers.2.self_attn.k_proj.weight": "model-00001-of-00050.safetensors", + "model.layers.2.self_attn.o_proj.weight": "model-00001-of-00050.safetensors", + "model.layers.2.self_attn.q_proj.weight": "model-00001-of-00050.safetensors", + "model.layers.2.self_attn.v_proj.weight": "model-00001-of-00050.safetensors", + "model.layers.20.input_layernorm.weight": "model-00008-of-00050.safetensors", + "model.layers.20.mlp.down_proj.weight": "model-00008-of-00050.safetensors", + "model.layers.20.mlp.gate_proj.weight": "model-00008-of-00050.safetensors", + "model.layers.20.mlp.up_proj.weight": "model-00008-of-00050.safetensors", + "model.layers.20.post_attention_layernorm.weight": "model-00008-of-00050.safetensors", + "model.layers.20.self_attn.k_proj.weight": "model-00008-of-00050.safetensors", + "model.layers.20.self_attn.o_proj.weight": "model-00008-of-00050.safetensors", + "model.layers.20.self_attn.q_proj.weight": "model-00008-of-00050.safetensors", + "model.layers.20.self_attn.v_proj.weight": "model-00008-of-00050.safetensors", + "model.layers.21.input_layernorm.weight": "model-00008-of-00050.safetensors", + "model.layers.21.mlp.down_proj.weight": "model-00008-of-00050.safetensors", + "model.layers.21.mlp.gate_proj.weight": "model-00008-of-00050.safetensors", + "model.layers.21.mlp.up_proj.weight": "model-00008-of-00050.safetensors", + "model.layers.21.post_attention_layernorm.weight": "model-00008-of-00050.safetensors", + "model.layers.21.self_attn.k_proj.weight": "model-00008-of-00050.safetensors", + "model.layers.21.self_attn.o_proj.weight": "model-00008-of-00050.safetensors", + "model.layers.21.self_attn.q_proj.weight": "model-00008-of-00050.safetensors", + "model.layers.21.self_attn.v_proj.weight": "model-00008-of-00050.safetensors", + "model.layers.22.input_layernorm.weight": "model-00009-of-00050.safetensors", + "model.layers.22.mlp.down_proj.weight": "model-00009-of-00050.safetensors", + "model.layers.22.mlp.gate_proj.weight": "model-00009-of-00050.safetensors", + "model.layers.22.mlp.up_proj.weight": "model-00009-of-00050.safetensors", + "model.layers.22.post_attention_layernorm.weight": "model-00009-of-00050.safetensors", + "model.layers.22.self_attn.k_proj.weight": "model-00008-of-00050.safetensors", + "model.layers.22.self_attn.o_proj.weight": "model-00009-of-00050.safetensors", + "model.layers.22.self_attn.q_proj.weight": "model-00008-of-00050.safetensors", + "model.layers.22.self_attn.v_proj.weight": "model-00008-of-00050.safetensors", + "model.layers.23.input_layernorm.weight": "model-00009-of-00050.safetensors", + "model.layers.23.mlp.down_proj.weight": "model-00009-of-00050.safetensors", + "model.layers.23.mlp.gate_proj.weight": "model-00009-of-00050.safetensors", + "model.layers.23.mlp.up_proj.weight": "model-00009-of-00050.safetensors", + "model.layers.23.post_attention_layernorm.weight": "model-00009-of-00050.safetensors", + "model.layers.23.self_attn.k_proj.weight": "model-00009-of-00050.safetensors", + "model.layers.23.self_attn.o_proj.weight": "model-00009-of-00050.safetensors", + "model.layers.23.self_attn.q_proj.weight": "model-00009-of-00050.safetensors", + "model.layers.23.self_attn.v_proj.weight": "model-00009-of-00050.safetensors", + "model.layers.24.input_layernorm.weight": "model-00009-of-00050.safetensors", + "model.layers.24.mlp.down_proj.weight": "model-00009-of-00050.safetensors", + "model.layers.24.mlp.gate_proj.weight": "model-00009-of-00050.safetensors", + "model.layers.24.mlp.up_proj.weight": "model-00009-of-00050.safetensors", + "model.layers.24.post_attention_layernorm.weight": "model-00009-of-00050.safetensors", + "model.layers.24.self_attn.k_proj.weight": "model-00009-of-00050.safetensors", + "model.layers.24.self_attn.o_proj.weight": "model-00009-of-00050.safetensors", + "model.layers.24.self_attn.q_proj.weight": "model-00009-of-00050.safetensors", + "model.layers.24.self_attn.v_proj.weight": "model-00009-of-00050.safetensors", + "model.layers.25.input_layernorm.weight": "model-00010-of-00050.safetensors", + "model.layers.25.mlp.down_proj.weight": "model-00010-of-00050.safetensors", + "model.layers.25.mlp.gate_proj.weight": "model-00010-of-00050.safetensors", + "model.layers.25.mlp.up_proj.weight": "model-00010-of-00050.safetensors", + "model.layers.25.post_attention_layernorm.weight": "model-00010-of-00050.safetensors", + "model.layers.25.self_attn.k_proj.weight": "model-00010-of-00050.safetensors", + "model.layers.25.self_attn.o_proj.weight": "model-00010-of-00050.safetensors", + "model.layers.25.self_attn.q_proj.weight": "model-00010-of-00050.safetensors", + "model.layers.25.self_attn.v_proj.weight": "model-00010-of-00050.safetensors", + "model.layers.26.input_layernorm.weight": "model-00010-of-00050.safetensors", + "model.layers.26.mlp.down_proj.weight": "model-00010-of-00050.safetensors", + "model.layers.26.mlp.gate_proj.weight": "model-00010-of-00050.safetensors", + "model.layers.26.mlp.up_proj.weight": "model-00010-of-00050.safetensors", + "model.layers.26.post_attention_layernorm.weight": "model-00010-of-00050.safetensors", + "model.layers.26.self_attn.k_proj.weight": "model-00010-of-00050.safetensors", + "model.layers.26.self_attn.o_proj.weight": "model-00010-of-00050.safetensors", + "model.layers.26.self_attn.q_proj.weight": "model-00010-of-00050.safetensors", + "model.layers.26.self_attn.v_proj.weight": "model-00010-of-00050.safetensors", + "model.layers.27.input_layernorm.weight": "model-00011-of-00050.safetensors", + "model.layers.27.mlp.down_proj.weight": "model-00011-of-00050.safetensors", + "model.layers.27.mlp.gate_proj.weight": "model-00010-of-00050.safetensors", + "model.layers.27.mlp.up_proj.weight": "model-00010-of-00050.safetensors", + "model.layers.27.post_attention_layernorm.weight": "model-00011-of-00050.safetensors", + "model.layers.27.self_attn.k_proj.weight": "model-00010-of-00050.safetensors", + "model.layers.27.self_attn.o_proj.weight": "model-00010-of-00050.safetensors", + "model.layers.27.self_attn.q_proj.weight": "model-00010-of-00050.safetensors", + "model.layers.27.self_attn.v_proj.weight": "model-00010-of-00050.safetensors", + "model.layers.28.input_layernorm.weight": "model-00011-of-00050.safetensors", + "model.layers.28.mlp.down_proj.weight": "model-00011-of-00050.safetensors", + "model.layers.28.mlp.gate_proj.weight": "model-00011-of-00050.safetensors", + "model.layers.28.mlp.up_proj.weight": "model-00011-of-00050.safetensors", + "model.layers.28.post_attention_layernorm.weight": "model-00011-of-00050.safetensors", + "model.layers.28.self_attn.k_proj.weight": "model-00011-of-00050.safetensors", + "model.layers.28.self_attn.o_proj.weight": "model-00011-of-00050.safetensors", + "model.layers.28.self_attn.q_proj.weight": "model-00011-of-00050.safetensors", + "model.layers.28.self_attn.v_proj.weight": "model-00011-of-00050.safetensors", + "model.layers.29.input_layernorm.weight": "model-00011-of-00050.safetensors", + "model.layers.29.mlp.down_proj.weight": "model-00011-of-00050.safetensors", + "model.layers.29.mlp.gate_proj.weight": "model-00011-of-00050.safetensors", + "model.layers.29.mlp.up_proj.weight": "model-00011-of-00050.safetensors", + "model.layers.29.post_attention_layernorm.weight": "model-00011-of-00050.safetensors", + "model.layers.29.self_attn.k_proj.weight": "model-00011-of-00050.safetensors", + "model.layers.29.self_attn.o_proj.weight": "model-00011-of-00050.safetensors", + "model.layers.29.self_attn.q_proj.weight": "model-00011-of-00050.safetensors", + "model.layers.29.self_attn.v_proj.weight": "model-00011-of-00050.safetensors", + "model.layers.3.input_layernorm.weight": "model-00002-of-00050.safetensors", + "model.layers.3.mlp.down_proj.weight": "model-00002-of-00050.safetensors", + "model.layers.3.mlp.gate_proj.weight": "model-00002-of-00050.safetensors", + "model.layers.3.mlp.up_proj.weight": "model-00002-of-00050.safetensors", + "model.layers.3.post_attention_layernorm.weight": "model-00002-of-00050.safetensors", + "model.layers.3.self_attn.k_proj.weight": "model-00002-of-00050.safetensors", + "model.layers.3.self_attn.o_proj.weight": "model-00002-of-00050.safetensors", + "model.layers.3.self_attn.q_proj.weight": "model-00002-of-00050.safetensors", + "model.layers.3.self_attn.v_proj.weight": "model-00002-of-00050.safetensors", + "model.layers.30.input_layernorm.weight": "model-00012-of-00050.safetensors", + "model.layers.30.mlp.down_proj.weight": "model-00012-of-00050.safetensors", + "model.layers.30.mlp.gate_proj.weight": "model-00011-of-00050.safetensors", + "model.layers.30.mlp.up_proj.weight": "model-00012-of-00050.safetensors", + "model.layers.30.post_attention_layernorm.weight": "model-00012-of-00050.safetensors", + "model.layers.30.self_attn.k_proj.weight": "model-00011-of-00050.safetensors", + "model.layers.30.self_attn.o_proj.weight": "model-00011-of-00050.safetensors", + "model.layers.30.self_attn.q_proj.weight": "model-00011-of-00050.safetensors", + "model.layers.30.self_attn.v_proj.weight": "model-00011-of-00050.safetensors", + "model.layers.31.input_layernorm.weight": "model-00012-of-00050.safetensors", + "model.layers.31.mlp.down_proj.weight": "model-00012-of-00050.safetensors", + "model.layers.31.mlp.gate_proj.weight": "model-00012-of-00050.safetensors", + "model.layers.31.mlp.up_proj.weight": "model-00012-of-00050.safetensors", + "model.layers.31.post_attention_layernorm.weight": "model-00012-of-00050.safetensors", + "model.layers.31.self_attn.k_proj.weight": "model-00012-of-00050.safetensors", + "model.layers.31.self_attn.o_proj.weight": "model-00012-of-00050.safetensors", + "model.layers.31.self_attn.q_proj.weight": "model-00012-of-00050.safetensors", + "model.layers.31.self_attn.v_proj.weight": "model-00012-of-00050.safetensors", + "model.layers.32.input_layernorm.weight": "model-00012-of-00050.safetensors", + "model.layers.32.mlp.down_proj.weight": "model-00012-of-00050.safetensors", + "model.layers.32.mlp.gate_proj.weight": "model-00012-of-00050.safetensors", + "model.layers.32.mlp.up_proj.weight": "model-00012-of-00050.safetensors", + "model.layers.32.post_attention_layernorm.weight": "model-00012-of-00050.safetensors", + "model.layers.32.self_attn.k_proj.weight": "model-00012-of-00050.safetensors", + "model.layers.32.self_attn.o_proj.weight": "model-00012-of-00050.safetensors", + "model.layers.32.self_attn.q_proj.weight": "model-00012-of-00050.safetensors", + "model.layers.32.self_attn.v_proj.weight": "model-00012-of-00050.safetensors", + "model.layers.33.input_layernorm.weight": "model-00013-of-00050.safetensors", + "model.layers.33.mlp.down_proj.weight": "model-00013-of-00050.safetensors", + "model.layers.33.mlp.gate_proj.weight": "model-00013-of-00050.safetensors", + "model.layers.33.mlp.up_proj.weight": "model-00013-of-00050.safetensors", + "model.layers.33.post_attention_layernorm.weight": "model-00013-of-00050.safetensors", + "model.layers.33.self_attn.k_proj.weight": "model-00012-of-00050.safetensors", + "model.layers.33.self_attn.o_proj.weight": "model-00012-of-00050.safetensors", + "model.layers.33.self_attn.q_proj.weight": "model-00012-of-00050.safetensors", + "model.layers.33.self_attn.v_proj.weight": "model-00012-of-00050.safetensors", + "model.layers.34.input_layernorm.weight": "model-00013-of-00050.safetensors", + "model.layers.34.mlp.down_proj.weight": "model-00013-of-00050.safetensors", + "model.layers.34.mlp.gate_proj.weight": "model-00013-of-00050.safetensors", + "model.layers.34.mlp.up_proj.weight": "model-00013-of-00050.safetensors", + "model.layers.34.post_attention_layernorm.weight": "model-00013-of-00050.safetensors", + "model.layers.34.self_attn.k_proj.weight": "model-00013-of-00050.safetensors", + "model.layers.34.self_attn.o_proj.weight": "model-00013-of-00050.safetensors", + "model.layers.34.self_attn.q_proj.weight": "model-00013-of-00050.safetensors", + "model.layers.34.self_attn.v_proj.weight": "model-00013-of-00050.safetensors", + "model.layers.35.input_layernorm.weight": "model-00013-of-00050.safetensors", + "model.layers.35.mlp.down_proj.weight": "model-00013-of-00050.safetensors", + "model.layers.35.mlp.gate_proj.weight": "model-00013-of-00050.safetensors", + "model.layers.35.mlp.up_proj.weight": "model-00013-of-00050.safetensors", + "model.layers.35.post_attention_layernorm.weight": "model-00013-of-00050.safetensors", + "model.layers.35.self_attn.k_proj.weight": "model-00013-of-00050.safetensors", + "model.layers.35.self_attn.o_proj.weight": "model-00013-of-00050.safetensors", + "model.layers.35.self_attn.q_proj.weight": "model-00013-of-00050.safetensors", + "model.layers.35.self_attn.v_proj.weight": "model-00013-of-00050.safetensors", + "model.layers.36.input_layernorm.weight": "model-00014-of-00050.safetensors", + "model.layers.36.mlp.down_proj.weight": "model-00014-of-00050.safetensors", + "model.layers.36.mlp.gate_proj.weight": "model-00014-of-00050.safetensors", + "model.layers.36.mlp.up_proj.weight": "model-00014-of-00050.safetensors", + "model.layers.36.post_attention_layernorm.weight": "model-00014-of-00050.safetensors", + "model.layers.36.self_attn.k_proj.weight": "model-00013-of-00050.safetensors", + "model.layers.36.self_attn.o_proj.weight": "model-00014-of-00050.safetensors", + "model.layers.36.self_attn.q_proj.weight": "model-00013-of-00050.safetensors", + "model.layers.36.self_attn.v_proj.weight": "model-00013-of-00050.safetensors", + "model.layers.37.input_layernorm.weight": "model-00014-of-00050.safetensors", + "model.layers.37.mlp.down_proj.weight": "model-00014-of-00050.safetensors", + "model.layers.37.mlp.gate_proj.weight": "model-00014-of-00050.safetensors", + "model.layers.37.mlp.up_proj.weight": "model-00014-of-00050.safetensors", + "model.layers.37.post_attention_layernorm.weight": "model-00014-of-00050.safetensors", + "model.layers.37.self_attn.k_proj.weight": "model-00014-of-00050.safetensors", + "model.layers.37.self_attn.o_proj.weight": "model-00014-of-00050.safetensors", + "model.layers.37.self_attn.q_proj.weight": "model-00014-of-00050.safetensors", + "model.layers.37.self_attn.v_proj.weight": "model-00014-of-00050.safetensors", + "model.layers.38.input_layernorm.weight": "model-00014-of-00050.safetensors", + "model.layers.38.mlp.down_proj.weight": "model-00014-of-00050.safetensors", + "model.layers.38.mlp.gate_proj.weight": "model-00014-of-00050.safetensors", + "model.layers.38.mlp.up_proj.weight": "model-00014-of-00050.safetensors", + "model.layers.38.post_attention_layernorm.weight": "model-00014-of-00050.safetensors", + "model.layers.38.self_attn.k_proj.weight": "model-00014-of-00050.safetensors", + "model.layers.38.self_attn.o_proj.weight": "model-00014-of-00050.safetensors", + "model.layers.38.self_attn.q_proj.weight": "model-00014-of-00050.safetensors", + "model.layers.38.self_attn.v_proj.weight": "model-00014-of-00050.safetensors", + "model.layers.39.input_layernorm.weight": "model-00015-of-00050.safetensors", + "model.layers.39.mlp.down_proj.weight": "model-00015-of-00050.safetensors", + "model.layers.39.mlp.gate_proj.weight": "model-00015-of-00050.safetensors", + "model.layers.39.mlp.up_proj.weight": "model-00015-of-00050.safetensors", + "model.layers.39.post_attention_layernorm.weight": "model-00015-of-00050.safetensors", + "model.layers.39.self_attn.k_proj.weight": "model-00015-of-00050.safetensors", + "model.layers.39.self_attn.o_proj.weight": "model-00015-of-00050.safetensors", + "model.layers.39.self_attn.q_proj.weight": "model-00015-of-00050.safetensors", + "model.layers.39.self_attn.v_proj.weight": "model-00015-of-00050.safetensors", + "model.layers.4.input_layernorm.weight": "model-00002-of-00050.safetensors", + "model.layers.4.mlp.down_proj.weight": "model-00002-of-00050.safetensors", + "model.layers.4.mlp.gate_proj.weight": "model-00002-of-00050.safetensors", + "model.layers.4.mlp.up_proj.weight": "model-00002-of-00050.safetensors", + "model.layers.4.post_attention_layernorm.weight": "model-00002-of-00050.safetensors", + "model.layers.4.self_attn.k_proj.weight": "model-00002-of-00050.safetensors", + "model.layers.4.self_attn.o_proj.weight": "model-00002-of-00050.safetensors", + "model.layers.4.self_attn.q_proj.weight": "model-00002-of-00050.safetensors", + "model.layers.4.self_attn.v_proj.weight": "model-00002-of-00050.safetensors", + "model.layers.40.input_layernorm.weight": "model-00015-of-00050.safetensors", + "model.layers.40.mlp.down_proj.weight": "model-00015-of-00050.safetensors", + "model.layers.40.mlp.gate_proj.weight": "model-00015-of-00050.safetensors", + "model.layers.40.mlp.up_proj.weight": "model-00015-of-00050.safetensors", + "model.layers.40.post_attention_layernorm.weight": "model-00015-of-00050.safetensors", + "model.layers.40.self_attn.k_proj.weight": "model-00015-of-00050.safetensors", + "model.layers.40.self_attn.o_proj.weight": "model-00015-of-00050.safetensors", + "model.layers.40.self_attn.q_proj.weight": "model-00015-of-00050.safetensors", + "model.layers.40.self_attn.v_proj.weight": "model-00015-of-00050.safetensors", + "model.layers.41.input_layernorm.weight": "model-00016-of-00050.safetensors", + "model.layers.41.mlp.down_proj.weight": "model-00016-of-00050.safetensors", + "model.layers.41.mlp.gate_proj.weight": "model-00015-of-00050.safetensors", + "model.layers.41.mlp.up_proj.weight": "model-00015-of-00050.safetensors", + "model.layers.41.post_attention_layernorm.weight": "model-00016-of-00050.safetensors", + "model.layers.41.self_attn.k_proj.weight": "model-00015-of-00050.safetensors", + "model.layers.41.self_attn.o_proj.weight": "model-00015-of-00050.safetensors", + "model.layers.41.self_attn.q_proj.weight": "model-00015-of-00050.safetensors", + "model.layers.41.self_attn.v_proj.weight": "model-00015-of-00050.safetensors", + "model.layers.42.input_layernorm.weight": "model-00016-of-00050.safetensors", + "model.layers.42.mlp.down_proj.weight": "model-00016-of-00050.safetensors", + "model.layers.42.mlp.gate_proj.weight": "model-00016-of-00050.safetensors", + "model.layers.42.mlp.up_proj.weight": "model-00016-of-00050.safetensors", + "model.layers.42.post_attention_layernorm.weight": "model-00016-of-00050.safetensors", + "model.layers.42.self_attn.k_proj.weight": "model-00016-of-00050.safetensors", + "model.layers.42.self_attn.o_proj.weight": "model-00016-of-00050.safetensors", + "model.layers.42.self_attn.q_proj.weight": "model-00016-of-00050.safetensors", + "model.layers.42.self_attn.v_proj.weight": "model-00016-of-00050.safetensors", + "model.layers.43.input_layernorm.weight": "model-00016-of-00050.safetensors", + "model.layers.43.mlp.down_proj.weight": "model-00016-of-00050.safetensors", + "model.layers.43.mlp.gate_proj.weight": "model-00016-of-00050.safetensors", + "model.layers.43.mlp.up_proj.weight": "model-00016-of-00050.safetensors", + "model.layers.43.post_attention_layernorm.weight": "model-00016-of-00050.safetensors", + "model.layers.43.self_attn.k_proj.weight": "model-00016-of-00050.safetensors", + "model.layers.43.self_attn.o_proj.weight": "model-00016-of-00050.safetensors", + "model.layers.43.self_attn.q_proj.weight": "model-00016-of-00050.safetensors", + "model.layers.43.self_attn.v_proj.weight": "model-00016-of-00050.safetensors", + "model.layers.44.input_layernorm.weight": "model-00017-of-00050.safetensors", + "model.layers.44.mlp.down_proj.weight": "model-00017-of-00050.safetensors", + "model.layers.44.mlp.gate_proj.weight": "model-00016-of-00050.safetensors", + "model.layers.44.mlp.up_proj.weight": "model-00017-of-00050.safetensors", + "model.layers.44.post_attention_layernorm.weight": "model-00017-of-00050.safetensors", + "model.layers.44.self_attn.k_proj.weight": "model-00016-of-00050.safetensors", + "model.layers.44.self_attn.o_proj.weight": "model-00016-of-00050.safetensors", + "model.layers.44.self_attn.q_proj.weight": "model-00016-of-00050.safetensors", + "model.layers.44.self_attn.v_proj.weight": "model-00016-of-00050.safetensors", + "model.layers.45.input_layernorm.weight": "model-00017-of-00050.safetensors", + "model.layers.45.mlp.down_proj.weight": "model-00017-of-00050.safetensors", + "model.layers.45.mlp.gate_proj.weight": "model-00017-of-00050.safetensors", + "model.layers.45.mlp.up_proj.weight": "model-00017-of-00050.safetensors", + "model.layers.45.post_attention_layernorm.weight": "model-00017-of-00050.safetensors", + "model.layers.45.self_attn.k_proj.weight": "model-00017-of-00050.safetensors", + "model.layers.45.self_attn.o_proj.weight": "model-00017-of-00050.safetensors", + "model.layers.45.self_attn.q_proj.weight": "model-00017-of-00050.safetensors", + "model.layers.45.self_attn.v_proj.weight": "model-00017-of-00050.safetensors", + "model.layers.46.input_layernorm.weight": "model-00017-of-00050.safetensors", + "model.layers.46.mlp.down_proj.weight": "model-00017-of-00050.safetensors", + "model.layers.46.mlp.gate_proj.weight": "model-00017-of-00050.safetensors", + "model.layers.46.mlp.up_proj.weight": "model-00017-of-00050.safetensors", + "model.layers.46.post_attention_layernorm.weight": "model-00017-of-00050.safetensors", + "model.layers.46.self_attn.k_proj.weight": "model-00017-of-00050.safetensors", + "model.layers.46.self_attn.o_proj.weight": "model-00017-of-00050.safetensors", + "model.layers.46.self_attn.q_proj.weight": "model-00017-of-00050.safetensors", + "model.layers.46.self_attn.v_proj.weight": "model-00017-of-00050.safetensors", + "model.layers.47.input_layernorm.weight": "model-00018-of-00050.safetensors", + "model.layers.47.mlp.down_proj.weight": "model-00018-of-00050.safetensors", + "model.layers.47.mlp.gate_proj.weight": "model-00018-of-00050.safetensors", + "model.layers.47.mlp.up_proj.weight": "model-00018-of-00050.safetensors", + "model.layers.47.post_attention_layernorm.weight": "model-00018-of-00050.safetensors", + "model.layers.47.self_attn.k_proj.weight": "model-00017-of-00050.safetensors", + "model.layers.47.self_attn.o_proj.weight": "model-00017-of-00050.safetensors", + "model.layers.47.self_attn.q_proj.weight": "model-00017-of-00050.safetensors", + "model.layers.47.self_attn.v_proj.weight": "model-00017-of-00050.safetensors", + "model.layers.48.input_layernorm.weight": "model-00018-of-00050.safetensors", + "model.layers.48.mlp.down_proj.weight": "model-00018-of-00050.safetensors", + "model.layers.48.mlp.gate_proj.weight": "model-00018-of-00050.safetensors", + "model.layers.48.mlp.up_proj.weight": "model-00018-of-00050.safetensors", + "model.layers.48.post_attention_layernorm.weight": "model-00018-of-00050.safetensors", + "model.layers.48.self_attn.k_proj.weight": "model-00018-of-00050.safetensors", + "model.layers.48.self_attn.o_proj.weight": "model-00018-of-00050.safetensors", + "model.layers.48.self_attn.q_proj.weight": "model-00018-of-00050.safetensors", + "model.layers.48.self_attn.v_proj.weight": "model-00018-of-00050.safetensors", + "model.layers.49.input_layernorm.weight": "model-00018-of-00050.safetensors", + "model.layers.49.mlp.down_proj.weight": "model-00018-of-00050.safetensors", + "model.layers.49.mlp.gate_proj.weight": "model-00018-of-00050.safetensors", + "model.layers.49.mlp.up_proj.weight": "model-00018-of-00050.safetensors", + "model.layers.49.post_attention_layernorm.weight": "model-00018-of-00050.safetensors", + "model.layers.49.self_attn.k_proj.weight": "model-00018-of-00050.safetensors", + "model.layers.49.self_attn.o_proj.weight": "model-00018-of-00050.safetensors", + "model.layers.49.self_attn.q_proj.weight": "model-00018-of-00050.safetensors", + "model.layers.49.self_attn.v_proj.weight": "model-00018-of-00050.safetensors", + "model.layers.5.input_layernorm.weight": "model-00003-of-00050.safetensors", + "model.layers.5.mlp.down_proj.weight": "model-00003-of-00050.safetensors", + "model.layers.5.mlp.gate_proj.weight": "model-00003-of-00050.safetensors", + "model.layers.5.mlp.up_proj.weight": "model-00003-of-00050.safetensors", + "model.layers.5.post_attention_layernorm.weight": "model-00003-of-00050.safetensors", + "model.layers.5.self_attn.k_proj.weight": "model-00002-of-00050.safetensors", + "model.layers.5.self_attn.o_proj.weight": "model-00002-of-00050.safetensors", + "model.layers.5.self_attn.q_proj.weight": "model-00002-of-00050.safetensors", + "model.layers.5.self_attn.v_proj.weight": "model-00002-of-00050.safetensors", + "model.layers.50.input_layernorm.weight": "model-00019-of-00050.safetensors", + "model.layers.50.mlp.down_proj.weight": "model-00019-of-00050.safetensors", + "model.layers.50.mlp.gate_proj.weight": "model-00019-of-00050.safetensors", + "model.layers.50.mlp.up_proj.weight": "model-00019-of-00050.safetensors", + "model.layers.50.post_attention_layernorm.weight": "model-00019-of-00050.safetensors", + "model.layers.50.self_attn.k_proj.weight": "model-00018-of-00050.safetensors", + "model.layers.50.self_attn.o_proj.weight": "model-00019-of-00050.safetensors", + "model.layers.50.self_attn.q_proj.weight": "model-00018-of-00050.safetensors", + "model.layers.50.self_attn.v_proj.weight": "model-00018-of-00050.safetensors", + "model.layers.51.input_layernorm.weight": "model-00019-of-00050.safetensors", + "model.layers.51.mlp.down_proj.weight": "model-00019-of-00050.safetensors", + "model.layers.51.mlp.gate_proj.weight": "model-00019-of-00050.safetensors", + "model.layers.51.mlp.up_proj.weight": "model-00019-of-00050.safetensors", + "model.layers.51.post_attention_layernorm.weight": "model-00019-of-00050.safetensors", + "model.layers.51.self_attn.k_proj.weight": "model-00019-of-00050.safetensors", + "model.layers.51.self_attn.o_proj.weight": "model-00019-of-00050.safetensors", + "model.layers.51.self_attn.q_proj.weight": "model-00019-of-00050.safetensors", + "model.layers.51.self_attn.v_proj.weight": "model-00019-of-00050.safetensors", + "model.layers.52.input_layernorm.weight": "model-00019-of-00050.safetensors", + "model.layers.52.mlp.down_proj.weight": "model-00019-of-00050.safetensors", + "model.layers.52.mlp.gate_proj.weight": "model-00019-of-00050.safetensors", + "model.layers.52.mlp.up_proj.weight": "model-00019-of-00050.safetensors", + "model.layers.52.post_attention_layernorm.weight": "model-00019-of-00050.safetensors", + "model.layers.52.self_attn.k_proj.weight": "model-00019-of-00050.safetensors", + "model.layers.52.self_attn.o_proj.weight": "model-00019-of-00050.safetensors", + "model.layers.52.self_attn.q_proj.weight": "model-00019-of-00050.safetensors", + "model.layers.52.self_attn.v_proj.weight": "model-00019-of-00050.safetensors", + "model.layers.53.input_layernorm.weight": "model-00020-of-00050.safetensors", + "model.layers.53.mlp.down_proj.weight": "model-00020-of-00050.safetensors", + "model.layers.53.mlp.gate_proj.weight": "model-00020-of-00050.safetensors", + "model.layers.53.mlp.up_proj.weight": "model-00020-of-00050.safetensors", + "model.layers.53.post_attention_layernorm.weight": "model-00020-of-00050.safetensors", + "model.layers.53.self_attn.k_proj.weight": "model-00020-of-00050.safetensors", + "model.layers.53.self_attn.o_proj.weight": "model-00020-of-00050.safetensors", + "model.layers.53.self_attn.q_proj.weight": "model-00020-of-00050.safetensors", + "model.layers.53.self_attn.v_proj.weight": "model-00020-of-00050.safetensors", + "model.layers.54.input_layernorm.weight": "model-00020-of-00050.safetensors", + "model.layers.54.mlp.down_proj.weight": "model-00020-of-00050.safetensors", + "model.layers.54.mlp.gate_proj.weight": "model-00020-of-00050.safetensors", + "model.layers.54.mlp.up_proj.weight": "model-00020-of-00050.safetensors", + "model.layers.54.post_attention_layernorm.weight": "model-00020-of-00050.safetensors", + "model.layers.54.self_attn.k_proj.weight": "model-00020-of-00050.safetensors", + "model.layers.54.self_attn.o_proj.weight": "model-00020-of-00050.safetensors", + "model.layers.54.self_attn.q_proj.weight": "model-00020-of-00050.safetensors", + "model.layers.54.self_attn.v_proj.weight": "model-00020-of-00050.safetensors", + "model.layers.55.input_layernorm.weight": "model-00021-of-00050.safetensors", + "model.layers.55.mlp.down_proj.weight": "model-00021-of-00050.safetensors", + "model.layers.55.mlp.gate_proj.weight": "model-00020-of-00050.safetensors", + "model.layers.55.mlp.up_proj.weight": "model-00020-of-00050.safetensors", + "model.layers.55.post_attention_layernorm.weight": "model-00021-of-00050.safetensors", + "model.layers.55.self_attn.k_proj.weight": "model-00020-of-00050.safetensors", + "model.layers.55.self_attn.o_proj.weight": "model-00020-of-00050.safetensors", + "model.layers.55.self_attn.q_proj.weight": "model-00020-of-00050.safetensors", + "model.layers.55.self_attn.v_proj.weight": "model-00020-of-00050.safetensors", + "model.layers.56.input_layernorm.weight": "model-00021-of-00050.safetensors", + "model.layers.56.mlp.down_proj.weight": "model-00021-of-00050.safetensors", + "model.layers.56.mlp.gate_proj.weight": "model-00021-of-00050.safetensors", + "model.layers.56.mlp.up_proj.weight": "model-00021-of-00050.safetensors", + "model.layers.56.post_attention_layernorm.weight": "model-00021-of-00050.safetensors", + "model.layers.56.self_attn.k_proj.weight": "model-00021-of-00050.safetensors", + "model.layers.56.self_attn.o_proj.weight": "model-00021-of-00050.safetensors", + "model.layers.56.self_attn.q_proj.weight": "model-00021-of-00050.safetensors", + "model.layers.56.self_attn.v_proj.weight": "model-00021-of-00050.safetensors", + "model.layers.57.input_layernorm.weight": "model-00021-of-00050.safetensors", + "model.layers.57.mlp.down_proj.weight": "model-00021-of-00050.safetensors", + "model.layers.57.mlp.gate_proj.weight": "model-00021-of-00050.safetensors", + "model.layers.57.mlp.up_proj.weight": "model-00021-of-00050.safetensors", + "model.layers.57.post_attention_layernorm.weight": "model-00021-of-00050.safetensors", + "model.layers.57.self_attn.k_proj.weight": "model-00021-of-00050.safetensors", + "model.layers.57.self_attn.o_proj.weight": "model-00021-of-00050.safetensors", + "model.layers.57.self_attn.q_proj.weight": "model-00021-of-00050.safetensors", + "model.layers.57.self_attn.v_proj.weight": "model-00021-of-00050.safetensors", + "model.layers.58.input_layernorm.weight": "model-00022-of-00050.safetensors", + "model.layers.58.mlp.down_proj.weight": "model-00022-of-00050.safetensors", + "model.layers.58.mlp.gate_proj.weight": "model-00021-of-00050.safetensors", + "model.layers.58.mlp.up_proj.weight": "model-00022-of-00050.safetensors", + "model.layers.58.post_attention_layernorm.weight": "model-00022-of-00050.safetensors", + "model.layers.58.self_attn.k_proj.weight": "model-00021-of-00050.safetensors", + "model.layers.58.self_attn.o_proj.weight": "model-00021-of-00050.safetensors", + "model.layers.58.self_attn.q_proj.weight": "model-00021-of-00050.safetensors", + "model.layers.58.self_attn.v_proj.weight": "model-00021-of-00050.safetensors", + "model.layers.59.input_layernorm.weight": "model-00022-of-00050.safetensors", + "model.layers.59.mlp.down_proj.weight": "model-00022-of-00050.safetensors", + "model.layers.59.mlp.gate_proj.weight": "model-00022-of-00050.safetensors", + "model.layers.59.mlp.up_proj.weight": "model-00022-of-00050.safetensors", + "model.layers.59.post_attention_layernorm.weight": "model-00022-of-00050.safetensors", + "model.layers.59.self_attn.k_proj.weight": "model-00022-of-00050.safetensors", + "model.layers.59.self_attn.o_proj.weight": "model-00022-of-00050.safetensors", + "model.layers.59.self_attn.q_proj.weight": "model-00022-of-00050.safetensors", + "model.layers.59.self_attn.v_proj.weight": "model-00022-of-00050.safetensors", + "model.layers.6.input_layernorm.weight": "model-00003-of-00050.safetensors", + "model.layers.6.mlp.down_proj.weight": "model-00003-of-00050.safetensors", + "model.layers.6.mlp.gate_proj.weight": "model-00003-of-00050.safetensors", + "model.layers.6.mlp.up_proj.weight": "model-00003-of-00050.safetensors", + "model.layers.6.post_attention_layernorm.weight": "model-00003-of-00050.safetensors", + "model.layers.6.self_attn.k_proj.weight": "model-00003-of-00050.safetensors", + "model.layers.6.self_attn.o_proj.weight": "model-00003-of-00050.safetensors", + "model.layers.6.self_attn.q_proj.weight": "model-00003-of-00050.safetensors", + "model.layers.6.self_attn.v_proj.weight": "model-00003-of-00050.safetensors", + "model.layers.60.input_layernorm.weight": "model-00022-of-00050.safetensors", + "model.layers.60.mlp.down_proj.weight": "model-00022-of-00050.safetensors", + "model.layers.60.mlp.gate_proj.weight": "model-00022-of-00050.safetensors", + "model.layers.60.mlp.up_proj.weight": "model-00022-of-00050.safetensors", + "model.layers.60.post_attention_layernorm.weight": "model-00022-of-00050.safetensors", + "model.layers.60.self_attn.k_proj.weight": "model-00022-of-00050.safetensors", + "model.layers.60.self_attn.o_proj.weight": "model-00022-of-00050.safetensors", + "model.layers.60.self_attn.q_proj.weight": "model-00022-of-00050.safetensors", + "model.layers.60.self_attn.v_proj.weight": "model-00022-of-00050.safetensors", + "model.layers.61.input_layernorm.weight": "model-00023-of-00050.safetensors", + "model.layers.61.mlp.down_proj.weight": "model-00023-of-00050.safetensors", + "model.layers.61.mlp.gate_proj.weight": "model-00023-of-00050.safetensors", + "model.layers.61.mlp.up_proj.weight": "model-00023-of-00050.safetensors", + "model.layers.61.post_attention_layernorm.weight": "model-00023-of-00050.safetensors", + "model.layers.61.self_attn.k_proj.weight": "model-00022-of-00050.safetensors", + "model.layers.61.self_attn.o_proj.weight": "model-00022-of-00050.safetensors", + "model.layers.61.self_attn.q_proj.weight": "model-00022-of-00050.safetensors", + "model.layers.61.self_attn.v_proj.weight": "model-00022-of-00050.safetensors", + "model.layers.62.input_layernorm.weight": "model-00023-of-00050.safetensors", + "model.layers.62.mlp.down_proj.weight": "model-00023-of-00050.safetensors", + "model.layers.62.mlp.gate_proj.weight": "model-00023-of-00050.safetensors", + "model.layers.62.mlp.up_proj.weight": "model-00023-of-00050.safetensors", + "model.layers.62.post_attention_layernorm.weight": "model-00023-of-00050.safetensors", + "model.layers.62.self_attn.k_proj.weight": "model-00023-of-00050.safetensors", + "model.layers.62.self_attn.o_proj.weight": "model-00023-of-00050.safetensors", + "model.layers.62.self_attn.q_proj.weight": "model-00023-of-00050.safetensors", + "model.layers.62.self_attn.v_proj.weight": "model-00023-of-00050.safetensors", + "model.layers.63.input_layernorm.weight": "model-00023-of-00050.safetensors", + "model.layers.63.mlp.down_proj.weight": "model-00023-of-00050.safetensors", + "model.layers.63.mlp.gate_proj.weight": "model-00023-of-00050.safetensors", + "model.layers.63.mlp.up_proj.weight": "model-00023-of-00050.safetensors", + "model.layers.63.post_attention_layernorm.weight": "model-00023-of-00050.safetensors", + "model.layers.63.self_attn.k_proj.weight": "model-00023-of-00050.safetensors", + "model.layers.63.self_attn.o_proj.weight": "model-00023-of-00050.safetensors", + "model.layers.63.self_attn.q_proj.weight": "model-00023-of-00050.safetensors", + "model.layers.63.self_attn.v_proj.weight": "model-00023-of-00050.safetensors", + "model.layers.64.input_layernorm.weight": "model-00024-of-00050.safetensors", + "model.layers.64.mlp.down_proj.weight": "model-00024-of-00050.safetensors", + "model.layers.64.mlp.gate_proj.weight": "model-00024-of-00050.safetensors", + "model.layers.64.mlp.up_proj.weight": "model-00024-of-00050.safetensors", + "model.layers.64.post_attention_layernorm.weight": "model-00024-of-00050.safetensors", + "model.layers.64.self_attn.k_proj.weight": "model-00023-of-00050.safetensors", + "model.layers.64.self_attn.o_proj.weight": "model-00024-of-00050.safetensors", + "model.layers.64.self_attn.q_proj.weight": "model-00023-of-00050.safetensors", + "model.layers.64.self_attn.v_proj.weight": "model-00023-of-00050.safetensors", + "model.layers.65.input_layernorm.weight": "model-00024-of-00050.safetensors", + "model.layers.65.mlp.down_proj.weight": "model-00024-of-00050.safetensors", + "model.layers.65.mlp.gate_proj.weight": "model-00024-of-00050.safetensors", + "model.layers.65.mlp.up_proj.weight": "model-00024-of-00050.safetensors", + "model.layers.65.post_attention_layernorm.weight": "model-00024-of-00050.safetensors", + "model.layers.65.self_attn.k_proj.weight": "model-00024-of-00050.safetensors", + "model.layers.65.self_attn.o_proj.weight": "model-00024-of-00050.safetensors", + "model.layers.65.self_attn.q_proj.weight": "model-00024-of-00050.safetensors", + "model.layers.65.self_attn.v_proj.weight": "model-00024-of-00050.safetensors", + "model.layers.66.input_layernorm.weight": "model-00024-of-00050.safetensors", + "model.layers.66.mlp.down_proj.weight": "model-00024-of-00050.safetensors", + "model.layers.66.mlp.gate_proj.weight": "model-00024-of-00050.safetensors", + "model.layers.66.mlp.up_proj.weight": "model-00024-of-00050.safetensors", + "model.layers.66.post_attention_layernorm.weight": "model-00024-of-00050.safetensors", + "model.layers.66.self_attn.k_proj.weight": "model-00024-of-00050.safetensors", + "model.layers.66.self_attn.o_proj.weight": "model-00024-of-00050.safetensors", + "model.layers.66.self_attn.q_proj.weight": "model-00024-of-00050.safetensors", + "model.layers.66.self_attn.v_proj.weight": "model-00024-of-00050.safetensors", + "model.layers.67.input_layernorm.weight": "model-00025-of-00050.safetensors", + "model.layers.67.mlp.down_proj.weight": "model-00025-of-00050.safetensors", + "model.layers.67.mlp.gate_proj.weight": "model-00025-of-00050.safetensors", + "model.layers.67.mlp.up_proj.weight": "model-00025-of-00050.safetensors", + "model.layers.67.post_attention_layernorm.weight": "model-00025-of-00050.safetensors", + "model.layers.67.self_attn.k_proj.weight": "model-00025-of-00050.safetensors", + "model.layers.67.self_attn.o_proj.weight": "model-00025-of-00050.safetensors", + "model.layers.67.self_attn.q_proj.weight": "model-00025-of-00050.safetensors", + "model.layers.67.self_attn.v_proj.weight": "model-00025-of-00050.safetensors", + "model.layers.68.input_layernorm.weight": "model-00025-of-00050.safetensors", + "model.layers.68.mlp.down_proj.weight": "model-00025-of-00050.safetensors", + "model.layers.68.mlp.gate_proj.weight": "model-00025-of-00050.safetensors", + "model.layers.68.mlp.up_proj.weight": "model-00025-of-00050.safetensors", + "model.layers.68.post_attention_layernorm.weight": "model-00025-of-00050.safetensors", + "model.layers.68.self_attn.k_proj.weight": "model-00025-of-00050.safetensors", + "model.layers.68.self_attn.o_proj.weight": "model-00025-of-00050.safetensors", + "model.layers.68.self_attn.q_proj.weight": "model-00025-of-00050.safetensors", + "model.layers.68.self_attn.v_proj.weight": "model-00025-of-00050.safetensors", + "model.layers.69.input_layernorm.weight": "model-00026-of-00050.safetensors", + "model.layers.69.mlp.down_proj.weight": "model-00026-of-00050.safetensors", + "model.layers.69.mlp.gate_proj.weight": "model-00025-of-00050.safetensors", + "model.layers.69.mlp.up_proj.weight": "model-00025-of-00050.safetensors", + "model.layers.69.post_attention_layernorm.weight": "model-00026-of-00050.safetensors", + "model.layers.69.self_attn.k_proj.weight": "model-00025-of-00050.safetensors", + "model.layers.69.self_attn.o_proj.weight": "model-00025-of-00050.safetensors", + "model.layers.69.self_attn.q_proj.weight": "model-00025-of-00050.safetensors", + "model.layers.69.self_attn.v_proj.weight": "model-00025-of-00050.safetensors", + "model.layers.7.input_layernorm.weight": "model-00003-of-00050.safetensors", + "model.layers.7.mlp.down_proj.weight": "model-00003-of-00050.safetensors", + "model.layers.7.mlp.gate_proj.weight": "model-00003-of-00050.safetensors", + "model.layers.7.mlp.up_proj.weight": "model-00003-of-00050.safetensors", + "model.layers.7.post_attention_layernorm.weight": "model-00003-of-00050.safetensors", + "model.layers.7.self_attn.k_proj.weight": "model-00003-of-00050.safetensors", + "model.layers.7.self_attn.o_proj.weight": "model-00003-of-00050.safetensors", + "model.layers.7.self_attn.q_proj.weight": "model-00003-of-00050.safetensors", + "model.layers.7.self_attn.v_proj.weight": "model-00003-of-00050.safetensors", + "model.layers.70.input_layernorm.weight": "model-00026-of-00050.safetensors", + "model.layers.70.mlp.down_proj.weight": "model-00026-of-00050.safetensors", + "model.layers.70.mlp.gate_proj.weight": "model-00026-of-00050.safetensors", + "model.layers.70.mlp.up_proj.weight": "model-00026-of-00050.safetensors", + "model.layers.70.post_attention_layernorm.weight": "model-00026-of-00050.safetensors", + "model.layers.70.self_attn.k_proj.weight": "model-00026-of-00050.safetensors", + "model.layers.70.self_attn.o_proj.weight": "model-00026-of-00050.safetensors", + "model.layers.70.self_attn.q_proj.weight": "model-00026-of-00050.safetensors", + "model.layers.70.self_attn.v_proj.weight": "model-00026-of-00050.safetensors", + "model.layers.71.input_layernorm.weight": "model-00026-of-00050.safetensors", + "model.layers.71.mlp.down_proj.weight": "model-00026-of-00050.safetensors", + "model.layers.71.mlp.gate_proj.weight": "model-00026-of-00050.safetensors", + "model.layers.71.mlp.up_proj.weight": "model-00026-of-00050.safetensors", + "model.layers.71.post_attention_layernorm.weight": "model-00026-of-00050.safetensors", + "model.layers.71.self_attn.k_proj.weight": "model-00026-of-00050.safetensors", + "model.layers.71.self_attn.o_proj.weight": "model-00026-of-00050.safetensors", + "model.layers.71.self_attn.q_proj.weight": "model-00026-of-00050.safetensors", + "model.layers.71.self_attn.v_proj.weight": "model-00026-of-00050.safetensors", + "model.layers.72.input_layernorm.weight": "model-00027-of-00050.safetensors", + "model.layers.72.mlp.down_proj.weight": "model-00027-of-00050.safetensors", + "model.layers.72.mlp.gate_proj.weight": "model-00026-of-00050.safetensors", + "model.layers.72.mlp.up_proj.weight": "model-00027-of-00050.safetensors", + "model.layers.72.post_attention_layernorm.weight": "model-00027-of-00050.safetensors", + "model.layers.72.self_attn.k_proj.weight": "model-00026-of-00050.safetensors", + "model.layers.72.self_attn.o_proj.weight": "model-00026-of-00050.safetensors", + "model.layers.72.self_attn.q_proj.weight": "model-00026-of-00050.safetensors", + "model.layers.72.self_attn.v_proj.weight": "model-00026-of-00050.safetensors", + "model.layers.73.input_layernorm.weight": "model-00027-of-00050.safetensors", + "model.layers.73.mlp.down_proj.weight": "model-00027-of-00050.safetensors", + "model.layers.73.mlp.gate_proj.weight": "model-00027-of-00050.safetensors", + "model.layers.73.mlp.up_proj.weight": "model-00027-of-00050.safetensors", + "model.layers.73.post_attention_layernorm.weight": "model-00027-of-00050.safetensors", + "model.layers.73.self_attn.k_proj.weight": "model-00027-of-00050.safetensors", + "model.layers.73.self_attn.o_proj.weight": "model-00027-of-00050.safetensors", + "model.layers.73.self_attn.q_proj.weight": "model-00027-of-00050.safetensors", + "model.layers.73.self_attn.v_proj.weight": "model-00027-of-00050.safetensors", + "model.layers.74.input_layernorm.weight": "model-00027-of-00050.safetensors", + "model.layers.74.mlp.down_proj.weight": "model-00027-of-00050.safetensors", + "model.layers.74.mlp.gate_proj.weight": "model-00027-of-00050.safetensors", + "model.layers.74.mlp.up_proj.weight": "model-00027-of-00050.safetensors", + "model.layers.74.post_attention_layernorm.weight": "model-00027-of-00050.safetensors", + "model.layers.74.self_attn.k_proj.weight": "model-00027-of-00050.safetensors", + "model.layers.74.self_attn.o_proj.weight": "model-00027-of-00050.safetensors", + "model.layers.74.self_attn.q_proj.weight": "model-00027-of-00050.safetensors", + "model.layers.74.self_attn.v_proj.weight": "model-00027-of-00050.safetensors", + "model.layers.75.input_layernorm.weight": "model-00028-of-00050.safetensors", + "model.layers.75.mlp.down_proj.weight": "model-00028-of-00050.safetensors", + "model.layers.75.mlp.gate_proj.weight": "model-00028-of-00050.safetensors", + "model.layers.75.mlp.up_proj.weight": "model-00028-of-00050.safetensors", + "model.layers.75.post_attention_layernorm.weight": "model-00028-of-00050.safetensors", + "model.layers.75.self_attn.k_proj.weight": "model-00027-of-00050.safetensors", + "model.layers.75.self_attn.o_proj.weight": "model-00027-of-00050.safetensors", + "model.layers.75.self_attn.q_proj.weight": "model-00027-of-00050.safetensors", + "model.layers.75.self_attn.v_proj.weight": "model-00027-of-00050.safetensors", + "model.layers.76.input_layernorm.weight": "model-00028-of-00050.safetensors", + "model.layers.76.mlp.down_proj.weight": "model-00028-of-00050.safetensors", + "model.layers.76.mlp.gate_proj.weight": "model-00028-of-00050.safetensors", + "model.layers.76.mlp.up_proj.weight": "model-00028-of-00050.safetensors", + "model.layers.76.post_attention_layernorm.weight": "model-00028-of-00050.safetensors", + "model.layers.76.self_attn.k_proj.weight": "model-00028-of-00050.safetensors", + "model.layers.76.self_attn.o_proj.weight": "model-00028-of-00050.safetensors", + "model.layers.76.self_attn.q_proj.weight": "model-00028-of-00050.safetensors", + "model.layers.76.self_attn.v_proj.weight": "model-00028-of-00050.safetensors", + "model.layers.77.input_layernorm.weight": "model-00028-of-00050.safetensors", + "model.layers.77.mlp.down_proj.weight": "model-00028-of-00050.safetensors", + "model.layers.77.mlp.gate_proj.weight": "model-00028-of-00050.safetensors", + "model.layers.77.mlp.up_proj.weight": "model-00028-of-00050.safetensors", + "model.layers.77.post_attention_layernorm.weight": "model-00028-of-00050.safetensors", + "model.layers.77.self_attn.k_proj.weight": "model-00028-of-00050.safetensors", + "model.layers.77.self_attn.o_proj.weight": "model-00028-of-00050.safetensors", + "model.layers.77.self_attn.q_proj.weight": "model-00028-of-00050.safetensors", + "model.layers.77.self_attn.v_proj.weight": "model-00028-of-00050.safetensors", + "model.layers.78.input_layernorm.weight": "model-00029-of-00050.safetensors", + "model.layers.78.mlp.down_proj.weight": "model-00029-of-00050.safetensors", + "model.layers.78.mlp.gate_proj.weight": "model-00029-of-00050.safetensors", + "model.layers.78.mlp.up_proj.weight": "model-00029-of-00050.safetensors", + "model.layers.78.post_attention_layernorm.weight": "model-00029-of-00050.safetensors", + "model.layers.78.self_attn.k_proj.weight": "model-00028-of-00050.safetensors", + "model.layers.78.self_attn.o_proj.weight": "model-00029-of-00050.safetensors", + "model.layers.78.self_attn.q_proj.weight": "model-00028-of-00050.safetensors", + "model.layers.78.self_attn.v_proj.weight": "model-00028-of-00050.safetensors", + "model.layers.79.input_layernorm.weight": "model-00029-of-00050.safetensors", + "model.layers.79.mlp.down_proj.weight": "model-00029-of-00050.safetensors", + "model.layers.79.mlp.gate_proj.weight": "model-00029-of-00050.safetensors", + "model.layers.79.mlp.up_proj.weight": "model-00029-of-00050.safetensors", + "model.layers.79.post_attention_layernorm.weight": "model-00029-of-00050.safetensors", + "model.layers.79.self_attn.k_proj.weight": "model-00029-of-00050.safetensors", + "model.layers.79.self_attn.o_proj.weight": "model-00029-of-00050.safetensors", + "model.layers.79.self_attn.q_proj.weight": "model-00029-of-00050.safetensors", + "model.layers.79.self_attn.v_proj.weight": "model-00029-of-00050.safetensors", + "model.layers.8.input_layernorm.weight": "model-00004-of-00050.safetensors", + "model.layers.8.mlp.down_proj.weight": "model-00004-of-00050.safetensors", + "model.layers.8.mlp.gate_proj.weight": "model-00004-of-00050.safetensors", + "model.layers.8.mlp.up_proj.weight": "model-00004-of-00050.safetensors", + "model.layers.8.post_attention_layernorm.weight": "model-00004-of-00050.safetensors", + "model.layers.8.self_attn.k_proj.weight": "model-00003-of-00050.safetensors", + "model.layers.8.self_attn.o_proj.weight": "model-00004-of-00050.safetensors", + "model.layers.8.self_attn.q_proj.weight": "model-00003-of-00050.safetensors", + "model.layers.8.self_attn.v_proj.weight": "model-00003-of-00050.safetensors", + "model.layers.80.input_layernorm.weight": "model-00029-of-00050.safetensors", + "model.layers.80.mlp.down_proj.weight": "model-00029-of-00050.safetensors", + "model.layers.80.mlp.gate_proj.weight": "model-00029-of-00050.safetensors", + "model.layers.80.mlp.up_proj.weight": "model-00029-of-00050.safetensors", + "model.layers.80.post_attention_layernorm.weight": "model-00029-of-00050.safetensors", + "model.layers.80.self_attn.k_proj.weight": "model-00029-of-00050.safetensors", + "model.layers.80.self_attn.o_proj.weight": "model-00029-of-00050.safetensors", + "model.layers.80.self_attn.q_proj.weight": "model-00029-of-00050.safetensors", + "model.layers.80.self_attn.v_proj.weight": "model-00029-of-00050.safetensors", + "model.layers.81.input_layernorm.weight": "model-00030-of-00050.safetensors", + "model.layers.81.mlp.down_proj.weight": "model-00030-of-00050.safetensors", + "model.layers.81.mlp.gate_proj.weight": "model-00030-of-00050.safetensors", + "model.layers.81.mlp.up_proj.weight": "model-00030-of-00050.safetensors", + "model.layers.81.post_attention_layernorm.weight": "model-00030-of-00050.safetensors", + "model.layers.81.self_attn.k_proj.weight": "model-00030-of-00050.safetensors", + "model.layers.81.self_attn.o_proj.weight": "model-00030-of-00050.safetensors", + "model.layers.81.self_attn.q_proj.weight": "model-00030-of-00050.safetensors", + "model.layers.81.self_attn.v_proj.weight": "model-00030-of-00050.safetensors", + "model.layers.82.input_layernorm.weight": "model-00030-of-00050.safetensors", + "model.layers.82.mlp.down_proj.weight": "model-00030-of-00050.safetensors", + "model.layers.82.mlp.gate_proj.weight": "model-00030-of-00050.safetensors", + "model.layers.82.mlp.up_proj.weight": "model-00030-of-00050.safetensors", + "model.layers.82.post_attention_layernorm.weight": "model-00030-of-00050.safetensors", + "model.layers.82.self_attn.k_proj.weight": "model-00030-of-00050.safetensors", + "model.layers.82.self_attn.o_proj.weight": "model-00030-of-00050.safetensors", + "model.layers.82.self_attn.q_proj.weight": "model-00030-of-00050.safetensors", + "model.layers.82.self_attn.v_proj.weight": "model-00030-of-00050.safetensors", + "model.layers.83.input_layernorm.weight": "model-00031-of-00050.safetensors", + "model.layers.83.mlp.down_proj.weight": "model-00031-of-00050.safetensors", + "model.layers.83.mlp.gate_proj.weight": "model-00030-of-00050.safetensors", + "model.layers.83.mlp.up_proj.weight": "model-00030-of-00050.safetensors", + "model.layers.83.post_attention_layernorm.weight": "model-00031-of-00050.safetensors", + "model.layers.83.self_attn.k_proj.weight": "model-00030-of-00050.safetensors", + "model.layers.83.self_attn.o_proj.weight": "model-00030-of-00050.safetensors", + "model.layers.83.self_attn.q_proj.weight": "model-00030-of-00050.safetensors", + "model.layers.83.self_attn.v_proj.weight": "model-00030-of-00050.safetensors", + "model.layers.84.input_layernorm.weight": "model-00031-of-00050.safetensors", + "model.layers.84.mlp.down_proj.weight": "model-00031-of-00050.safetensors", + "model.layers.84.mlp.gate_proj.weight": "model-00031-of-00050.safetensors", + "model.layers.84.mlp.up_proj.weight": "model-00031-of-00050.safetensors", + "model.layers.84.post_attention_layernorm.weight": "model-00031-of-00050.safetensors", + "model.layers.84.self_attn.k_proj.weight": "model-00031-of-00050.safetensors", + "model.layers.84.self_attn.o_proj.weight": "model-00031-of-00050.safetensors", + "model.layers.84.self_attn.q_proj.weight": "model-00031-of-00050.safetensors", + "model.layers.84.self_attn.v_proj.weight": "model-00031-of-00050.safetensors", + "model.layers.85.input_layernorm.weight": "model-00031-of-00050.safetensors", + "model.layers.85.mlp.down_proj.weight": "model-00031-of-00050.safetensors", + "model.layers.85.mlp.gate_proj.weight": "model-00031-of-00050.safetensors", + "model.layers.85.mlp.up_proj.weight": "model-00031-of-00050.safetensors", + "model.layers.85.post_attention_layernorm.weight": "model-00031-of-00050.safetensors", + "model.layers.85.self_attn.k_proj.weight": "model-00031-of-00050.safetensors", + "model.layers.85.self_attn.o_proj.weight": "model-00031-of-00050.safetensors", + "model.layers.85.self_attn.q_proj.weight": "model-00031-of-00050.safetensors", + "model.layers.85.self_attn.v_proj.weight": "model-00031-of-00050.safetensors", + "model.layers.86.input_layernorm.weight": "model-00032-of-00050.safetensors", + "model.layers.86.mlp.down_proj.weight": "model-00032-of-00050.safetensors", + "model.layers.86.mlp.gate_proj.weight": "model-00031-of-00050.safetensors", + "model.layers.86.mlp.up_proj.weight": "model-00032-of-00050.safetensors", + "model.layers.86.post_attention_layernorm.weight": "model-00032-of-00050.safetensors", + "model.layers.86.self_attn.k_proj.weight": "model-00031-of-00050.safetensors", + "model.layers.86.self_attn.o_proj.weight": "model-00031-of-00050.safetensors", + "model.layers.86.self_attn.q_proj.weight": "model-00031-of-00050.safetensors", + "model.layers.86.self_attn.v_proj.weight": "model-00031-of-00050.safetensors", + "model.layers.87.input_layernorm.weight": "model-00032-of-00050.safetensors", + "model.layers.87.mlp.down_proj.weight": "model-00032-of-00050.safetensors", + "model.layers.87.mlp.gate_proj.weight": "model-00032-of-00050.safetensors", + "model.layers.87.mlp.up_proj.weight": "model-00032-of-00050.safetensors", + "model.layers.87.post_attention_layernorm.weight": "model-00032-of-00050.safetensors", + "model.layers.87.self_attn.k_proj.weight": "model-00032-of-00050.safetensors", + "model.layers.87.self_attn.o_proj.weight": "model-00032-of-00050.safetensors", + "model.layers.87.self_attn.q_proj.weight": "model-00032-of-00050.safetensors", + "model.layers.87.self_attn.v_proj.weight": "model-00032-of-00050.safetensors", + "model.layers.88.input_layernorm.weight": "model-00032-of-00050.safetensors", + "model.layers.88.mlp.down_proj.weight": "model-00032-of-00050.safetensors", + "model.layers.88.mlp.gate_proj.weight": "model-00032-of-00050.safetensors", + "model.layers.88.mlp.up_proj.weight": "model-00032-of-00050.safetensors", + "model.layers.88.post_attention_layernorm.weight": "model-00032-of-00050.safetensors", + "model.layers.88.self_attn.k_proj.weight": "model-00032-of-00050.safetensors", + "model.layers.88.self_attn.o_proj.weight": "model-00032-of-00050.safetensors", + "model.layers.88.self_attn.q_proj.weight": "model-00032-of-00050.safetensors", + "model.layers.88.self_attn.v_proj.weight": "model-00032-of-00050.safetensors", + "model.layers.89.input_layernorm.weight": "model-00033-of-00050.safetensors", + "model.layers.89.mlp.down_proj.weight": "model-00033-of-00050.safetensors", + "model.layers.89.mlp.gate_proj.weight": "model-00033-of-00050.safetensors", + "model.layers.89.mlp.up_proj.weight": "model-00033-of-00050.safetensors", + "model.layers.89.post_attention_layernorm.weight": "model-00033-of-00050.safetensors", + "model.layers.89.self_attn.k_proj.weight": "model-00032-of-00050.safetensors", + "model.layers.89.self_attn.o_proj.weight": "model-00032-of-00050.safetensors", + "model.layers.89.self_attn.q_proj.weight": "model-00032-of-00050.safetensors", + "model.layers.89.self_attn.v_proj.weight": "model-00032-of-00050.safetensors", + "model.layers.9.input_layernorm.weight": "model-00004-of-00050.safetensors", + "model.layers.9.mlp.down_proj.weight": "model-00004-of-00050.safetensors", + "model.layers.9.mlp.gate_proj.weight": "model-00004-of-00050.safetensors", + "model.layers.9.mlp.up_proj.weight": "model-00004-of-00050.safetensors", + "model.layers.9.post_attention_layernorm.weight": "model-00004-of-00050.safetensors", + "model.layers.9.self_attn.k_proj.weight": "model-00004-of-00050.safetensors", + "model.layers.9.self_attn.o_proj.weight": "model-00004-of-00050.safetensors", + "model.layers.9.self_attn.q_proj.weight": "model-00004-of-00050.safetensors", + "model.layers.9.self_attn.v_proj.weight": "model-00004-of-00050.safetensors", + "model.layers.90.input_layernorm.weight": "model-00033-of-00050.safetensors", + "model.layers.90.mlp.down_proj.weight": "model-00033-of-00050.safetensors", + "model.layers.90.mlp.gate_proj.weight": "model-00033-of-00050.safetensors", + "model.layers.90.mlp.up_proj.weight": "model-00033-of-00050.safetensors", + "model.layers.90.post_attention_layernorm.weight": "model-00033-of-00050.safetensors", + "model.layers.90.self_attn.k_proj.weight": "model-00033-of-00050.safetensors", + "model.layers.90.self_attn.o_proj.weight": "model-00033-of-00050.safetensors", + "model.layers.90.self_attn.q_proj.weight": "model-00033-of-00050.safetensors", + "model.layers.90.self_attn.v_proj.weight": "model-00033-of-00050.safetensors", + "model.layers.91.input_layernorm.weight": "model-00033-of-00050.safetensors", + "model.layers.91.mlp.down_proj.weight": "model-00033-of-00050.safetensors", + "model.layers.91.mlp.gate_proj.weight": "model-00033-of-00050.safetensors", + "model.layers.91.mlp.up_proj.weight": "model-00033-of-00050.safetensors", + "model.layers.91.post_attention_layernorm.weight": "model-00033-of-00050.safetensors", + "model.layers.91.self_attn.k_proj.weight": "model-00033-of-00050.safetensors", + "model.layers.91.self_attn.o_proj.weight": "model-00033-of-00050.safetensors", + "model.layers.91.self_attn.q_proj.weight": "model-00033-of-00050.safetensors", + "model.layers.91.self_attn.v_proj.weight": "model-00033-of-00050.safetensors", + "model.layers.92.input_layernorm.weight": "model-00034-of-00050.safetensors", + "model.layers.92.mlp.down_proj.weight": "model-00034-of-00050.safetensors", + "model.layers.92.mlp.gate_proj.weight": "model-00034-of-00050.safetensors", + "model.layers.92.mlp.up_proj.weight": "model-00034-of-00050.safetensors", + "model.layers.92.post_attention_layernorm.weight": "model-00034-of-00050.safetensors", + "model.layers.92.self_attn.k_proj.weight": "model-00033-of-00050.safetensors", + "model.layers.92.self_attn.o_proj.weight": "model-00034-of-00050.safetensors", + "model.layers.92.self_attn.q_proj.weight": "model-00033-of-00050.safetensors", + "model.layers.92.self_attn.v_proj.weight": "model-00033-of-00050.safetensors", + "model.layers.93.input_layernorm.weight": "model-00034-of-00050.safetensors", + "model.layers.93.mlp.down_proj.weight": "model-00034-of-00050.safetensors", + "model.layers.93.mlp.gate_proj.weight": "model-00034-of-00050.safetensors", + "model.layers.93.mlp.up_proj.weight": "model-00034-of-00050.safetensors", + "model.layers.93.post_attention_layernorm.weight": "model-00034-of-00050.safetensors", + "model.layers.93.self_attn.k_proj.weight": "model-00034-of-00050.safetensors", + "model.layers.93.self_attn.o_proj.weight": "model-00034-of-00050.safetensors", + "model.layers.93.self_attn.q_proj.weight": "model-00034-of-00050.safetensors", + "model.layers.93.self_attn.v_proj.weight": "model-00034-of-00050.safetensors", + "model.layers.94.input_layernorm.weight": "model-00034-of-00050.safetensors", + "model.layers.94.mlp.down_proj.weight": "model-00034-of-00050.safetensors", + "model.layers.94.mlp.gate_proj.weight": "model-00034-of-00050.safetensors", + "model.layers.94.mlp.up_proj.weight": "model-00034-of-00050.safetensors", + "model.layers.94.post_attention_layernorm.weight": "model-00034-of-00050.safetensors", + "model.layers.94.self_attn.k_proj.weight": "model-00034-of-00050.safetensors", + "model.layers.94.self_attn.o_proj.weight": "model-00034-of-00050.safetensors", + "model.layers.94.self_attn.q_proj.weight": "model-00034-of-00050.safetensors", + "model.layers.94.self_attn.v_proj.weight": "model-00034-of-00050.safetensors", + "model.layers.95.input_layernorm.weight": "model-00035-of-00050.safetensors", + "model.layers.95.mlp.down_proj.weight": "model-00035-of-00050.safetensors", + "model.layers.95.mlp.gate_proj.weight": "model-00035-of-00050.safetensors", + "model.layers.95.mlp.up_proj.weight": "model-00035-of-00050.safetensors", + "model.layers.95.post_attention_layernorm.weight": "model-00035-of-00050.safetensors", + "model.layers.95.self_attn.k_proj.weight": "model-00035-of-00050.safetensors", + "model.layers.95.self_attn.o_proj.weight": "model-00035-of-00050.safetensors", + "model.layers.95.self_attn.q_proj.weight": "model-00035-of-00050.safetensors", + "model.layers.95.self_attn.v_proj.weight": "model-00035-of-00050.safetensors", + "model.layers.96.input_layernorm.weight": "model-00035-of-00050.safetensors", + "model.layers.96.mlp.down_proj.weight": "model-00035-of-00050.safetensors", + "model.layers.96.mlp.gate_proj.weight": "model-00035-of-00050.safetensors", + "model.layers.96.mlp.up_proj.weight": "model-00035-of-00050.safetensors", + "model.layers.96.post_attention_layernorm.weight": "model-00035-of-00050.safetensors", + "model.layers.96.self_attn.k_proj.weight": "model-00035-of-00050.safetensors", + "model.layers.96.self_attn.o_proj.weight": "model-00035-of-00050.safetensors", + "model.layers.96.self_attn.q_proj.weight": "model-00035-of-00050.safetensors", + "model.layers.96.self_attn.v_proj.weight": "model-00035-of-00050.safetensors", + "model.layers.97.input_layernorm.weight": "model-00036-of-00050.safetensors", + "model.layers.97.mlp.down_proj.weight": "model-00036-of-00050.safetensors", + "model.layers.97.mlp.gate_proj.weight": "model-00035-of-00050.safetensors", + "model.layers.97.mlp.up_proj.weight": "model-00035-of-00050.safetensors", + "model.layers.97.post_attention_layernorm.weight": "model-00036-of-00050.safetensors", + "model.layers.97.self_attn.k_proj.weight": "model-00035-of-00050.safetensors", + "model.layers.97.self_attn.o_proj.weight": "model-00035-of-00050.safetensors", + "model.layers.97.self_attn.q_proj.weight": "model-00035-of-00050.safetensors", + "model.layers.97.self_attn.v_proj.weight": "model-00035-of-00050.safetensors", + "model.layers.98.input_layernorm.weight": "model-00036-of-00050.safetensors", + "model.layers.98.mlp.down_proj.weight": "model-00036-of-00050.safetensors", + "model.layers.98.mlp.gate_proj.weight": "model-00036-of-00050.safetensors", + "model.layers.98.mlp.up_proj.weight": "model-00036-of-00050.safetensors", + "model.layers.98.post_attention_layernorm.weight": "model-00036-of-00050.safetensors", + "model.layers.98.self_attn.k_proj.weight": "model-00036-of-00050.safetensors", + "model.layers.98.self_attn.o_proj.weight": "model-00036-of-00050.safetensors", + "model.layers.98.self_attn.q_proj.weight": "model-00036-of-00050.safetensors", + "model.layers.98.self_attn.v_proj.weight": "model-00036-of-00050.safetensors", + "model.layers.99.input_layernorm.weight": "model-00036-of-00050.safetensors", + "model.layers.99.mlp.down_proj.weight": "model-00036-of-00050.safetensors", + "model.layers.99.mlp.gate_proj.weight": "model-00036-of-00050.safetensors", + "model.layers.99.mlp.up_proj.weight": "model-00036-of-00050.safetensors", + "model.layers.99.post_attention_layernorm.weight": "model-00036-of-00050.safetensors", + "model.layers.99.self_attn.k_proj.weight": "model-00036-of-00050.safetensors", + "model.layers.99.self_attn.o_proj.weight": "model-00036-of-00050.safetensors", + "model.layers.99.self_attn.q_proj.weight": "model-00036-of-00050.safetensors", + "model.layers.99.self_attn.v_proj.weight": "model-00036-of-00050.safetensors", + "model.norm.weight": "model-00049-of-00050.safetensors" + } +} \ No newline at end of file