diff --git a/README.md b/README.md new file mode 100644 index 0000000000000000000000000000000000000000..bc5f30d6632ac0efdc7be2e9095e9e9579af2e33 --- /dev/null +++ b/README.md @@ -0,0 +1,199 @@ +--- +library_name: transformers +tags: [] +--- + +# Model Card for Model ID + + + + + +## Model Details + +### Model Description + + + +This is the model card of a 🤗 transformers model that has been pushed on the Hub. This model card has been automatically generated. + +- **Developed by:** [More Information Needed] +- **Funded by [optional]:** [More Information Needed] +- **Shared by [optional]:** [More Information Needed] +- **Model type:** [More Information Needed] +- **Language(s) (NLP):** [More Information Needed] +- **License:** [More Information Needed] +- **Finetuned from model [optional]:** [More Information Needed] + +### Model Sources [optional] + + + +- **Repository:** [More Information Needed] +- **Paper [optional]:** [More Information Needed] +- **Demo [optional]:** [More Information Needed] + +## Uses + + + +### Direct Use + + + +[More Information Needed] + +### Downstream Use [optional] + + + +[More Information Needed] + +### Out-of-Scope Use + + + +[More Information Needed] + +## Bias, Risks, and Limitations + + + +[More Information Needed] + +### Recommendations + + + +Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. + +## How to Get Started with the Model + +Use the code below to get started with the model. + +[More Information Needed] + +## Training Details + +### Training Data + + + +[More Information Needed] + +### Training Procedure + + + +#### Preprocessing [optional] + +[More Information Needed] + + +#### Training Hyperparameters + +- **Training regime:** [More Information Needed] + +#### Speeds, Sizes, Times [optional] + + + +[More Information Needed] + +## Evaluation + + + +### Testing Data, Factors & Metrics + +#### Testing Data + + + +[More Information Needed] + +#### Factors + + + +[More Information Needed] + +#### Metrics + + + +[More Information Needed] + +### Results + +[More Information Needed] + +#### Summary + + + +## Model Examination [optional] + + + +[More Information Needed] + +## Environmental Impact + + + +Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700). + +- **Hardware Type:** [More Information Needed] +- **Hours used:** [More Information Needed] +- **Cloud Provider:** [More Information Needed] +- **Compute Region:** [More Information Needed] +- **Carbon Emitted:** [More Information Needed] + +## Technical Specifications [optional] + +### Model Architecture and Objective + +[More Information Needed] + +### Compute Infrastructure + +[More Information Needed] + +#### Hardware + +[More Information Needed] + +#### Software + +[More Information Needed] + +## Citation [optional] + + + +**BibTeX:** + +[More Information Needed] + +**APA:** + +[More Information Needed] + +## Glossary [optional] + + + +[More Information Needed] + +## More Information [optional] + +[More Information Needed] + +## Model Card Authors [optional] + +[More Information Needed] + +## Model Card Contact + +[More Information Needed] \ No newline at end of file diff --git a/config.json b/config.json new file mode 100644 index 0000000000000000000000000000000000000000..16c1b8e9ea85d8b5da8832024c3f433d6e24abfb --- /dev/null +++ b/config.json @@ -0,0 +1,28 @@ +{ + "_name_or_path": "./merged_model", + "architectures": [ + "Qwen2ForCausalLM" + ], + "attention_dropout": 0.0, + "eos_token_id": 151645, + "hidden_act": "silu", + "hidden_size": 8192, + "initializer_range": 0.02, + "intermediate_size": 29568, + "max_position_embeddings": 32768, + "max_window_layers": 80, + "model_type": "qwen2", + "num_attention_heads": 64, + "num_hidden_layers": 80, + "num_key_value_heads": 8, + "rms_norm_eps": 1e-06, + "rope_scaling": null, + "rope_theta": 1000000.0, + "sliding_window": null, + "tie_word_embeddings": false, + "torch_dtype": "float32", + "transformers_version": "4.48.1", + "use_cache": false, + "use_sliding_window": false, + "vocab_size": 151646 +} diff --git a/generation_config.json b/generation_config.json new file mode 100644 index 0000000000000000000000000000000000000000..e76d8ffc31bd7eadb4d4f9d40737f04ee1741269 --- /dev/null +++ b/generation_config.json @@ -0,0 +1,6 @@ +{ + "_from_model_config": true, + "eos_token_id": 151645, + "transformers_version": "4.48.1", + "use_cache": false +} diff --git a/model-00001-of-00063.safetensors b/model-00001-of-00063.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..29a1a9a7932265efa004239c37586f0c82abce05 --- /dev/null +++ b/model-00001-of-00063.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:52c8739fe97a288699a6a64961d28a9ae6a2c68f15a8afca9cec0b21e19e5cdf +size 4969136264 diff --git a/model-00002-of-00063.safetensors b/model-00002-of-00063.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..86eaa030cc99764f1b60fdf87f434d2c168fe1c1 --- /dev/null +++ b/model-00002-of-00063.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:d70752295a337b66db3bc0f724039ea332b665975df0793a6f754211f3b29851 +size 4114761856 diff --git a/model-00003-of-00063.safetensors b/model-00003-of-00063.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..c283b4d53f1488759b2994e355bfb12def9ce939 --- /dev/null +++ b/model-00003-of-00063.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:49bffedf7544315b5aa137c225f4af3072cd8d1323ed1069853907c9e595ad70 +size 4479624664 diff --git a/model-00004-of-00063.safetensors b/model-00004-of-00063.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..d2157fb095cf1acaec4189be62cdfcfbd2cdb0af --- /dev/null +++ b/model-00004-of-00063.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:6ed518baa24523fae234f5568eb51b468fc442a369620004fbdd5d309bf4cf00 +size 4479624664 diff --git a/model-00005-of-00063.safetensors b/model-00005-of-00063.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..13be16bb608c6cf56599a59be1cc21a33af08c73 --- /dev/null +++ b/model-00005-of-00063.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:05eda129d8de289390e388454b0f1e09e1a68377a69df6f9394272c9c846c6a9 +size 4815276384 diff --git a/model-00006-of-00063.safetensors b/model-00006-of-00063.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..6e642e35a8f16b0d4e2dcc907f874fef3131d8df --- /dev/null +++ b/model-00006-of-00063.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:5b70d52edfc2a05b64617cbe3fce554c95203dcc119fc143fd7a7c39955ed4db +size 4748060240 diff --git a/model-00007-of-00063.safetensors b/model-00007-of-00063.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..2e94a7d57f91f1978d09a724d05a929e3fe738b2 --- /dev/null +++ b/model-00007-of-00063.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:affd680c99665d3ddb9d40402026b59c38d00840cad8c5e3c4926b4b18f04883 +size 4479624664 diff --git a/model-00008-of-00063.safetensors b/model-00008-of-00063.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..d78e06e845e38ef89feb514f44be5936e7dcd83c --- /dev/null +++ b/model-00008-of-00063.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:2f1310b74415936f22a735a5726b556c0d5614be16f95d79d51c8c1fc6b8a98f +size 4815276384 diff --git a/model-00009-of-00063.safetensors b/model-00009-of-00063.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..a7259e0cfecf282fcb5514240e6a406f20014c0e --- /dev/null +++ b/model-00009-of-00063.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:6fb64196ebaf54755a1f2da44a9135988377ce145ca255d8d1ff1ab9d8155da2 +size 4748060256 diff --git a/model-00010-of-00063.safetensors b/model-00010-of-00063.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..7b47ab6f7e1648d1a09ff9c711d42f7503e72f26 --- /dev/null +++ b/model-00010-of-00063.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:a8fa99dcaa65b3bec3764ac0d12a3ffb31a405a29126f2cd28718e75681d3e22 +size 4479624672 diff --git a/model-00011-of-00063.safetensors b/model-00011-of-00063.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..2e4496ed0c32b82ee82b13d564e7d6688323f71d --- /dev/null +++ b/model-00011-of-00063.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:722cf2e27f9dcbdd7df42091c97bb66219941f3fc20daba9058d817f765f0fdf +size 4815276400 diff --git a/model-00012-of-00063.safetensors b/model-00012-of-00063.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..2f18e60620f49f90cd068eca0ef336733c486da8 --- /dev/null +++ b/model-00012-of-00063.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:0db64777cff98190067c29fa85b1ac8713e5c99618b54d40e3cf9777e1776aab +size 4748060248 diff --git a/model-00013-of-00063.safetensors b/model-00013-of-00063.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..6d6cec531180aa80f70da97c28f4d66d8672c363 --- /dev/null +++ b/model-00013-of-00063.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:96fb5a96646e04e9b0c49931044d9441f7d6e4274e387cfe40a30494dc11119e +size 4479624672 diff --git a/model-00014-of-00063.safetensors b/model-00014-of-00063.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..e4e92bbf6d4878fa221579a106d03bc14f413234 --- /dev/null +++ b/model-00014-of-00063.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:5743389d2b76d692a682876e458ab27b57f29a719b11ce2deb92867ffa57f76f +size 4815276400 diff --git a/model-00015-of-00063.safetensors b/model-00015-of-00063.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..bdd41b86303fa41016bde63670884298be676c82 --- /dev/null +++ b/model-00015-of-00063.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:4347f6f10707b6862adb14f34503be58fe985928dec6ca99bca8eec7a82004c7 +size 4748060248 diff --git a/model-00016-of-00063.safetensors b/model-00016-of-00063.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..a18c0e5f0784017d9953a8760234245b3c804420 --- /dev/null +++ b/model-00016-of-00063.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:13b85d1b419e87c15dc4252e84acbbcca03a759b4b88715c5d99b60298536dde +size 4479624672 diff --git a/model-00017-of-00063.safetensors b/model-00017-of-00063.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..1f8bdb58a10eb4948b55a8634079a99b089cfc6e --- /dev/null +++ b/model-00017-of-00063.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:71d1fef57c1c80405ee12c81a630d0293e8c10f0cbc2baae69d01885c75ef0f6 +size 4815276400 diff --git a/model-00018-of-00063.safetensors b/model-00018-of-00063.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..e3769eb18a162cf4e1465a3658a8ffea33243dd7 --- /dev/null +++ b/model-00018-of-00063.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:b0283c98c2c774e68b4c36da5e879c3d4b2169486a556e074ddd64f03435ca69 +size 4748060248 diff --git a/model-00019-of-00063.safetensors b/model-00019-of-00063.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..62574b6214075d9d4f8ea8325f2854a931f982ec --- /dev/null +++ b/model-00019-of-00063.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:65fc043ca7aa1e360c044cfb4f6e735caba5c018b60e0b35ac9da72dd7f38b09 +size 4479624672 diff --git a/model-00020-of-00063.safetensors b/model-00020-of-00063.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..0ed06b036ac90dcf40f9a86f3cedbc8daf995be6 --- /dev/null +++ b/model-00020-of-00063.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:881a62be425f35c5fc1606faf07418862b5784dcb0e367249d8f1a6bc54497bc +size 4815276400 diff --git a/model-00021-of-00063.safetensors b/model-00021-of-00063.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..65640deca6fd120ae170c3aaab5e13ac9a8b313f --- /dev/null +++ b/model-00021-of-00063.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:9ae9b863fba1a3cdc7d271838b3619eacd87f507356b484ef8bf33854ff098b4 +size 4748060248 diff --git a/model-00022-of-00063.safetensors b/model-00022-of-00063.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..c5bf88e544514a37689244ac9e0f29f141363440 --- /dev/null +++ b/model-00022-of-00063.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:47bddb24571a368a9984a2aaa94d02c9b9e4c66740cdbe72df7e36317fcd94e3 +size 4479624672 diff --git a/model-00023-of-00063.safetensors b/model-00023-of-00063.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..69c78585d1ec28bca78c499a12c9e4930a7108de --- /dev/null +++ b/model-00023-of-00063.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:1681fabfaa1d51b6e3a8b2ee50759246fadf5ce11d03c15350eab285d62fc70f +size 4815276400 diff --git a/model-00024-of-00063.safetensors b/model-00024-of-00063.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..487d4f9c63ce634fc5934f920004d479ba7fb134 --- /dev/null +++ b/model-00024-of-00063.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:205a123ce16d092a1a7c807ec78c82943bae755cadf19ffaea6acf4ea871187e +size 4748060248 diff --git a/model-00025-of-00063.safetensors b/model-00025-of-00063.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..f652718fe682883852f825c7678fbdd887ae0267 --- /dev/null +++ b/model-00025-of-00063.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:6d0c47fe934e21977aefb1b175b3142a55ce735394ef6eb4be5deac18517091f +size 4479624672 diff --git a/model-00026-of-00063.safetensors b/model-00026-of-00063.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..a143a41c3e8c7a31a090a849c7be04eb0ff368ea --- /dev/null +++ b/model-00026-of-00063.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:371e1e20817efdc5acb381676889a1fd64124b2fd4bbd6734ad3a13e8afa3cb4 +size 4815276400 diff --git a/model-00027-of-00063.safetensors b/model-00027-of-00063.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..85bb0210ecf8d43caa77b2e1404e539f4e741d2e --- /dev/null +++ b/model-00027-of-00063.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:3e54a76d30725c92937936890cc3e2a452d379ed9c1351365cefbf9174ca4823 +size 4748060248 diff --git a/model-00028-of-00063.safetensors b/model-00028-of-00063.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..4a0184b6240e79fe0bc77b5b05eded5be5f7e792 --- /dev/null +++ b/model-00028-of-00063.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:ff5d5095d5240f857ad870eab75337500e3141cdf6211d37ae7210d706aa0fe7 +size 4479624672 diff --git a/model-00029-of-00063.safetensors b/model-00029-of-00063.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..e7257c910ed9780b7bb18b14123470460a3f0367 --- /dev/null +++ b/model-00029-of-00063.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:d7db3d59a05bf2611c4d68713edf67609c523bc9abc54449952fe2aed741459c +size 4815276400 diff --git a/model-00030-of-00063.safetensors b/model-00030-of-00063.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..9de7b99103df4e9740d3bd122e04d71d302a75c2 --- /dev/null +++ b/model-00030-of-00063.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:571de628f0637a5058fd155d4018384498e6b2b02a0f0024a6025c1dab68cf1a +size 4748060248 diff --git a/model-00031-of-00063.safetensors b/model-00031-of-00063.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..c8fbc4285bfba5532aff900a0e43fe9e0eedb0d9 --- /dev/null +++ b/model-00031-of-00063.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:077c53926e2ba27cbe4ce3dcd77b79eca7f572bcfad75b0fa9d56283bb9fa312 +size 4479624672 diff --git a/model-00032-of-00063.safetensors b/model-00032-of-00063.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..f9b8a827f5bd9385c00de6d06f1d3c4b8f50f588 --- /dev/null +++ b/model-00032-of-00063.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:8d403701fdab2ca267faf8e00cfbbb9586bb763f0bf436e25bf8348a779bf47a +size 4815276400 diff --git a/model-00033-of-00063.safetensors b/model-00033-of-00063.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..349d5a60fc7b2215469fbdcec28dbbd024f4623e --- /dev/null +++ b/model-00033-of-00063.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:76ecbc7c6900387173a5bf43da7b0a2d303f96455e5b8b0806debe1ee9694b37 +size 4748060248 diff --git a/model-00034-of-00063.safetensors b/model-00034-of-00063.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..aa49e6e144d9883973dd5db392c2077495020355 --- /dev/null +++ b/model-00034-of-00063.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:f7af0c30b57fbe373b7f81dc121277e890187b4a414189d8af51df11b689c1e1 +size 4479624672 diff --git a/model-00035-of-00063.safetensors b/model-00035-of-00063.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..7f6257064fa631e0c0b4838353f9b217b4699d94 --- /dev/null +++ b/model-00035-of-00063.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:d009931d72b2ab71ff7c6f8123df64b564f191a8dc49ed13c843334d78528681 +size 4815276400 diff --git a/model-00036-of-00063.safetensors b/model-00036-of-00063.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..e44382c65be3a17765bfe6ba378177f937e391c1 --- /dev/null +++ b/model-00036-of-00063.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:0190755a24bbbd220dbebccdafd81029cf621ec8828b58a83757d9efde6e4170 +size 4748060248 diff --git a/model-00037-of-00063.safetensors b/model-00037-of-00063.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..23abedc25e0ca04006ba24a8eec13dfe97b052c7 --- /dev/null +++ b/model-00037-of-00063.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:6e830d2cd61f1c935cc458a51e8909f8466f339a3e45fd2a3eca4e87255fc598 +size 4479624672 diff --git a/model-00038-of-00063.safetensors b/model-00038-of-00063.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..366b193cd65e4b2668cc6b7ba9545e8f412e7289 --- /dev/null +++ b/model-00038-of-00063.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:64bfd0e7531241e30d79a16c45fc9771cdab290489bcb5520a77946bf37924b7 +size 4815276400 diff --git a/model-00039-of-00063.safetensors b/model-00039-of-00063.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..3d7ecea9519d33da8739e43532aa2ebd436d6529 --- /dev/null +++ b/model-00039-of-00063.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:7abb85db14265a845614ba251408c64cbeddaea7e0147c92b739b01b28143a45 +size 4748060248 diff --git a/model-00040-of-00063.safetensors b/model-00040-of-00063.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..81f75f204bfe9f5786385ba664175cf08d2df4ee --- /dev/null +++ b/model-00040-of-00063.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:45e3a7c1b818c6e427e7a7a1eb43e8e00da88f737d9d4bb584b3599e591f246f +size 4479624672 diff --git a/model-00041-of-00063.safetensors b/model-00041-of-00063.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..82fc5a7f21b891c69f3eb6e83f29dc3766a40ea0 --- /dev/null +++ b/model-00041-of-00063.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:1dc05ff9b20f42cec11e479e10fae338810827d060c56b9d676693e099c1a440 +size 4815276400 diff --git a/model-00042-of-00063.safetensors b/model-00042-of-00063.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..c7921d117fcd7c977ea3696844cb3a9fb7c8d6bb --- /dev/null +++ b/model-00042-of-00063.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:b023c20e6590ae9693ffcec734594c3b57bc903a4f138cadad09e8eed0ab9b00 +size 4748060248 diff --git a/model-00043-of-00063.safetensors b/model-00043-of-00063.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..c0bad8053737b1fb60c039a4e95c6a9bd833e3d3 --- /dev/null +++ b/model-00043-of-00063.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:269905b60793af1fcbb400e61ab4334824a6136c84374bc90fb1adfdc41304b5 +size 4479624672 diff --git a/model-00044-of-00063.safetensors b/model-00044-of-00063.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..c8f0bbd1e09015f49f998830a1ba4a5c4af063ca --- /dev/null +++ b/model-00044-of-00063.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:a37051ac3c5a82d7eb4777c89240d66c51f783ee8fe5542993fba96431118b00 +size 4815276400 diff --git a/model-00045-of-00063.safetensors b/model-00045-of-00063.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..49024638525a6c698fc8894fb1c898a00282d71f --- /dev/null +++ b/model-00045-of-00063.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:7455d803a6e3f684f7136c390930c78a7c72cc15578e85413093c2530cb9072f +size 4748060248 diff --git a/model-00046-of-00063.safetensors b/model-00046-of-00063.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..b9252299f77f848a945e9bd018b10609a706d79b --- /dev/null +++ b/model-00046-of-00063.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:8fa2d0c994a42ea6c91eaa5d96eab33e1d15280201d98a27416b3825acea5dbe +size 4479624672 diff --git a/model-00047-of-00063.safetensors b/model-00047-of-00063.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..1cbbc1d2c87e12b200b0e7784fa39af2f53b2ebc --- /dev/null +++ b/model-00047-of-00063.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:1d450278fc59e71d950e495fb009d7ac551a9bb9bad8671f858c62f3f3f63471 +size 4815276400 diff --git a/model-00048-of-00063.safetensors b/model-00048-of-00063.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..aebb481a181e01fc23e343c5968d395e1b6acc61 --- /dev/null +++ b/model-00048-of-00063.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:aefb26c4e5df95b56302026e73434fa1fae37514945b37959b3adfbcba060428 +size 4748060248 diff --git a/model-00049-of-00063.safetensors b/model-00049-of-00063.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..1bd171241c70714423185d7514f476bab93827fd --- /dev/null +++ b/model-00049-of-00063.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:e7564eec02caf9acbd7a1dd73e527c9f357e8d60fd38b19081bbb468c8a466e7 +size 4479624672 diff --git a/model-00050-of-00063.safetensors b/model-00050-of-00063.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..727b22159140acc33059c53da50fa361205d0436 --- /dev/null +++ b/model-00050-of-00063.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:e4d4720b0ccba5b21c3afdc817f1c8375ae7829102c1afcd7cd79f1d7f168c29 +size 4815276400 diff --git a/model-00051-of-00063.safetensors b/model-00051-of-00063.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..82f6d585e683ef3029da9bb28fdf9b7348dc61c8 --- /dev/null +++ b/model-00051-of-00063.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:33a6edf2b9b39cba4e51ea795250bd7240af9337c692071628a3628e6336681a +size 4748060248 diff --git a/model-00052-of-00063.safetensors b/model-00052-of-00063.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..01f7181cf94234bf09e129602232ba6043b76d3e --- /dev/null +++ b/model-00052-of-00063.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:f426da2ac246ee7326a1aa368f236ee6692844291543e96efef35f27daf679bd +size 4479624672 diff --git a/model-00053-of-00063.safetensors b/model-00053-of-00063.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..48f5c9b9beeb80d763f939119618e2fb4527ba12 --- /dev/null +++ b/model-00053-of-00063.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:5f56cd3409620f243aad55bb6f3c8e45c7a9d39fcd6a7eae91f73ba22d5a530b +size 4815276400 diff --git a/model-00054-of-00063.safetensors b/model-00054-of-00063.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..8208790aae2e7c7df7e4c5e229ddfd8a7cbd6f74 --- /dev/null +++ b/model-00054-of-00063.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:e35e6a8907be5f49490647100ece2aeea58b17c5ab423694eeb98feff34f7888 +size 4748060248 diff --git a/model-00055-of-00063.safetensors b/model-00055-of-00063.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..9e82a491e4421aed80151bc277e0494e81bb085f --- /dev/null +++ b/model-00055-of-00063.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:ce98268baf91e33f4b483084b1213de31f2140f134e9478a2cbec477cc4ddafd +size 4479624672 diff --git a/model-00056-of-00063.safetensors b/model-00056-of-00063.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..9e39a1d3b169a7f35faa2132756f669bf9ecb47d --- /dev/null +++ b/model-00056-of-00063.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:d19ea98deb034193e8080f98fa753f158b42b856fb4697b8e7dfa52944dd2123 +size 4815276400 diff --git a/model-00057-of-00063.safetensors b/model-00057-of-00063.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..b9b66ab8b71cf3ca63bab3570b992153453943d9 --- /dev/null +++ b/model-00057-of-00063.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:19ed8f408d1d3620b297d08a166dfba68fcb524ba354a604896d954fc9681486 +size 4748060248 diff --git a/model-00058-of-00063.safetensors b/model-00058-of-00063.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..67342acab06354a4251b795b56e29d1df6b99492 --- /dev/null +++ b/model-00058-of-00063.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:e9046d162ef2b214b1673b2754239f01dc21c9f18b700040b7a6ae9bf7e06dbb +size 4479624672 diff --git a/model-00059-of-00063.safetensors b/model-00059-of-00063.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..b174a86b1149bf3aad13f332d258a1144bd49748 --- /dev/null +++ b/model-00059-of-00063.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:23f23db33e115d0b212055979693464ead7cc174caf1dfd8515cb640df067737 +size 4815276400 diff --git a/model-00060-of-00063.safetensors b/model-00060-of-00063.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..7411710f070c3b353869fa474fb2228fd10d6de3 --- /dev/null +++ b/model-00060-of-00063.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:c9d3d9f606a0b7d5ac1c4e2097ea188e6332008bc84f8337c318f9875e4f343e +size 4748060248 diff --git a/model-00061-of-00063.safetensors b/model-00061-of-00063.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..f464302b39da40b51522fbc8d8ffead379d8055a --- /dev/null +++ b/model-00061-of-00063.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:046a7f8b202bee5991ce1e38259427867ddccfd2b1f39d42eecf30591eb5cbbb +size 4479624672 diff --git a/model-00062-of-00063.safetensors b/model-00062-of-00063.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..97ea511907687542a32f955de830dfc0dd9a592a --- /dev/null +++ b/model-00062-of-00063.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:fb31fc81ed028d88f14694c69437ab830b53f27e2c43118a4da935e01a5f1bbe +size 968982984 diff --git a/model-00063-of-00063.safetensors b/model-00063-of-00063.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..37e2527b673a32eb975d68d586dcee841a56260e --- /dev/null +++ b/model-00063-of-00063.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:867d406c5562ee2416da392f102b1ec5cd3bf728d90c95774eb690cb0806a33f +size 4969136256 diff --git a/model.safetensors.index.json b/model.safetensors.index.json new file mode 100644 index 0000000000000000000000000000000000000000..dc1176ab53f9ed87802ea2c7f001b1246b414ae4 --- /dev/null +++ b/model.safetensors.index.json @@ -0,0 +1,970 @@ +{ + "metadata": { + "total_size": 290797420544 + }, + "weight_map": { + "lm_head.weight": "model-00063-of-00063.safetensors", + "model.embed_tokens.weight": "model-00001-of-00063.safetensors", + "model.layers.0.input_layernorm.weight": "model-00002-of-00063.safetensors", + "model.layers.0.mlp.down_proj.weight": "model-00002-of-00063.safetensors", + "model.layers.0.mlp.gate_proj.weight": "model-00002-of-00063.safetensors", + "model.layers.0.mlp.up_proj.weight": "model-00002-of-00063.safetensors", + "model.layers.0.post_attention_layernorm.weight": "model-00002-of-00063.safetensors", + "model.layers.0.self_attn.k_proj.bias": "model-00002-of-00063.safetensors", + "model.layers.0.self_attn.k_proj.weight": "model-00002-of-00063.safetensors", + "model.layers.0.self_attn.o_proj.weight": "model-00002-of-00063.safetensors", + "model.layers.0.self_attn.q_proj.bias": "model-00002-of-00063.safetensors", + "model.layers.0.self_attn.q_proj.weight": "model-00002-of-00063.safetensors", + "model.layers.0.self_attn.v_proj.bias": "model-00002-of-00063.safetensors", + "model.layers.0.self_attn.v_proj.weight": "model-00002-of-00063.safetensors", + "model.layers.1.input_layernorm.weight": "model-00003-of-00063.safetensors", + "model.layers.1.mlp.down_proj.weight": "model-00003-of-00063.safetensors", + "model.layers.1.mlp.gate_proj.weight": "model-00003-of-00063.safetensors", + "model.layers.1.mlp.up_proj.weight": "model-00003-of-00063.safetensors", + "model.layers.1.post_attention_layernorm.weight": "model-00003-of-00063.safetensors", + "model.layers.1.self_attn.k_proj.bias": "model-00002-of-00063.safetensors", + "model.layers.1.self_attn.k_proj.weight": "model-00002-of-00063.safetensors", + "model.layers.1.self_attn.o_proj.weight": "model-00002-of-00063.safetensors", + "model.layers.1.self_attn.q_proj.bias": "model-00002-of-00063.safetensors", + "model.layers.1.self_attn.q_proj.weight": "model-00002-of-00063.safetensors", + "model.layers.1.self_attn.v_proj.bias": "model-00002-of-00063.safetensors", + "model.layers.1.self_attn.v_proj.weight": "model-00002-of-00063.safetensors", + "model.layers.10.input_layernorm.weight": "model-00010-of-00063.safetensors", + "model.layers.10.mlp.down_proj.weight": "model-00010-of-00063.safetensors", + "model.layers.10.mlp.gate_proj.weight": "model-00009-of-00063.safetensors", + "model.layers.10.mlp.up_proj.weight": "model-00010-of-00063.safetensors", + "model.layers.10.post_attention_layernorm.weight": "model-00010-of-00063.safetensors", + "model.layers.10.self_attn.k_proj.bias": "model-00009-of-00063.safetensors", + "model.layers.10.self_attn.k_proj.weight": "model-00009-of-00063.safetensors", + "model.layers.10.self_attn.o_proj.weight": "model-00009-of-00063.safetensors", + "model.layers.10.self_attn.q_proj.bias": "model-00009-of-00063.safetensors", + "model.layers.10.self_attn.q_proj.weight": "model-00009-of-00063.safetensors", + "model.layers.10.self_attn.v_proj.bias": "model-00009-of-00063.safetensors", + "model.layers.10.self_attn.v_proj.weight": "model-00009-of-00063.safetensors", + "model.layers.11.input_layernorm.weight": "model-00011-of-00063.safetensors", + "model.layers.11.mlp.down_proj.weight": "model-00011-of-00063.safetensors", + "model.layers.11.mlp.gate_proj.weight": "model-00010-of-00063.safetensors", + "model.layers.11.mlp.up_proj.weight": "model-00010-of-00063.safetensors", + "model.layers.11.post_attention_layernorm.weight": "model-00011-of-00063.safetensors", + "model.layers.11.self_attn.k_proj.bias": "model-00010-of-00063.safetensors", + "model.layers.11.self_attn.k_proj.weight": "model-00010-of-00063.safetensors", + "model.layers.11.self_attn.o_proj.weight": "model-00010-of-00063.safetensors", + "model.layers.11.self_attn.q_proj.bias": "model-00010-of-00063.safetensors", + "model.layers.11.self_attn.q_proj.weight": "model-00010-of-00063.safetensors", + "model.layers.11.self_attn.v_proj.bias": "model-00010-of-00063.safetensors", + "model.layers.11.self_attn.v_proj.weight": "model-00010-of-00063.safetensors", + "model.layers.12.input_layernorm.weight": "model-00011-of-00063.safetensors", + "model.layers.12.mlp.down_proj.weight": "model-00011-of-00063.safetensors", + "model.layers.12.mlp.gate_proj.weight": "model-00011-of-00063.safetensors", + "model.layers.12.mlp.up_proj.weight": "model-00011-of-00063.safetensors", + "model.layers.12.post_attention_layernorm.weight": "model-00011-of-00063.safetensors", + "model.layers.12.self_attn.k_proj.bias": "model-00011-of-00063.safetensors", + "model.layers.12.self_attn.k_proj.weight": "model-00011-of-00063.safetensors", + "model.layers.12.self_attn.o_proj.weight": "model-00011-of-00063.safetensors", + "model.layers.12.self_attn.q_proj.bias": "model-00011-of-00063.safetensors", + "model.layers.12.self_attn.q_proj.weight": "model-00011-of-00063.safetensors", + "model.layers.12.self_attn.v_proj.bias": "model-00011-of-00063.safetensors", + "model.layers.12.self_attn.v_proj.weight": "model-00011-of-00063.safetensors", + "model.layers.13.input_layernorm.weight": "model-00012-of-00063.safetensors", + "model.layers.13.mlp.down_proj.weight": "model-00012-of-00063.safetensors", + "model.layers.13.mlp.gate_proj.weight": "model-00012-of-00063.safetensors", + "model.layers.13.mlp.up_proj.weight": "model-00012-of-00063.safetensors", + "model.layers.13.post_attention_layernorm.weight": "model-00012-of-00063.safetensors", + "model.layers.13.self_attn.k_proj.bias": "model-00011-of-00063.safetensors", + "model.layers.13.self_attn.k_proj.weight": "model-00011-of-00063.safetensors", + "model.layers.13.self_attn.o_proj.weight": "model-00012-of-00063.safetensors", + "model.layers.13.self_attn.q_proj.bias": "model-00011-of-00063.safetensors", + "model.layers.13.self_attn.q_proj.weight": "model-00011-of-00063.safetensors", + "model.layers.13.self_attn.v_proj.bias": "model-00011-of-00063.safetensors", + "model.layers.13.self_attn.v_proj.weight": "model-00011-of-00063.safetensors", + "model.layers.14.input_layernorm.weight": "model-00013-of-00063.safetensors", + "model.layers.14.mlp.down_proj.weight": "model-00013-of-00063.safetensors", + "model.layers.14.mlp.gate_proj.weight": "model-00012-of-00063.safetensors", + "model.layers.14.mlp.up_proj.weight": "model-00013-of-00063.safetensors", + "model.layers.14.post_attention_layernorm.weight": "model-00013-of-00063.safetensors", + "model.layers.14.self_attn.k_proj.bias": "model-00012-of-00063.safetensors", + "model.layers.14.self_attn.k_proj.weight": "model-00012-of-00063.safetensors", + "model.layers.14.self_attn.o_proj.weight": "model-00012-of-00063.safetensors", + "model.layers.14.self_attn.q_proj.bias": "model-00012-of-00063.safetensors", + "model.layers.14.self_attn.q_proj.weight": "model-00012-of-00063.safetensors", + "model.layers.14.self_attn.v_proj.bias": "model-00012-of-00063.safetensors", + "model.layers.14.self_attn.v_proj.weight": "model-00012-of-00063.safetensors", + "model.layers.15.input_layernorm.weight": "model-00014-of-00063.safetensors", + "model.layers.15.mlp.down_proj.weight": "model-00014-of-00063.safetensors", + "model.layers.15.mlp.gate_proj.weight": "model-00013-of-00063.safetensors", + "model.layers.15.mlp.up_proj.weight": "model-00013-of-00063.safetensors", + "model.layers.15.post_attention_layernorm.weight": "model-00014-of-00063.safetensors", + "model.layers.15.self_attn.k_proj.bias": "model-00013-of-00063.safetensors", + "model.layers.15.self_attn.k_proj.weight": "model-00013-of-00063.safetensors", + "model.layers.15.self_attn.o_proj.weight": "model-00013-of-00063.safetensors", + "model.layers.15.self_attn.q_proj.bias": "model-00013-of-00063.safetensors", + "model.layers.15.self_attn.q_proj.weight": "model-00013-of-00063.safetensors", + "model.layers.15.self_attn.v_proj.bias": "model-00013-of-00063.safetensors", + "model.layers.15.self_attn.v_proj.weight": "model-00013-of-00063.safetensors", + "model.layers.16.input_layernorm.weight": "model-00014-of-00063.safetensors", + "model.layers.16.mlp.down_proj.weight": "model-00014-of-00063.safetensors", + "model.layers.16.mlp.gate_proj.weight": "model-00014-of-00063.safetensors", + "model.layers.16.mlp.up_proj.weight": "model-00014-of-00063.safetensors", + "model.layers.16.post_attention_layernorm.weight": "model-00014-of-00063.safetensors", + "model.layers.16.self_attn.k_proj.bias": "model-00014-of-00063.safetensors", + "model.layers.16.self_attn.k_proj.weight": "model-00014-of-00063.safetensors", + "model.layers.16.self_attn.o_proj.weight": "model-00014-of-00063.safetensors", + "model.layers.16.self_attn.q_proj.bias": "model-00014-of-00063.safetensors", + "model.layers.16.self_attn.q_proj.weight": "model-00014-of-00063.safetensors", + "model.layers.16.self_attn.v_proj.bias": "model-00014-of-00063.safetensors", + "model.layers.16.self_attn.v_proj.weight": "model-00014-of-00063.safetensors", + "model.layers.17.input_layernorm.weight": "model-00015-of-00063.safetensors", + "model.layers.17.mlp.down_proj.weight": "model-00015-of-00063.safetensors", + "model.layers.17.mlp.gate_proj.weight": "model-00015-of-00063.safetensors", + "model.layers.17.mlp.up_proj.weight": "model-00015-of-00063.safetensors", + "model.layers.17.post_attention_layernorm.weight": "model-00015-of-00063.safetensors", + "model.layers.17.self_attn.k_proj.bias": "model-00014-of-00063.safetensors", + "model.layers.17.self_attn.k_proj.weight": "model-00014-of-00063.safetensors", + "model.layers.17.self_attn.o_proj.weight": "model-00015-of-00063.safetensors", + "model.layers.17.self_attn.q_proj.bias": "model-00014-of-00063.safetensors", + "model.layers.17.self_attn.q_proj.weight": "model-00014-of-00063.safetensors", + "model.layers.17.self_attn.v_proj.bias": "model-00014-of-00063.safetensors", + "model.layers.17.self_attn.v_proj.weight": "model-00014-of-00063.safetensors", + "model.layers.18.input_layernorm.weight": "model-00016-of-00063.safetensors", + "model.layers.18.mlp.down_proj.weight": "model-00016-of-00063.safetensors", + "model.layers.18.mlp.gate_proj.weight": "model-00015-of-00063.safetensors", + "model.layers.18.mlp.up_proj.weight": "model-00016-of-00063.safetensors", + "model.layers.18.post_attention_layernorm.weight": "model-00016-of-00063.safetensors", + "model.layers.18.self_attn.k_proj.bias": "model-00015-of-00063.safetensors", + "model.layers.18.self_attn.k_proj.weight": "model-00015-of-00063.safetensors", + "model.layers.18.self_attn.o_proj.weight": "model-00015-of-00063.safetensors", + "model.layers.18.self_attn.q_proj.bias": "model-00015-of-00063.safetensors", + "model.layers.18.self_attn.q_proj.weight": "model-00015-of-00063.safetensors", + "model.layers.18.self_attn.v_proj.bias": "model-00015-of-00063.safetensors", + "model.layers.18.self_attn.v_proj.weight": "model-00015-of-00063.safetensors", + "model.layers.19.input_layernorm.weight": "model-00017-of-00063.safetensors", + "model.layers.19.mlp.down_proj.weight": "model-00017-of-00063.safetensors", + "model.layers.19.mlp.gate_proj.weight": "model-00016-of-00063.safetensors", + "model.layers.19.mlp.up_proj.weight": "model-00016-of-00063.safetensors", + "model.layers.19.post_attention_layernorm.weight": "model-00017-of-00063.safetensors", + "model.layers.19.self_attn.k_proj.bias": "model-00016-of-00063.safetensors", + "model.layers.19.self_attn.k_proj.weight": "model-00016-of-00063.safetensors", + "model.layers.19.self_attn.o_proj.weight": "model-00016-of-00063.safetensors", + "model.layers.19.self_attn.q_proj.bias": "model-00016-of-00063.safetensors", + "model.layers.19.self_attn.q_proj.weight": "model-00016-of-00063.safetensors", + "model.layers.19.self_attn.v_proj.bias": "model-00016-of-00063.safetensors", + "model.layers.19.self_attn.v_proj.weight": "model-00016-of-00063.safetensors", + "model.layers.2.input_layernorm.weight": "model-00004-of-00063.safetensors", + "model.layers.2.mlp.down_proj.weight": "model-00004-of-00063.safetensors", + "model.layers.2.mlp.gate_proj.weight": "model-00003-of-00063.safetensors", + "model.layers.2.mlp.up_proj.weight": "model-00004-of-00063.safetensors", + "model.layers.2.post_attention_layernorm.weight": "model-00004-of-00063.safetensors", + "model.layers.2.self_attn.k_proj.bias": "model-00003-of-00063.safetensors", + "model.layers.2.self_attn.k_proj.weight": "model-00003-of-00063.safetensors", + "model.layers.2.self_attn.o_proj.weight": "model-00003-of-00063.safetensors", + "model.layers.2.self_attn.q_proj.bias": "model-00003-of-00063.safetensors", + "model.layers.2.self_attn.q_proj.weight": "model-00003-of-00063.safetensors", + "model.layers.2.self_attn.v_proj.bias": "model-00003-of-00063.safetensors", + "model.layers.2.self_attn.v_proj.weight": "model-00003-of-00063.safetensors", + "model.layers.20.input_layernorm.weight": "model-00017-of-00063.safetensors", + "model.layers.20.mlp.down_proj.weight": "model-00017-of-00063.safetensors", + "model.layers.20.mlp.gate_proj.weight": "model-00017-of-00063.safetensors", + "model.layers.20.mlp.up_proj.weight": "model-00017-of-00063.safetensors", + "model.layers.20.post_attention_layernorm.weight": "model-00017-of-00063.safetensors", + "model.layers.20.self_attn.k_proj.bias": "model-00017-of-00063.safetensors", + "model.layers.20.self_attn.k_proj.weight": "model-00017-of-00063.safetensors", + "model.layers.20.self_attn.o_proj.weight": "model-00017-of-00063.safetensors", + "model.layers.20.self_attn.q_proj.bias": "model-00017-of-00063.safetensors", + "model.layers.20.self_attn.q_proj.weight": "model-00017-of-00063.safetensors", + "model.layers.20.self_attn.v_proj.bias": "model-00017-of-00063.safetensors", + "model.layers.20.self_attn.v_proj.weight": "model-00017-of-00063.safetensors", + "model.layers.21.input_layernorm.weight": "model-00018-of-00063.safetensors", + "model.layers.21.mlp.down_proj.weight": "model-00018-of-00063.safetensors", + "model.layers.21.mlp.gate_proj.weight": "model-00018-of-00063.safetensors", + "model.layers.21.mlp.up_proj.weight": "model-00018-of-00063.safetensors", + "model.layers.21.post_attention_layernorm.weight": "model-00018-of-00063.safetensors", + "model.layers.21.self_attn.k_proj.bias": "model-00017-of-00063.safetensors", + "model.layers.21.self_attn.k_proj.weight": "model-00017-of-00063.safetensors", + "model.layers.21.self_attn.o_proj.weight": "model-00018-of-00063.safetensors", + "model.layers.21.self_attn.q_proj.bias": "model-00017-of-00063.safetensors", + "model.layers.21.self_attn.q_proj.weight": "model-00017-of-00063.safetensors", + "model.layers.21.self_attn.v_proj.bias": "model-00017-of-00063.safetensors", + "model.layers.21.self_attn.v_proj.weight": "model-00017-of-00063.safetensors", + "model.layers.22.input_layernorm.weight": "model-00019-of-00063.safetensors", + "model.layers.22.mlp.down_proj.weight": "model-00019-of-00063.safetensors", + "model.layers.22.mlp.gate_proj.weight": "model-00018-of-00063.safetensors", + "model.layers.22.mlp.up_proj.weight": "model-00019-of-00063.safetensors", + "model.layers.22.post_attention_layernorm.weight": "model-00019-of-00063.safetensors", + "model.layers.22.self_attn.k_proj.bias": "model-00018-of-00063.safetensors", + "model.layers.22.self_attn.k_proj.weight": "model-00018-of-00063.safetensors", + "model.layers.22.self_attn.o_proj.weight": "model-00018-of-00063.safetensors", + "model.layers.22.self_attn.q_proj.bias": "model-00018-of-00063.safetensors", + "model.layers.22.self_attn.q_proj.weight": "model-00018-of-00063.safetensors", + "model.layers.22.self_attn.v_proj.bias": "model-00018-of-00063.safetensors", + "model.layers.22.self_attn.v_proj.weight": "model-00018-of-00063.safetensors", + "model.layers.23.input_layernorm.weight": "model-00020-of-00063.safetensors", + "model.layers.23.mlp.down_proj.weight": "model-00020-of-00063.safetensors", + "model.layers.23.mlp.gate_proj.weight": "model-00019-of-00063.safetensors", + "model.layers.23.mlp.up_proj.weight": "model-00019-of-00063.safetensors", + "model.layers.23.post_attention_layernorm.weight": "model-00020-of-00063.safetensors", + "model.layers.23.self_attn.k_proj.bias": "model-00019-of-00063.safetensors", + "model.layers.23.self_attn.k_proj.weight": "model-00019-of-00063.safetensors", + "model.layers.23.self_attn.o_proj.weight": "model-00019-of-00063.safetensors", + "model.layers.23.self_attn.q_proj.bias": "model-00019-of-00063.safetensors", + "model.layers.23.self_attn.q_proj.weight": "model-00019-of-00063.safetensors", + "model.layers.23.self_attn.v_proj.bias": "model-00019-of-00063.safetensors", + "model.layers.23.self_attn.v_proj.weight": "model-00019-of-00063.safetensors", + "model.layers.24.input_layernorm.weight": "model-00020-of-00063.safetensors", + "model.layers.24.mlp.down_proj.weight": "model-00020-of-00063.safetensors", + "model.layers.24.mlp.gate_proj.weight": "model-00020-of-00063.safetensors", + "model.layers.24.mlp.up_proj.weight": "model-00020-of-00063.safetensors", + "model.layers.24.post_attention_layernorm.weight": "model-00020-of-00063.safetensors", + "model.layers.24.self_attn.k_proj.bias": "model-00020-of-00063.safetensors", + "model.layers.24.self_attn.k_proj.weight": "model-00020-of-00063.safetensors", + "model.layers.24.self_attn.o_proj.weight": "model-00020-of-00063.safetensors", + "model.layers.24.self_attn.q_proj.bias": "model-00020-of-00063.safetensors", + "model.layers.24.self_attn.q_proj.weight": "model-00020-of-00063.safetensors", + "model.layers.24.self_attn.v_proj.bias": "model-00020-of-00063.safetensors", + "model.layers.24.self_attn.v_proj.weight": "model-00020-of-00063.safetensors", + "model.layers.25.input_layernorm.weight": "model-00021-of-00063.safetensors", + "model.layers.25.mlp.down_proj.weight": "model-00021-of-00063.safetensors", + "model.layers.25.mlp.gate_proj.weight": "model-00021-of-00063.safetensors", + "model.layers.25.mlp.up_proj.weight": "model-00021-of-00063.safetensors", + "model.layers.25.post_attention_layernorm.weight": "model-00021-of-00063.safetensors", + "model.layers.25.self_attn.k_proj.bias": "model-00020-of-00063.safetensors", + "model.layers.25.self_attn.k_proj.weight": "model-00020-of-00063.safetensors", + "model.layers.25.self_attn.o_proj.weight": "model-00021-of-00063.safetensors", + "model.layers.25.self_attn.q_proj.bias": "model-00020-of-00063.safetensors", + "model.layers.25.self_attn.q_proj.weight": "model-00020-of-00063.safetensors", + "model.layers.25.self_attn.v_proj.bias": "model-00020-of-00063.safetensors", + "model.layers.25.self_attn.v_proj.weight": "model-00020-of-00063.safetensors", + "model.layers.26.input_layernorm.weight": "model-00022-of-00063.safetensors", + "model.layers.26.mlp.down_proj.weight": "model-00022-of-00063.safetensors", + "model.layers.26.mlp.gate_proj.weight": "model-00021-of-00063.safetensors", + "model.layers.26.mlp.up_proj.weight": "model-00022-of-00063.safetensors", + "model.layers.26.post_attention_layernorm.weight": "model-00022-of-00063.safetensors", + "model.layers.26.self_attn.k_proj.bias": "model-00021-of-00063.safetensors", + "model.layers.26.self_attn.k_proj.weight": "model-00021-of-00063.safetensors", + "model.layers.26.self_attn.o_proj.weight": "model-00021-of-00063.safetensors", + "model.layers.26.self_attn.q_proj.bias": "model-00021-of-00063.safetensors", + "model.layers.26.self_attn.q_proj.weight": "model-00021-of-00063.safetensors", + "model.layers.26.self_attn.v_proj.bias": "model-00021-of-00063.safetensors", + "model.layers.26.self_attn.v_proj.weight": "model-00021-of-00063.safetensors", + "model.layers.27.input_layernorm.weight": "model-00023-of-00063.safetensors", + "model.layers.27.mlp.down_proj.weight": "model-00023-of-00063.safetensors", + "model.layers.27.mlp.gate_proj.weight": "model-00022-of-00063.safetensors", + "model.layers.27.mlp.up_proj.weight": "model-00022-of-00063.safetensors", + "model.layers.27.post_attention_layernorm.weight": "model-00023-of-00063.safetensors", + "model.layers.27.self_attn.k_proj.bias": "model-00022-of-00063.safetensors", + "model.layers.27.self_attn.k_proj.weight": "model-00022-of-00063.safetensors", + "model.layers.27.self_attn.o_proj.weight": "model-00022-of-00063.safetensors", + "model.layers.27.self_attn.q_proj.bias": "model-00022-of-00063.safetensors", + "model.layers.27.self_attn.q_proj.weight": "model-00022-of-00063.safetensors", + "model.layers.27.self_attn.v_proj.bias": "model-00022-of-00063.safetensors", + "model.layers.27.self_attn.v_proj.weight": "model-00022-of-00063.safetensors", + "model.layers.28.input_layernorm.weight": "model-00023-of-00063.safetensors", + "model.layers.28.mlp.down_proj.weight": "model-00023-of-00063.safetensors", + "model.layers.28.mlp.gate_proj.weight": "model-00023-of-00063.safetensors", + "model.layers.28.mlp.up_proj.weight": "model-00023-of-00063.safetensors", + "model.layers.28.post_attention_layernorm.weight": "model-00023-of-00063.safetensors", + "model.layers.28.self_attn.k_proj.bias": "model-00023-of-00063.safetensors", + "model.layers.28.self_attn.k_proj.weight": "model-00023-of-00063.safetensors", + "model.layers.28.self_attn.o_proj.weight": "model-00023-of-00063.safetensors", + "model.layers.28.self_attn.q_proj.bias": "model-00023-of-00063.safetensors", + "model.layers.28.self_attn.q_proj.weight": "model-00023-of-00063.safetensors", + "model.layers.28.self_attn.v_proj.bias": "model-00023-of-00063.safetensors", + "model.layers.28.self_attn.v_proj.weight": "model-00023-of-00063.safetensors", + "model.layers.29.input_layernorm.weight": "model-00024-of-00063.safetensors", + "model.layers.29.mlp.down_proj.weight": "model-00024-of-00063.safetensors", + "model.layers.29.mlp.gate_proj.weight": "model-00024-of-00063.safetensors", + "model.layers.29.mlp.up_proj.weight": "model-00024-of-00063.safetensors", + "model.layers.29.post_attention_layernorm.weight": "model-00024-of-00063.safetensors", + "model.layers.29.self_attn.k_proj.bias": "model-00023-of-00063.safetensors", + "model.layers.29.self_attn.k_proj.weight": "model-00023-of-00063.safetensors", + "model.layers.29.self_attn.o_proj.weight": "model-00024-of-00063.safetensors", + "model.layers.29.self_attn.q_proj.bias": "model-00023-of-00063.safetensors", + "model.layers.29.self_attn.q_proj.weight": "model-00023-of-00063.safetensors", + "model.layers.29.self_attn.v_proj.bias": "model-00023-of-00063.safetensors", + "model.layers.29.self_attn.v_proj.weight": "model-00023-of-00063.safetensors", + "model.layers.3.input_layernorm.weight": "model-00005-of-00063.safetensors", + "model.layers.3.mlp.down_proj.weight": "model-00005-of-00063.safetensors", + "model.layers.3.mlp.gate_proj.weight": "model-00004-of-00063.safetensors", + "model.layers.3.mlp.up_proj.weight": "model-00004-of-00063.safetensors", + "model.layers.3.post_attention_layernorm.weight": "model-00005-of-00063.safetensors", + "model.layers.3.self_attn.k_proj.bias": "model-00004-of-00063.safetensors", + "model.layers.3.self_attn.k_proj.weight": "model-00004-of-00063.safetensors", + "model.layers.3.self_attn.o_proj.weight": "model-00004-of-00063.safetensors", + "model.layers.3.self_attn.q_proj.bias": "model-00004-of-00063.safetensors", + "model.layers.3.self_attn.q_proj.weight": "model-00004-of-00063.safetensors", + "model.layers.3.self_attn.v_proj.bias": "model-00004-of-00063.safetensors", + "model.layers.3.self_attn.v_proj.weight": "model-00004-of-00063.safetensors", + "model.layers.30.input_layernorm.weight": "model-00025-of-00063.safetensors", + "model.layers.30.mlp.down_proj.weight": "model-00025-of-00063.safetensors", + "model.layers.30.mlp.gate_proj.weight": "model-00024-of-00063.safetensors", + "model.layers.30.mlp.up_proj.weight": "model-00025-of-00063.safetensors", + "model.layers.30.post_attention_layernorm.weight": "model-00025-of-00063.safetensors", + "model.layers.30.self_attn.k_proj.bias": "model-00024-of-00063.safetensors", + "model.layers.30.self_attn.k_proj.weight": "model-00024-of-00063.safetensors", + "model.layers.30.self_attn.o_proj.weight": "model-00024-of-00063.safetensors", + "model.layers.30.self_attn.q_proj.bias": "model-00024-of-00063.safetensors", + "model.layers.30.self_attn.q_proj.weight": "model-00024-of-00063.safetensors", + "model.layers.30.self_attn.v_proj.bias": "model-00024-of-00063.safetensors", + "model.layers.30.self_attn.v_proj.weight": "model-00024-of-00063.safetensors", + "model.layers.31.input_layernorm.weight": "model-00026-of-00063.safetensors", + "model.layers.31.mlp.down_proj.weight": "model-00026-of-00063.safetensors", + "model.layers.31.mlp.gate_proj.weight": "model-00025-of-00063.safetensors", + "model.layers.31.mlp.up_proj.weight": "model-00025-of-00063.safetensors", + "model.layers.31.post_attention_layernorm.weight": "model-00026-of-00063.safetensors", + "model.layers.31.self_attn.k_proj.bias": "model-00025-of-00063.safetensors", + "model.layers.31.self_attn.k_proj.weight": "model-00025-of-00063.safetensors", + "model.layers.31.self_attn.o_proj.weight": "model-00025-of-00063.safetensors", + "model.layers.31.self_attn.q_proj.bias": "model-00025-of-00063.safetensors", + "model.layers.31.self_attn.q_proj.weight": "model-00025-of-00063.safetensors", + "model.layers.31.self_attn.v_proj.bias": "model-00025-of-00063.safetensors", + "model.layers.31.self_attn.v_proj.weight": "model-00025-of-00063.safetensors", + "model.layers.32.input_layernorm.weight": "model-00026-of-00063.safetensors", + "model.layers.32.mlp.down_proj.weight": "model-00026-of-00063.safetensors", + "model.layers.32.mlp.gate_proj.weight": "model-00026-of-00063.safetensors", + "model.layers.32.mlp.up_proj.weight": "model-00026-of-00063.safetensors", + "model.layers.32.post_attention_layernorm.weight": "model-00026-of-00063.safetensors", + "model.layers.32.self_attn.k_proj.bias": "model-00026-of-00063.safetensors", + "model.layers.32.self_attn.k_proj.weight": "model-00026-of-00063.safetensors", + "model.layers.32.self_attn.o_proj.weight": "model-00026-of-00063.safetensors", + "model.layers.32.self_attn.q_proj.bias": "model-00026-of-00063.safetensors", + "model.layers.32.self_attn.q_proj.weight": "model-00026-of-00063.safetensors", + "model.layers.32.self_attn.v_proj.bias": "model-00026-of-00063.safetensors", + "model.layers.32.self_attn.v_proj.weight": "model-00026-of-00063.safetensors", + "model.layers.33.input_layernorm.weight": "model-00027-of-00063.safetensors", + "model.layers.33.mlp.down_proj.weight": "model-00027-of-00063.safetensors", + "model.layers.33.mlp.gate_proj.weight": "model-00027-of-00063.safetensors", + "model.layers.33.mlp.up_proj.weight": "model-00027-of-00063.safetensors", + "model.layers.33.post_attention_layernorm.weight": "model-00027-of-00063.safetensors", + "model.layers.33.self_attn.k_proj.bias": "model-00026-of-00063.safetensors", + "model.layers.33.self_attn.k_proj.weight": "model-00026-of-00063.safetensors", + "model.layers.33.self_attn.o_proj.weight": "model-00027-of-00063.safetensors", + "model.layers.33.self_attn.q_proj.bias": "model-00026-of-00063.safetensors", + "model.layers.33.self_attn.q_proj.weight": "model-00026-of-00063.safetensors", + "model.layers.33.self_attn.v_proj.bias": "model-00026-of-00063.safetensors", + "model.layers.33.self_attn.v_proj.weight": "model-00026-of-00063.safetensors", + "model.layers.34.input_layernorm.weight": "model-00028-of-00063.safetensors", + "model.layers.34.mlp.down_proj.weight": "model-00028-of-00063.safetensors", + "model.layers.34.mlp.gate_proj.weight": "model-00027-of-00063.safetensors", + "model.layers.34.mlp.up_proj.weight": "model-00028-of-00063.safetensors", + "model.layers.34.post_attention_layernorm.weight": "model-00028-of-00063.safetensors", + "model.layers.34.self_attn.k_proj.bias": "model-00027-of-00063.safetensors", + "model.layers.34.self_attn.k_proj.weight": "model-00027-of-00063.safetensors", + "model.layers.34.self_attn.o_proj.weight": "model-00027-of-00063.safetensors", + "model.layers.34.self_attn.q_proj.bias": "model-00027-of-00063.safetensors", + "model.layers.34.self_attn.q_proj.weight": "model-00027-of-00063.safetensors", + "model.layers.34.self_attn.v_proj.bias": "model-00027-of-00063.safetensors", + "model.layers.34.self_attn.v_proj.weight": "model-00027-of-00063.safetensors", + "model.layers.35.input_layernorm.weight": "model-00029-of-00063.safetensors", + "model.layers.35.mlp.down_proj.weight": "model-00029-of-00063.safetensors", + "model.layers.35.mlp.gate_proj.weight": "model-00028-of-00063.safetensors", + "model.layers.35.mlp.up_proj.weight": "model-00028-of-00063.safetensors", + "model.layers.35.post_attention_layernorm.weight": "model-00029-of-00063.safetensors", + "model.layers.35.self_attn.k_proj.bias": "model-00028-of-00063.safetensors", + "model.layers.35.self_attn.k_proj.weight": "model-00028-of-00063.safetensors", + "model.layers.35.self_attn.o_proj.weight": "model-00028-of-00063.safetensors", + "model.layers.35.self_attn.q_proj.bias": "model-00028-of-00063.safetensors", + "model.layers.35.self_attn.q_proj.weight": "model-00028-of-00063.safetensors", + "model.layers.35.self_attn.v_proj.bias": "model-00028-of-00063.safetensors", + "model.layers.35.self_attn.v_proj.weight": "model-00028-of-00063.safetensors", + "model.layers.36.input_layernorm.weight": "model-00029-of-00063.safetensors", + "model.layers.36.mlp.down_proj.weight": "model-00029-of-00063.safetensors", + "model.layers.36.mlp.gate_proj.weight": "model-00029-of-00063.safetensors", + "model.layers.36.mlp.up_proj.weight": "model-00029-of-00063.safetensors", + "model.layers.36.post_attention_layernorm.weight": "model-00029-of-00063.safetensors", + "model.layers.36.self_attn.k_proj.bias": "model-00029-of-00063.safetensors", + "model.layers.36.self_attn.k_proj.weight": "model-00029-of-00063.safetensors", + "model.layers.36.self_attn.o_proj.weight": "model-00029-of-00063.safetensors", + "model.layers.36.self_attn.q_proj.bias": "model-00029-of-00063.safetensors", + "model.layers.36.self_attn.q_proj.weight": "model-00029-of-00063.safetensors", + "model.layers.36.self_attn.v_proj.bias": "model-00029-of-00063.safetensors", + "model.layers.36.self_attn.v_proj.weight": "model-00029-of-00063.safetensors", + "model.layers.37.input_layernorm.weight": "model-00030-of-00063.safetensors", + "model.layers.37.mlp.down_proj.weight": "model-00030-of-00063.safetensors", + "model.layers.37.mlp.gate_proj.weight": "model-00030-of-00063.safetensors", + "model.layers.37.mlp.up_proj.weight": "model-00030-of-00063.safetensors", + "model.layers.37.post_attention_layernorm.weight": "model-00030-of-00063.safetensors", + "model.layers.37.self_attn.k_proj.bias": "model-00029-of-00063.safetensors", + "model.layers.37.self_attn.k_proj.weight": "model-00029-of-00063.safetensors", + "model.layers.37.self_attn.o_proj.weight": "model-00030-of-00063.safetensors", + "model.layers.37.self_attn.q_proj.bias": "model-00029-of-00063.safetensors", + "model.layers.37.self_attn.q_proj.weight": "model-00029-of-00063.safetensors", + "model.layers.37.self_attn.v_proj.bias": "model-00029-of-00063.safetensors", + "model.layers.37.self_attn.v_proj.weight": "model-00029-of-00063.safetensors", + "model.layers.38.input_layernorm.weight": "model-00031-of-00063.safetensors", + "model.layers.38.mlp.down_proj.weight": "model-00031-of-00063.safetensors", + "model.layers.38.mlp.gate_proj.weight": "model-00030-of-00063.safetensors", + "model.layers.38.mlp.up_proj.weight": "model-00031-of-00063.safetensors", + "model.layers.38.post_attention_layernorm.weight": "model-00031-of-00063.safetensors", + "model.layers.38.self_attn.k_proj.bias": "model-00030-of-00063.safetensors", + "model.layers.38.self_attn.k_proj.weight": "model-00030-of-00063.safetensors", + "model.layers.38.self_attn.o_proj.weight": "model-00030-of-00063.safetensors", + "model.layers.38.self_attn.q_proj.bias": "model-00030-of-00063.safetensors", + "model.layers.38.self_attn.q_proj.weight": "model-00030-of-00063.safetensors", + "model.layers.38.self_attn.v_proj.bias": "model-00030-of-00063.safetensors", + "model.layers.38.self_attn.v_proj.weight": "model-00030-of-00063.safetensors", + "model.layers.39.input_layernorm.weight": "model-00032-of-00063.safetensors", + "model.layers.39.mlp.down_proj.weight": "model-00032-of-00063.safetensors", + "model.layers.39.mlp.gate_proj.weight": "model-00031-of-00063.safetensors", + "model.layers.39.mlp.up_proj.weight": "model-00031-of-00063.safetensors", + "model.layers.39.post_attention_layernorm.weight": "model-00032-of-00063.safetensors", + "model.layers.39.self_attn.k_proj.bias": "model-00031-of-00063.safetensors", + "model.layers.39.self_attn.k_proj.weight": "model-00031-of-00063.safetensors", + "model.layers.39.self_attn.o_proj.weight": "model-00031-of-00063.safetensors", + "model.layers.39.self_attn.q_proj.bias": "model-00031-of-00063.safetensors", + "model.layers.39.self_attn.q_proj.weight": "model-00031-of-00063.safetensors", + "model.layers.39.self_attn.v_proj.bias": "model-00031-of-00063.safetensors", + "model.layers.39.self_attn.v_proj.weight": "model-00031-of-00063.safetensors", + "model.layers.4.input_layernorm.weight": "model-00005-of-00063.safetensors", + "model.layers.4.mlp.down_proj.weight": "model-00005-of-00063.safetensors", + "model.layers.4.mlp.gate_proj.weight": "model-00005-of-00063.safetensors", + "model.layers.4.mlp.up_proj.weight": "model-00005-of-00063.safetensors", + "model.layers.4.post_attention_layernorm.weight": "model-00005-of-00063.safetensors", + "model.layers.4.self_attn.k_proj.bias": "model-00005-of-00063.safetensors", + "model.layers.4.self_attn.k_proj.weight": "model-00005-of-00063.safetensors", + "model.layers.4.self_attn.o_proj.weight": "model-00005-of-00063.safetensors", + "model.layers.4.self_attn.q_proj.bias": "model-00005-of-00063.safetensors", + "model.layers.4.self_attn.q_proj.weight": "model-00005-of-00063.safetensors", + "model.layers.4.self_attn.v_proj.bias": "model-00005-of-00063.safetensors", + "model.layers.4.self_attn.v_proj.weight": "model-00005-of-00063.safetensors", + "model.layers.40.input_layernorm.weight": "model-00032-of-00063.safetensors", + "model.layers.40.mlp.down_proj.weight": "model-00032-of-00063.safetensors", + "model.layers.40.mlp.gate_proj.weight": "model-00032-of-00063.safetensors", + "model.layers.40.mlp.up_proj.weight": "model-00032-of-00063.safetensors", + "model.layers.40.post_attention_layernorm.weight": "model-00032-of-00063.safetensors", + "model.layers.40.self_attn.k_proj.bias": "model-00032-of-00063.safetensors", + "model.layers.40.self_attn.k_proj.weight": "model-00032-of-00063.safetensors", + "model.layers.40.self_attn.o_proj.weight": "model-00032-of-00063.safetensors", + "model.layers.40.self_attn.q_proj.bias": "model-00032-of-00063.safetensors", + "model.layers.40.self_attn.q_proj.weight": "model-00032-of-00063.safetensors", + "model.layers.40.self_attn.v_proj.bias": "model-00032-of-00063.safetensors", + "model.layers.40.self_attn.v_proj.weight": "model-00032-of-00063.safetensors", + "model.layers.41.input_layernorm.weight": "model-00033-of-00063.safetensors", + "model.layers.41.mlp.down_proj.weight": "model-00033-of-00063.safetensors", + "model.layers.41.mlp.gate_proj.weight": "model-00033-of-00063.safetensors", + "model.layers.41.mlp.up_proj.weight": "model-00033-of-00063.safetensors", + "model.layers.41.post_attention_layernorm.weight": "model-00033-of-00063.safetensors", + "model.layers.41.self_attn.k_proj.bias": "model-00032-of-00063.safetensors", + "model.layers.41.self_attn.k_proj.weight": "model-00032-of-00063.safetensors", + "model.layers.41.self_attn.o_proj.weight": "model-00033-of-00063.safetensors", + "model.layers.41.self_attn.q_proj.bias": "model-00032-of-00063.safetensors", + "model.layers.41.self_attn.q_proj.weight": "model-00032-of-00063.safetensors", + "model.layers.41.self_attn.v_proj.bias": "model-00032-of-00063.safetensors", + "model.layers.41.self_attn.v_proj.weight": "model-00032-of-00063.safetensors", + "model.layers.42.input_layernorm.weight": "model-00034-of-00063.safetensors", + "model.layers.42.mlp.down_proj.weight": "model-00034-of-00063.safetensors", + "model.layers.42.mlp.gate_proj.weight": "model-00033-of-00063.safetensors", + "model.layers.42.mlp.up_proj.weight": "model-00034-of-00063.safetensors", + "model.layers.42.post_attention_layernorm.weight": "model-00034-of-00063.safetensors", + "model.layers.42.self_attn.k_proj.bias": "model-00033-of-00063.safetensors", + "model.layers.42.self_attn.k_proj.weight": "model-00033-of-00063.safetensors", + "model.layers.42.self_attn.o_proj.weight": "model-00033-of-00063.safetensors", + "model.layers.42.self_attn.q_proj.bias": "model-00033-of-00063.safetensors", + "model.layers.42.self_attn.q_proj.weight": "model-00033-of-00063.safetensors", + "model.layers.42.self_attn.v_proj.bias": "model-00033-of-00063.safetensors", + "model.layers.42.self_attn.v_proj.weight": "model-00033-of-00063.safetensors", + "model.layers.43.input_layernorm.weight": "model-00035-of-00063.safetensors", + "model.layers.43.mlp.down_proj.weight": "model-00035-of-00063.safetensors", + "model.layers.43.mlp.gate_proj.weight": "model-00034-of-00063.safetensors", + "model.layers.43.mlp.up_proj.weight": "model-00034-of-00063.safetensors", + "model.layers.43.post_attention_layernorm.weight": "model-00035-of-00063.safetensors", + "model.layers.43.self_attn.k_proj.bias": "model-00034-of-00063.safetensors", + "model.layers.43.self_attn.k_proj.weight": "model-00034-of-00063.safetensors", + "model.layers.43.self_attn.o_proj.weight": "model-00034-of-00063.safetensors", + "model.layers.43.self_attn.q_proj.bias": "model-00034-of-00063.safetensors", + "model.layers.43.self_attn.q_proj.weight": "model-00034-of-00063.safetensors", + "model.layers.43.self_attn.v_proj.bias": "model-00034-of-00063.safetensors", + "model.layers.43.self_attn.v_proj.weight": "model-00034-of-00063.safetensors", + "model.layers.44.input_layernorm.weight": "model-00035-of-00063.safetensors", + "model.layers.44.mlp.down_proj.weight": "model-00035-of-00063.safetensors", + "model.layers.44.mlp.gate_proj.weight": "model-00035-of-00063.safetensors", + "model.layers.44.mlp.up_proj.weight": "model-00035-of-00063.safetensors", + "model.layers.44.post_attention_layernorm.weight": "model-00035-of-00063.safetensors", + "model.layers.44.self_attn.k_proj.bias": "model-00035-of-00063.safetensors", + "model.layers.44.self_attn.k_proj.weight": "model-00035-of-00063.safetensors", + "model.layers.44.self_attn.o_proj.weight": "model-00035-of-00063.safetensors", + "model.layers.44.self_attn.q_proj.bias": "model-00035-of-00063.safetensors", + "model.layers.44.self_attn.q_proj.weight": "model-00035-of-00063.safetensors", + "model.layers.44.self_attn.v_proj.bias": "model-00035-of-00063.safetensors", + "model.layers.44.self_attn.v_proj.weight": "model-00035-of-00063.safetensors", + "model.layers.45.input_layernorm.weight": "model-00036-of-00063.safetensors", + "model.layers.45.mlp.down_proj.weight": "model-00036-of-00063.safetensors", + "model.layers.45.mlp.gate_proj.weight": "model-00036-of-00063.safetensors", + "model.layers.45.mlp.up_proj.weight": "model-00036-of-00063.safetensors", + "model.layers.45.post_attention_layernorm.weight": "model-00036-of-00063.safetensors", + "model.layers.45.self_attn.k_proj.bias": "model-00035-of-00063.safetensors", + "model.layers.45.self_attn.k_proj.weight": "model-00035-of-00063.safetensors", + "model.layers.45.self_attn.o_proj.weight": "model-00036-of-00063.safetensors", + "model.layers.45.self_attn.q_proj.bias": "model-00035-of-00063.safetensors", + "model.layers.45.self_attn.q_proj.weight": "model-00035-of-00063.safetensors", + "model.layers.45.self_attn.v_proj.bias": "model-00035-of-00063.safetensors", + "model.layers.45.self_attn.v_proj.weight": "model-00035-of-00063.safetensors", + "model.layers.46.input_layernorm.weight": "model-00037-of-00063.safetensors", + "model.layers.46.mlp.down_proj.weight": "model-00037-of-00063.safetensors", + "model.layers.46.mlp.gate_proj.weight": "model-00036-of-00063.safetensors", + "model.layers.46.mlp.up_proj.weight": "model-00037-of-00063.safetensors", + "model.layers.46.post_attention_layernorm.weight": "model-00037-of-00063.safetensors", + "model.layers.46.self_attn.k_proj.bias": "model-00036-of-00063.safetensors", + "model.layers.46.self_attn.k_proj.weight": "model-00036-of-00063.safetensors", + "model.layers.46.self_attn.o_proj.weight": "model-00036-of-00063.safetensors", + "model.layers.46.self_attn.q_proj.bias": "model-00036-of-00063.safetensors", + "model.layers.46.self_attn.q_proj.weight": "model-00036-of-00063.safetensors", + "model.layers.46.self_attn.v_proj.bias": "model-00036-of-00063.safetensors", + "model.layers.46.self_attn.v_proj.weight": "model-00036-of-00063.safetensors", + "model.layers.47.input_layernorm.weight": "model-00038-of-00063.safetensors", + "model.layers.47.mlp.down_proj.weight": "model-00038-of-00063.safetensors", + "model.layers.47.mlp.gate_proj.weight": "model-00037-of-00063.safetensors", + "model.layers.47.mlp.up_proj.weight": "model-00037-of-00063.safetensors", + "model.layers.47.post_attention_layernorm.weight": "model-00038-of-00063.safetensors", + "model.layers.47.self_attn.k_proj.bias": "model-00037-of-00063.safetensors", + "model.layers.47.self_attn.k_proj.weight": "model-00037-of-00063.safetensors", + "model.layers.47.self_attn.o_proj.weight": "model-00037-of-00063.safetensors", + "model.layers.47.self_attn.q_proj.bias": "model-00037-of-00063.safetensors", + "model.layers.47.self_attn.q_proj.weight": "model-00037-of-00063.safetensors", + "model.layers.47.self_attn.v_proj.bias": "model-00037-of-00063.safetensors", + "model.layers.47.self_attn.v_proj.weight": "model-00037-of-00063.safetensors", + "model.layers.48.input_layernorm.weight": "model-00038-of-00063.safetensors", + "model.layers.48.mlp.down_proj.weight": "model-00038-of-00063.safetensors", + "model.layers.48.mlp.gate_proj.weight": "model-00038-of-00063.safetensors", + "model.layers.48.mlp.up_proj.weight": "model-00038-of-00063.safetensors", + "model.layers.48.post_attention_layernorm.weight": "model-00038-of-00063.safetensors", + "model.layers.48.self_attn.k_proj.bias": "model-00038-of-00063.safetensors", + "model.layers.48.self_attn.k_proj.weight": "model-00038-of-00063.safetensors", + "model.layers.48.self_attn.o_proj.weight": "model-00038-of-00063.safetensors", + "model.layers.48.self_attn.q_proj.bias": "model-00038-of-00063.safetensors", + "model.layers.48.self_attn.q_proj.weight": "model-00038-of-00063.safetensors", + "model.layers.48.self_attn.v_proj.bias": "model-00038-of-00063.safetensors", + "model.layers.48.self_attn.v_proj.weight": "model-00038-of-00063.safetensors", + "model.layers.49.input_layernorm.weight": "model-00039-of-00063.safetensors", + "model.layers.49.mlp.down_proj.weight": "model-00039-of-00063.safetensors", + "model.layers.49.mlp.gate_proj.weight": "model-00039-of-00063.safetensors", + "model.layers.49.mlp.up_proj.weight": "model-00039-of-00063.safetensors", + "model.layers.49.post_attention_layernorm.weight": "model-00039-of-00063.safetensors", + "model.layers.49.self_attn.k_proj.bias": "model-00038-of-00063.safetensors", + "model.layers.49.self_attn.k_proj.weight": "model-00038-of-00063.safetensors", + "model.layers.49.self_attn.o_proj.weight": "model-00039-of-00063.safetensors", + "model.layers.49.self_attn.q_proj.bias": "model-00038-of-00063.safetensors", + "model.layers.49.self_attn.q_proj.weight": "model-00038-of-00063.safetensors", + "model.layers.49.self_attn.v_proj.bias": "model-00038-of-00063.safetensors", + "model.layers.49.self_attn.v_proj.weight": "model-00038-of-00063.safetensors", + "model.layers.5.input_layernorm.weight": "model-00006-of-00063.safetensors", + "model.layers.5.mlp.down_proj.weight": "model-00006-of-00063.safetensors", + "model.layers.5.mlp.gate_proj.weight": "model-00006-of-00063.safetensors", + "model.layers.5.mlp.up_proj.weight": "model-00006-of-00063.safetensors", + "model.layers.5.post_attention_layernorm.weight": "model-00006-of-00063.safetensors", + "model.layers.5.self_attn.k_proj.bias": "model-00005-of-00063.safetensors", + "model.layers.5.self_attn.k_proj.weight": "model-00005-of-00063.safetensors", + "model.layers.5.self_attn.o_proj.weight": "model-00006-of-00063.safetensors", + "model.layers.5.self_attn.q_proj.bias": "model-00005-of-00063.safetensors", + "model.layers.5.self_attn.q_proj.weight": "model-00005-of-00063.safetensors", + "model.layers.5.self_attn.v_proj.bias": "model-00005-of-00063.safetensors", + "model.layers.5.self_attn.v_proj.weight": "model-00005-of-00063.safetensors", + "model.layers.50.input_layernorm.weight": "model-00040-of-00063.safetensors", + "model.layers.50.mlp.down_proj.weight": "model-00040-of-00063.safetensors", + "model.layers.50.mlp.gate_proj.weight": "model-00039-of-00063.safetensors", + "model.layers.50.mlp.up_proj.weight": "model-00040-of-00063.safetensors", + "model.layers.50.post_attention_layernorm.weight": "model-00040-of-00063.safetensors", + "model.layers.50.self_attn.k_proj.bias": "model-00039-of-00063.safetensors", + "model.layers.50.self_attn.k_proj.weight": "model-00039-of-00063.safetensors", + "model.layers.50.self_attn.o_proj.weight": "model-00039-of-00063.safetensors", + "model.layers.50.self_attn.q_proj.bias": "model-00039-of-00063.safetensors", + "model.layers.50.self_attn.q_proj.weight": "model-00039-of-00063.safetensors", + "model.layers.50.self_attn.v_proj.bias": "model-00039-of-00063.safetensors", + "model.layers.50.self_attn.v_proj.weight": "model-00039-of-00063.safetensors", + "model.layers.51.input_layernorm.weight": "model-00041-of-00063.safetensors", + "model.layers.51.mlp.down_proj.weight": "model-00041-of-00063.safetensors", + "model.layers.51.mlp.gate_proj.weight": "model-00040-of-00063.safetensors", + "model.layers.51.mlp.up_proj.weight": "model-00040-of-00063.safetensors", + "model.layers.51.post_attention_layernorm.weight": "model-00041-of-00063.safetensors", + "model.layers.51.self_attn.k_proj.bias": "model-00040-of-00063.safetensors", + "model.layers.51.self_attn.k_proj.weight": "model-00040-of-00063.safetensors", + "model.layers.51.self_attn.o_proj.weight": "model-00040-of-00063.safetensors", + "model.layers.51.self_attn.q_proj.bias": "model-00040-of-00063.safetensors", + "model.layers.51.self_attn.q_proj.weight": "model-00040-of-00063.safetensors", + "model.layers.51.self_attn.v_proj.bias": "model-00040-of-00063.safetensors", + "model.layers.51.self_attn.v_proj.weight": "model-00040-of-00063.safetensors", + "model.layers.52.input_layernorm.weight": "model-00041-of-00063.safetensors", + "model.layers.52.mlp.down_proj.weight": "model-00041-of-00063.safetensors", + "model.layers.52.mlp.gate_proj.weight": "model-00041-of-00063.safetensors", + "model.layers.52.mlp.up_proj.weight": "model-00041-of-00063.safetensors", + "model.layers.52.post_attention_layernorm.weight": "model-00041-of-00063.safetensors", + "model.layers.52.self_attn.k_proj.bias": "model-00041-of-00063.safetensors", + "model.layers.52.self_attn.k_proj.weight": "model-00041-of-00063.safetensors", + "model.layers.52.self_attn.o_proj.weight": "model-00041-of-00063.safetensors", + "model.layers.52.self_attn.q_proj.bias": "model-00041-of-00063.safetensors", + "model.layers.52.self_attn.q_proj.weight": "model-00041-of-00063.safetensors", + "model.layers.52.self_attn.v_proj.bias": "model-00041-of-00063.safetensors", + "model.layers.52.self_attn.v_proj.weight": "model-00041-of-00063.safetensors", + "model.layers.53.input_layernorm.weight": "model-00042-of-00063.safetensors", + "model.layers.53.mlp.down_proj.weight": "model-00042-of-00063.safetensors", + "model.layers.53.mlp.gate_proj.weight": "model-00042-of-00063.safetensors", + "model.layers.53.mlp.up_proj.weight": "model-00042-of-00063.safetensors", + "model.layers.53.post_attention_layernorm.weight": "model-00042-of-00063.safetensors", + "model.layers.53.self_attn.k_proj.bias": "model-00041-of-00063.safetensors", + "model.layers.53.self_attn.k_proj.weight": "model-00041-of-00063.safetensors", + "model.layers.53.self_attn.o_proj.weight": "model-00042-of-00063.safetensors", + "model.layers.53.self_attn.q_proj.bias": "model-00041-of-00063.safetensors", + "model.layers.53.self_attn.q_proj.weight": "model-00041-of-00063.safetensors", + "model.layers.53.self_attn.v_proj.bias": "model-00041-of-00063.safetensors", + "model.layers.53.self_attn.v_proj.weight": "model-00041-of-00063.safetensors", + "model.layers.54.input_layernorm.weight": "model-00043-of-00063.safetensors", + "model.layers.54.mlp.down_proj.weight": "model-00043-of-00063.safetensors", + "model.layers.54.mlp.gate_proj.weight": "model-00042-of-00063.safetensors", + "model.layers.54.mlp.up_proj.weight": "model-00043-of-00063.safetensors", + "model.layers.54.post_attention_layernorm.weight": "model-00043-of-00063.safetensors", + "model.layers.54.self_attn.k_proj.bias": "model-00042-of-00063.safetensors", + "model.layers.54.self_attn.k_proj.weight": "model-00042-of-00063.safetensors", + "model.layers.54.self_attn.o_proj.weight": "model-00042-of-00063.safetensors", + "model.layers.54.self_attn.q_proj.bias": "model-00042-of-00063.safetensors", + "model.layers.54.self_attn.q_proj.weight": "model-00042-of-00063.safetensors", + "model.layers.54.self_attn.v_proj.bias": "model-00042-of-00063.safetensors", + "model.layers.54.self_attn.v_proj.weight": "model-00042-of-00063.safetensors", + "model.layers.55.input_layernorm.weight": "model-00044-of-00063.safetensors", + "model.layers.55.mlp.down_proj.weight": "model-00044-of-00063.safetensors", + "model.layers.55.mlp.gate_proj.weight": "model-00043-of-00063.safetensors", + "model.layers.55.mlp.up_proj.weight": "model-00043-of-00063.safetensors", + "model.layers.55.post_attention_layernorm.weight": "model-00044-of-00063.safetensors", + "model.layers.55.self_attn.k_proj.bias": "model-00043-of-00063.safetensors", + "model.layers.55.self_attn.k_proj.weight": "model-00043-of-00063.safetensors", + "model.layers.55.self_attn.o_proj.weight": "model-00043-of-00063.safetensors", + "model.layers.55.self_attn.q_proj.bias": "model-00043-of-00063.safetensors", + "model.layers.55.self_attn.q_proj.weight": "model-00043-of-00063.safetensors", + "model.layers.55.self_attn.v_proj.bias": "model-00043-of-00063.safetensors", + "model.layers.55.self_attn.v_proj.weight": "model-00043-of-00063.safetensors", + "model.layers.56.input_layernorm.weight": "model-00044-of-00063.safetensors", + "model.layers.56.mlp.down_proj.weight": "model-00044-of-00063.safetensors", + "model.layers.56.mlp.gate_proj.weight": "model-00044-of-00063.safetensors", + "model.layers.56.mlp.up_proj.weight": "model-00044-of-00063.safetensors", + "model.layers.56.post_attention_layernorm.weight": "model-00044-of-00063.safetensors", + "model.layers.56.self_attn.k_proj.bias": "model-00044-of-00063.safetensors", + "model.layers.56.self_attn.k_proj.weight": "model-00044-of-00063.safetensors", + "model.layers.56.self_attn.o_proj.weight": "model-00044-of-00063.safetensors", + "model.layers.56.self_attn.q_proj.bias": "model-00044-of-00063.safetensors", + "model.layers.56.self_attn.q_proj.weight": "model-00044-of-00063.safetensors", + "model.layers.56.self_attn.v_proj.bias": "model-00044-of-00063.safetensors", + "model.layers.56.self_attn.v_proj.weight": "model-00044-of-00063.safetensors", + "model.layers.57.input_layernorm.weight": "model-00045-of-00063.safetensors", + "model.layers.57.mlp.down_proj.weight": "model-00045-of-00063.safetensors", + "model.layers.57.mlp.gate_proj.weight": "model-00045-of-00063.safetensors", + "model.layers.57.mlp.up_proj.weight": "model-00045-of-00063.safetensors", + "model.layers.57.post_attention_layernorm.weight": "model-00045-of-00063.safetensors", + "model.layers.57.self_attn.k_proj.bias": "model-00044-of-00063.safetensors", + "model.layers.57.self_attn.k_proj.weight": "model-00044-of-00063.safetensors", + "model.layers.57.self_attn.o_proj.weight": "model-00045-of-00063.safetensors", + "model.layers.57.self_attn.q_proj.bias": "model-00044-of-00063.safetensors", + "model.layers.57.self_attn.q_proj.weight": "model-00044-of-00063.safetensors", + "model.layers.57.self_attn.v_proj.bias": "model-00044-of-00063.safetensors", + "model.layers.57.self_attn.v_proj.weight": "model-00044-of-00063.safetensors", + "model.layers.58.input_layernorm.weight": "model-00046-of-00063.safetensors", + "model.layers.58.mlp.down_proj.weight": "model-00046-of-00063.safetensors", + "model.layers.58.mlp.gate_proj.weight": "model-00045-of-00063.safetensors", + "model.layers.58.mlp.up_proj.weight": "model-00046-of-00063.safetensors", + "model.layers.58.post_attention_layernorm.weight": "model-00046-of-00063.safetensors", + "model.layers.58.self_attn.k_proj.bias": "model-00045-of-00063.safetensors", + "model.layers.58.self_attn.k_proj.weight": "model-00045-of-00063.safetensors", + "model.layers.58.self_attn.o_proj.weight": "model-00045-of-00063.safetensors", + "model.layers.58.self_attn.q_proj.bias": "model-00045-of-00063.safetensors", + "model.layers.58.self_attn.q_proj.weight": "model-00045-of-00063.safetensors", + "model.layers.58.self_attn.v_proj.bias": "model-00045-of-00063.safetensors", + "model.layers.58.self_attn.v_proj.weight": "model-00045-of-00063.safetensors", + "model.layers.59.input_layernorm.weight": "model-00047-of-00063.safetensors", + "model.layers.59.mlp.down_proj.weight": "model-00047-of-00063.safetensors", + "model.layers.59.mlp.gate_proj.weight": "model-00046-of-00063.safetensors", + "model.layers.59.mlp.up_proj.weight": "model-00046-of-00063.safetensors", + "model.layers.59.post_attention_layernorm.weight": "model-00047-of-00063.safetensors", + "model.layers.59.self_attn.k_proj.bias": "model-00046-of-00063.safetensors", + "model.layers.59.self_attn.k_proj.weight": "model-00046-of-00063.safetensors", + "model.layers.59.self_attn.o_proj.weight": "model-00046-of-00063.safetensors", + "model.layers.59.self_attn.q_proj.bias": "model-00046-of-00063.safetensors", + "model.layers.59.self_attn.q_proj.weight": "model-00046-of-00063.safetensors", + "model.layers.59.self_attn.v_proj.bias": "model-00046-of-00063.safetensors", + "model.layers.59.self_attn.v_proj.weight": "model-00046-of-00063.safetensors", + "model.layers.6.input_layernorm.weight": "model-00007-of-00063.safetensors", + "model.layers.6.mlp.down_proj.weight": "model-00007-of-00063.safetensors", + "model.layers.6.mlp.gate_proj.weight": "model-00006-of-00063.safetensors", + "model.layers.6.mlp.up_proj.weight": "model-00007-of-00063.safetensors", + "model.layers.6.post_attention_layernorm.weight": "model-00007-of-00063.safetensors", + "model.layers.6.self_attn.k_proj.bias": "model-00006-of-00063.safetensors", + "model.layers.6.self_attn.k_proj.weight": "model-00006-of-00063.safetensors", + "model.layers.6.self_attn.o_proj.weight": "model-00006-of-00063.safetensors", + "model.layers.6.self_attn.q_proj.bias": "model-00006-of-00063.safetensors", + "model.layers.6.self_attn.q_proj.weight": "model-00006-of-00063.safetensors", + "model.layers.6.self_attn.v_proj.bias": "model-00006-of-00063.safetensors", + "model.layers.6.self_attn.v_proj.weight": "model-00006-of-00063.safetensors", + "model.layers.60.input_layernorm.weight": "model-00047-of-00063.safetensors", + "model.layers.60.mlp.down_proj.weight": "model-00047-of-00063.safetensors", + "model.layers.60.mlp.gate_proj.weight": "model-00047-of-00063.safetensors", + "model.layers.60.mlp.up_proj.weight": "model-00047-of-00063.safetensors", + "model.layers.60.post_attention_layernorm.weight": "model-00047-of-00063.safetensors", + "model.layers.60.self_attn.k_proj.bias": "model-00047-of-00063.safetensors", + "model.layers.60.self_attn.k_proj.weight": "model-00047-of-00063.safetensors", + "model.layers.60.self_attn.o_proj.weight": "model-00047-of-00063.safetensors", + "model.layers.60.self_attn.q_proj.bias": "model-00047-of-00063.safetensors", + "model.layers.60.self_attn.q_proj.weight": "model-00047-of-00063.safetensors", + "model.layers.60.self_attn.v_proj.bias": "model-00047-of-00063.safetensors", + "model.layers.60.self_attn.v_proj.weight": "model-00047-of-00063.safetensors", + "model.layers.61.input_layernorm.weight": "model-00048-of-00063.safetensors", + "model.layers.61.mlp.down_proj.weight": "model-00048-of-00063.safetensors", + "model.layers.61.mlp.gate_proj.weight": "model-00048-of-00063.safetensors", + "model.layers.61.mlp.up_proj.weight": "model-00048-of-00063.safetensors", + "model.layers.61.post_attention_layernorm.weight": "model-00048-of-00063.safetensors", + "model.layers.61.self_attn.k_proj.bias": "model-00047-of-00063.safetensors", + "model.layers.61.self_attn.k_proj.weight": "model-00047-of-00063.safetensors", + "model.layers.61.self_attn.o_proj.weight": "model-00048-of-00063.safetensors", + "model.layers.61.self_attn.q_proj.bias": "model-00047-of-00063.safetensors", + "model.layers.61.self_attn.q_proj.weight": "model-00047-of-00063.safetensors", + "model.layers.61.self_attn.v_proj.bias": "model-00047-of-00063.safetensors", + "model.layers.61.self_attn.v_proj.weight": "model-00047-of-00063.safetensors", + "model.layers.62.input_layernorm.weight": "model-00049-of-00063.safetensors", + "model.layers.62.mlp.down_proj.weight": "model-00049-of-00063.safetensors", + "model.layers.62.mlp.gate_proj.weight": "model-00048-of-00063.safetensors", + "model.layers.62.mlp.up_proj.weight": "model-00049-of-00063.safetensors", + "model.layers.62.post_attention_layernorm.weight": "model-00049-of-00063.safetensors", + "model.layers.62.self_attn.k_proj.bias": "model-00048-of-00063.safetensors", + "model.layers.62.self_attn.k_proj.weight": "model-00048-of-00063.safetensors", + "model.layers.62.self_attn.o_proj.weight": "model-00048-of-00063.safetensors", + "model.layers.62.self_attn.q_proj.bias": "model-00048-of-00063.safetensors", + "model.layers.62.self_attn.q_proj.weight": "model-00048-of-00063.safetensors", + "model.layers.62.self_attn.v_proj.bias": "model-00048-of-00063.safetensors", + "model.layers.62.self_attn.v_proj.weight": "model-00048-of-00063.safetensors", + "model.layers.63.input_layernorm.weight": "model-00050-of-00063.safetensors", + "model.layers.63.mlp.down_proj.weight": "model-00050-of-00063.safetensors", + "model.layers.63.mlp.gate_proj.weight": "model-00049-of-00063.safetensors", + "model.layers.63.mlp.up_proj.weight": "model-00049-of-00063.safetensors", + "model.layers.63.post_attention_layernorm.weight": "model-00050-of-00063.safetensors", + "model.layers.63.self_attn.k_proj.bias": "model-00049-of-00063.safetensors", + "model.layers.63.self_attn.k_proj.weight": "model-00049-of-00063.safetensors", + "model.layers.63.self_attn.o_proj.weight": "model-00049-of-00063.safetensors", + "model.layers.63.self_attn.q_proj.bias": "model-00049-of-00063.safetensors", + "model.layers.63.self_attn.q_proj.weight": "model-00049-of-00063.safetensors", + "model.layers.63.self_attn.v_proj.bias": "model-00049-of-00063.safetensors", + "model.layers.63.self_attn.v_proj.weight": "model-00049-of-00063.safetensors", + "model.layers.64.input_layernorm.weight": "model-00050-of-00063.safetensors", + "model.layers.64.mlp.down_proj.weight": "model-00050-of-00063.safetensors", + "model.layers.64.mlp.gate_proj.weight": "model-00050-of-00063.safetensors", + "model.layers.64.mlp.up_proj.weight": "model-00050-of-00063.safetensors", + "model.layers.64.post_attention_layernorm.weight": "model-00050-of-00063.safetensors", + "model.layers.64.self_attn.k_proj.bias": "model-00050-of-00063.safetensors", + "model.layers.64.self_attn.k_proj.weight": "model-00050-of-00063.safetensors", + "model.layers.64.self_attn.o_proj.weight": "model-00050-of-00063.safetensors", + "model.layers.64.self_attn.q_proj.bias": "model-00050-of-00063.safetensors", + "model.layers.64.self_attn.q_proj.weight": "model-00050-of-00063.safetensors", + "model.layers.64.self_attn.v_proj.bias": "model-00050-of-00063.safetensors", + "model.layers.64.self_attn.v_proj.weight": "model-00050-of-00063.safetensors", + "model.layers.65.input_layernorm.weight": "model-00051-of-00063.safetensors", + "model.layers.65.mlp.down_proj.weight": "model-00051-of-00063.safetensors", + "model.layers.65.mlp.gate_proj.weight": "model-00051-of-00063.safetensors", + "model.layers.65.mlp.up_proj.weight": "model-00051-of-00063.safetensors", + "model.layers.65.post_attention_layernorm.weight": "model-00051-of-00063.safetensors", + "model.layers.65.self_attn.k_proj.bias": "model-00050-of-00063.safetensors", + "model.layers.65.self_attn.k_proj.weight": "model-00050-of-00063.safetensors", + "model.layers.65.self_attn.o_proj.weight": "model-00051-of-00063.safetensors", + "model.layers.65.self_attn.q_proj.bias": "model-00050-of-00063.safetensors", + "model.layers.65.self_attn.q_proj.weight": "model-00050-of-00063.safetensors", + "model.layers.65.self_attn.v_proj.bias": "model-00050-of-00063.safetensors", + "model.layers.65.self_attn.v_proj.weight": "model-00050-of-00063.safetensors", + "model.layers.66.input_layernorm.weight": "model-00052-of-00063.safetensors", + "model.layers.66.mlp.down_proj.weight": "model-00052-of-00063.safetensors", + "model.layers.66.mlp.gate_proj.weight": "model-00051-of-00063.safetensors", + "model.layers.66.mlp.up_proj.weight": "model-00052-of-00063.safetensors", + "model.layers.66.post_attention_layernorm.weight": "model-00052-of-00063.safetensors", + "model.layers.66.self_attn.k_proj.bias": "model-00051-of-00063.safetensors", + "model.layers.66.self_attn.k_proj.weight": "model-00051-of-00063.safetensors", + "model.layers.66.self_attn.o_proj.weight": "model-00051-of-00063.safetensors", + "model.layers.66.self_attn.q_proj.bias": "model-00051-of-00063.safetensors", + "model.layers.66.self_attn.q_proj.weight": "model-00051-of-00063.safetensors", + "model.layers.66.self_attn.v_proj.bias": "model-00051-of-00063.safetensors", + "model.layers.66.self_attn.v_proj.weight": "model-00051-of-00063.safetensors", + "model.layers.67.input_layernorm.weight": "model-00053-of-00063.safetensors", + "model.layers.67.mlp.down_proj.weight": "model-00053-of-00063.safetensors", + "model.layers.67.mlp.gate_proj.weight": "model-00052-of-00063.safetensors", + "model.layers.67.mlp.up_proj.weight": "model-00052-of-00063.safetensors", + "model.layers.67.post_attention_layernorm.weight": "model-00053-of-00063.safetensors", + "model.layers.67.self_attn.k_proj.bias": "model-00052-of-00063.safetensors", + "model.layers.67.self_attn.k_proj.weight": "model-00052-of-00063.safetensors", + "model.layers.67.self_attn.o_proj.weight": "model-00052-of-00063.safetensors", + "model.layers.67.self_attn.q_proj.bias": "model-00052-of-00063.safetensors", + "model.layers.67.self_attn.q_proj.weight": "model-00052-of-00063.safetensors", + "model.layers.67.self_attn.v_proj.bias": "model-00052-of-00063.safetensors", + "model.layers.67.self_attn.v_proj.weight": "model-00052-of-00063.safetensors", + "model.layers.68.input_layernorm.weight": "model-00053-of-00063.safetensors", + "model.layers.68.mlp.down_proj.weight": "model-00053-of-00063.safetensors", + "model.layers.68.mlp.gate_proj.weight": "model-00053-of-00063.safetensors", + "model.layers.68.mlp.up_proj.weight": "model-00053-of-00063.safetensors", + "model.layers.68.post_attention_layernorm.weight": "model-00053-of-00063.safetensors", + "model.layers.68.self_attn.k_proj.bias": "model-00053-of-00063.safetensors", + "model.layers.68.self_attn.k_proj.weight": "model-00053-of-00063.safetensors", + "model.layers.68.self_attn.o_proj.weight": "model-00053-of-00063.safetensors", + "model.layers.68.self_attn.q_proj.bias": "model-00053-of-00063.safetensors", + "model.layers.68.self_attn.q_proj.weight": "model-00053-of-00063.safetensors", + "model.layers.68.self_attn.v_proj.bias": "model-00053-of-00063.safetensors", + "model.layers.68.self_attn.v_proj.weight": "model-00053-of-00063.safetensors", + "model.layers.69.input_layernorm.weight": "model-00054-of-00063.safetensors", + "model.layers.69.mlp.down_proj.weight": "model-00054-of-00063.safetensors", + "model.layers.69.mlp.gate_proj.weight": "model-00054-of-00063.safetensors", + "model.layers.69.mlp.up_proj.weight": "model-00054-of-00063.safetensors", + "model.layers.69.post_attention_layernorm.weight": "model-00054-of-00063.safetensors", + "model.layers.69.self_attn.k_proj.bias": "model-00053-of-00063.safetensors", + "model.layers.69.self_attn.k_proj.weight": "model-00053-of-00063.safetensors", + "model.layers.69.self_attn.o_proj.weight": "model-00054-of-00063.safetensors", + "model.layers.69.self_attn.q_proj.bias": "model-00053-of-00063.safetensors", + "model.layers.69.self_attn.q_proj.weight": "model-00053-of-00063.safetensors", + "model.layers.69.self_attn.v_proj.bias": "model-00053-of-00063.safetensors", + "model.layers.69.self_attn.v_proj.weight": "model-00053-of-00063.safetensors", + "model.layers.7.input_layernorm.weight": "model-00008-of-00063.safetensors", + "model.layers.7.mlp.down_proj.weight": "model-00008-of-00063.safetensors", + "model.layers.7.mlp.gate_proj.weight": "model-00007-of-00063.safetensors", + "model.layers.7.mlp.up_proj.weight": "model-00007-of-00063.safetensors", + "model.layers.7.post_attention_layernorm.weight": "model-00008-of-00063.safetensors", + "model.layers.7.self_attn.k_proj.bias": "model-00007-of-00063.safetensors", + "model.layers.7.self_attn.k_proj.weight": "model-00007-of-00063.safetensors", + "model.layers.7.self_attn.o_proj.weight": "model-00007-of-00063.safetensors", + "model.layers.7.self_attn.q_proj.bias": "model-00007-of-00063.safetensors", + "model.layers.7.self_attn.q_proj.weight": "model-00007-of-00063.safetensors", + "model.layers.7.self_attn.v_proj.bias": "model-00007-of-00063.safetensors", + "model.layers.7.self_attn.v_proj.weight": "model-00007-of-00063.safetensors", + "model.layers.70.input_layernorm.weight": "model-00055-of-00063.safetensors", + "model.layers.70.mlp.down_proj.weight": "model-00055-of-00063.safetensors", + "model.layers.70.mlp.gate_proj.weight": "model-00054-of-00063.safetensors", + "model.layers.70.mlp.up_proj.weight": "model-00055-of-00063.safetensors", + "model.layers.70.post_attention_layernorm.weight": "model-00055-of-00063.safetensors", + "model.layers.70.self_attn.k_proj.bias": "model-00054-of-00063.safetensors", + "model.layers.70.self_attn.k_proj.weight": "model-00054-of-00063.safetensors", + "model.layers.70.self_attn.o_proj.weight": "model-00054-of-00063.safetensors", + "model.layers.70.self_attn.q_proj.bias": "model-00054-of-00063.safetensors", + "model.layers.70.self_attn.q_proj.weight": "model-00054-of-00063.safetensors", + "model.layers.70.self_attn.v_proj.bias": "model-00054-of-00063.safetensors", + "model.layers.70.self_attn.v_proj.weight": "model-00054-of-00063.safetensors", + "model.layers.71.input_layernorm.weight": "model-00056-of-00063.safetensors", + "model.layers.71.mlp.down_proj.weight": "model-00056-of-00063.safetensors", + "model.layers.71.mlp.gate_proj.weight": "model-00055-of-00063.safetensors", + "model.layers.71.mlp.up_proj.weight": "model-00055-of-00063.safetensors", + "model.layers.71.post_attention_layernorm.weight": "model-00056-of-00063.safetensors", + "model.layers.71.self_attn.k_proj.bias": "model-00055-of-00063.safetensors", + "model.layers.71.self_attn.k_proj.weight": "model-00055-of-00063.safetensors", + "model.layers.71.self_attn.o_proj.weight": "model-00055-of-00063.safetensors", + "model.layers.71.self_attn.q_proj.bias": "model-00055-of-00063.safetensors", + "model.layers.71.self_attn.q_proj.weight": "model-00055-of-00063.safetensors", + "model.layers.71.self_attn.v_proj.bias": "model-00055-of-00063.safetensors", + "model.layers.71.self_attn.v_proj.weight": "model-00055-of-00063.safetensors", + "model.layers.72.input_layernorm.weight": "model-00056-of-00063.safetensors", + "model.layers.72.mlp.down_proj.weight": "model-00056-of-00063.safetensors", + "model.layers.72.mlp.gate_proj.weight": "model-00056-of-00063.safetensors", + "model.layers.72.mlp.up_proj.weight": "model-00056-of-00063.safetensors", + "model.layers.72.post_attention_layernorm.weight": "model-00056-of-00063.safetensors", + "model.layers.72.self_attn.k_proj.bias": "model-00056-of-00063.safetensors", + "model.layers.72.self_attn.k_proj.weight": "model-00056-of-00063.safetensors", + "model.layers.72.self_attn.o_proj.weight": "model-00056-of-00063.safetensors", + "model.layers.72.self_attn.q_proj.bias": "model-00056-of-00063.safetensors", + "model.layers.72.self_attn.q_proj.weight": "model-00056-of-00063.safetensors", + "model.layers.72.self_attn.v_proj.bias": "model-00056-of-00063.safetensors", + "model.layers.72.self_attn.v_proj.weight": "model-00056-of-00063.safetensors", + "model.layers.73.input_layernorm.weight": "model-00057-of-00063.safetensors", + "model.layers.73.mlp.down_proj.weight": "model-00057-of-00063.safetensors", + "model.layers.73.mlp.gate_proj.weight": "model-00057-of-00063.safetensors", + "model.layers.73.mlp.up_proj.weight": "model-00057-of-00063.safetensors", + "model.layers.73.post_attention_layernorm.weight": "model-00057-of-00063.safetensors", + "model.layers.73.self_attn.k_proj.bias": "model-00056-of-00063.safetensors", + "model.layers.73.self_attn.k_proj.weight": "model-00056-of-00063.safetensors", + "model.layers.73.self_attn.o_proj.weight": "model-00057-of-00063.safetensors", + "model.layers.73.self_attn.q_proj.bias": "model-00056-of-00063.safetensors", + "model.layers.73.self_attn.q_proj.weight": "model-00056-of-00063.safetensors", + "model.layers.73.self_attn.v_proj.bias": "model-00056-of-00063.safetensors", + "model.layers.73.self_attn.v_proj.weight": "model-00056-of-00063.safetensors", + "model.layers.74.input_layernorm.weight": "model-00058-of-00063.safetensors", + "model.layers.74.mlp.down_proj.weight": "model-00058-of-00063.safetensors", + "model.layers.74.mlp.gate_proj.weight": "model-00057-of-00063.safetensors", + "model.layers.74.mlp.up_proj.weight": "model-00058-of-00063.safetensors", + "model.layers.74.post_attention_layernorm.weight": "model-00058-of-00063.safetensors", + "model.layers.74.self_attn.k_proj.bias": "model-00057-of-00063.safetensors", + "model.layers.74.self_attn.k_proj.weight": "model-00057-of-00063.safetensors", + "model.layers.74.self_attn.o_proj.weight": "model-00057-of-00063.safetensors", + "model.layers.74.self_attn.q_proj.bias": "model-00057-of-00063.safetensors", + "model.layers.74.self_attn.q_proj.weight": "model-00057-of-00063.safetensors", + "model.layers.74.self_attn.v_proj.bias": "model-00057-of-00063.safetensors", + "model.layers.74.self_attn.v_proj.weight": "model-00057-of-00063.safetensors", + "model.layers.75.input_layernorm.weight": "model-00059-of-00063.safetensors", + "model.layers.75.mlp.down_proj.weight": "model-00059-of-00063.safetensors", + "model.layers.75.mlp.gate_proj.weight": "model-00058-of-00063.safetensors", + "model.layers.75.mlp.up_proj.weight": "model-00058-of-00063.safetensors", + "model.layers.75.post_attention_layernorm.weight": "model-00059-of-00063.safetensors", + "model.layers.75.self_attn.k_proj.bias": "model-00058-of-00063.safetensors", + "model.layers.75.self_attn.k_proj.weight": "model-00058-of-00063.safetensors", + "model.layers.75.self_attn.o_proj.weight": "model-00058-of-00063.safetensors", + "model.layers.75.self_attn.q_proj.bias": "model-00058-of-00063.safetensors", + "model.layers.75.self_attn.q_proj.weight": "model-00058-of-00063.safetensors", + "model.layers.75.self_attn.v_proj.bias": "model-00058-of-00063.safetensors", + "model.layers.75.self_attn.v_proj.weight": "model-00058-of-00063.safetensors", + "model.layers.76.input_layernorm.weight": "model-00059-of-00063.safetensors", + "model.layers.76.mlp.down_proj.weight": "model-00059-of-00063.safetensors", + "model.layers.76.mlp.gate_proj.weight": "model-00059-of-00063.safetensors", + "model.layers.76.mlp.up_proj.weight": "model-00059-of-00063.safetensors", + "model.layers.76.post_attention_layernorm.weight": "model-00059-of-00063.safetensors", + "model.layers.76.self_attn.k_proj.bias": "model-00059-of-00063.safetensors", + "model.layers.76.self_attn.k_proj.weight": "model-00059-of-00063.safetensors", + "model.layers.76.self_attn.o_proj.weight": "model-00059-of-00063.safetensors", + "model.layers.76.self_attn.q_proj.bias": "model-00059-of-00063.safetensors", + "model.layers.76.self_attn.q_proj.weight": "model-00059-of-00063.safetensors", + "model.layers.76.self_attn.v_proj.bias": "model-00059-of-00063.safetensors", + "model.layers.76.self_attn.v_proj.weight": "model-00059-of-00063.safetensors", + "model.layers.77.input_layernorm.weight": "model-00060-of-00063.safetensors", + "model.layers.77.mlp.down_proj.weight": "model-00060-of-00063.safetensors", + "model.layers.77.mlp.gate_proj.weight": "model-00060-of-00063.safetensors", + "model.layers.77.mlp.up_proj.weight": "model-00060-of-00063.safetensors", + "model.layers.77.post_attention_layernorm.weight": "model-00060-of-00063.safetensors", + "model.layers.77.self_attn.k_proj.bias": "model-00059-of-00063.safetensors", + "model.layers.77.self_attn.k_proj.weight": "model-00059-of-00063.safetensors", + "model.layers.77.self_attn.o_proj.weight": "model-00060-of-00063.safetensors", + "model.layers.77.self_attn.q_proj.bias": "model-00059-of-00063.safetensors", + "model.layers.77.self_attn.q_proj.weight": "model-00059-of-00063.safetensors", + "model.layers.77.self_attn.v_proj.bias": "model-00059-of-00063.safetensors", + "model.layers.77.self_attn.v_proj.weight": "model-00059-of-00063.safetensors", + "model.layers.78.input_layernorm.weight": "model-00061-of-00063.safetensors", + "model.layers.78.mlp.down_proj.weight": "model-00061-of-00063.safetensors", + "model.layers.78.mlp.gate_proj.weight": "model-00060-of-00063.safetensors", + "model.layers.78.mlp.up_proj.weight": "model-00061-of-00063.safetensors", + "model.layers.78.post_attention_layernorm.weight": "model-00061-of-00063.safetensors", + "model.layers.78.self_attn.k_proj.bias": "model-00060-of-00063.safetensors", + "model.layers.78.self_attn.k_proj.weight": "model-00060-of-00063.safetensors", + "model.layers.78.self_attn.o_proj.weight": "model-00060-of-00063.safetensors", + "model.layers.78.self_attn.q_proj.bias": "model-00060-of-00063.safetensors", + "model.layers.78.self_attn.q_proj.weight": "model-00060-of-00063.safetensors", + "model.layers.78.self_attn.v_proj.bias": "model-00060-of-00063.safetensors", + "model.layers.78.self_attn.v_proj.weight": "model-00060-of-00063.safetensors", + "model.layers.79.input_layernorm.weight": "model-00062-of-00063.safetensors", + "model.layers.79.mlp.down_proj.weight": "model-00062-of-00063.safetensors", + "model.layers.79.mlp.gate_proj.weight": "model-00061-of-00063.safetensors", + "model.layers.79.mlp.up_proj.weight": "model-00061-of-00063.safetensors", + "model.layers.79.post_attention_layernorm.weight": "model-00062-of-00063.safetensors", + "model.layers.79.self_attn.k_proj.bias": "model-00061-of-00063.safetensors", + "model.layers.79.self_attn.k_proj.weight": "model-00061-of-00063.safetensors", + "model.layers.79.self_attn.o_proj.weight": "model-00061-of-00063.safetensors", + "model.layers.79.self_attn.q_proj.bias": "model-00061-of-00063.safetensors", + "model.layers.79.self_attn.q_proj.weight": "model-00061-of-00063.safetensors", + "model.layers.79.self_attn.v_proj.bias": "model-00061-of-00063.safetensors", + "model.layers.79.self_attn.v_proj.weight": "model-00061-of-00063.safetensors", + "model.layers.8.input_layernorm.weight": "model-00008-of-00063.safetensors", + "model.layers.8.mlp.down_proj.weight": "model-00008-of-00063.safetensors", + "model.layers.8.mlp.gate_proj.weight": "model-00008-of-00063.safetensors", + "model.layers.8.mlp.up_proj.weight": "model-00008-of-00063.safetensors", + "model.layers.8.post_attention_layernorm.weight": "model-00008-of-00063.safetensors", + "model.layers.8.self_attn.k_proj.bias": "model-00008-of-00063.safetensors", + "model.layers.8.self_attn.k_proj.weight": "model-00008-of-00063.safetensors", + "model.layers.8.self_attn.o_proj.weight": "model-00008-of-00063.safetensors", + "model.layers.8.self_attn.q_proj.bias": "model-00008-of-00063.safetensors", + "model.layers.8.self_attn.q_proj.weight": "model-00008-of-00063.safetensors", + "model.layers.8.self_attn.v_proj.bias": "model-00008-of-00063.safetensors", + "model.layers.8.self_attn.v_proj.weight": "model-00008-of-00063.safetensors", + "model.layers.9.input_layernorm.weight": "model-00009-of-00063.safetensors", + "model.layers.9.mlp.down_proj.weight": "model-00009-of-00063.safetensors", + "model.layers.9.mlp.gate_proj.weight": "model-00009-of-00063.safetensors", + "model.layers.9.mlp.up_proj.weight": "model-00009-of-00063.safetensors", + "model.layers.9.post_attention_layernorm.weight": "model-00009-of-00063.safetensors", + "model.layers.9.self_attn.k_proj.bias": "model-00008-of-00063.safetensors", + "model.layers.9.self_attn.k_proj.weight": "model-00008-of-00063.safetensors", + "model.layers.9.self_attn.o_proj.weight": "model-00009-of-00063.safetensors", + "model.layers.9.self_attn.q_proj.bias": "model-00008-of-00063.safetensors", + "model.layers.9.self_attn.q_proj.weight": "model-00008-of-00063.safetensors", + "model.layers.9.self_attn.v_proj.bias": "model-00008-of-00063.safetensors", + "model.layers.9.self_attn.v_proj.weight": "model-00008-of-00063.safetensors", + "model.norm.weight": "model-00062-of-00063.safetensors" + } +}