sandiago21 commited on
Commit
a82427a
1 Parent(s): f42fd22

committing original decapoda-research/llama-13b-hf model

Browse files
This view is limited to 50 files because it contains too many changes.   See raw diff
Files changed (50) hide show
  1. .gitattributes +44 -0
  2. LICENSE.txt +76 -0
  3. config.json +1 -0
  4. generation_config.json +1 -0
  5. gpt-training-Trainer-conversations-Inference.ipynb +0 -0
  6. huggingface_notebook.ipynb +68 -0
  7. pytorch_model-00000-of-00041.bin +3 -0
  8. pytorch_model-00001-of-00041.bin +3 -0
  9. pytorch_model-00002-of-00041.bin +3 -0
  10. pytorch_model-00003-of-00041.bin +3 -0
  11. pytorch_model-00004-of-00041.bin +3 -0
  12. pytorch_model-00005-of-00041.bin +3 -0
  13. pytorch_model-00006-of-00041.bin +3 -0
  14. pytorch_model-00007-of-00041.bin +3 -0
  15. pytorch_model-00008-of-00041.bin +3 -0
  16. pytorch_model-00009-of-00041.bin +3 -0
  17. pytorch_model-00010-of-00041.bin +3 -0
  18. pytorch_model-00011-of-00041.bin +3 -0
  19. pytorch_model-00012-of-00041.bin +3 -0
  20. pytorch_model-00013-of-00041.bin +3 -0
  21. pytorch_model-00014-of-00041.bin +3 -0
  22. pytorch_model-00015-of-00041.bin +3 -0
  23. pytorch_model-00016-of-00041.bin +3 -0
  24. pytorch_model-00017-of-00041.bin +3 -0
  25. pytorch_model-00018-of-00041.bin +3 -0
  26. pytorch_model-00019-of-00041.bin +3 -0
  27. pytorch_model-00020-of-00041.bin +3 -0
  28. pytorch_model-00021-of-00041.bin +3 -0
  29. pytorch_model-00022-of-00041.bin +3 -0
  30. pytorch_model-00023-of-00041.bin +3 -0
  31. pytorch_model-00024-of-00041.bin +3 -0
  32. pytorch_model-00025-of-00041.bin +3 -0
  33. pytorch_model-00026-of-00041.bin +3 -0
  34. pytorch_model-00027-of-00041.bin +3 -0
  35. pytorch_model-00028-of-00041.bin +3 -0
  36. pytorch_model-00029-of-00041.bin +3 -0
  37. pytorch_model-00030-of-00041.bin +3 -0
  38. pytorch_model-00031-of-00041.bin +3 -0
  39. pytorch_model-00032-of-00041.bin +3 -0
  40. pytorch_model-00033-of-00041.bin +3 -0
  41. pytorch_model-00034-of-00041.bin +3 -0
  42. pytorch_model-00035-of-00041.bin +3 -0
  43. pytorch_model-00036-of-00041.bin +3 -0
  44. pytorch_model-00037-of-00041.bin +3 -0
  45. pytorch_model-00038-of-00041.bin +3 -0
  46. pytorch_model-00039-of-00041.bin +3 -0
  47. pytorch_model-00040-of-00041.bin +3 -0
  48. pytorch_model-00041-of-00041.bin +3 -0
  49. pytorch_model.bin.index.json +1 -0
  50. special_tokens_map.json +1 -0
.gitattributes CHANGED
@@ -32,3 +32,47 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
32
  *.zip filter=lfs diff=lfs merge=lfs -text
33
  *.zst filter=lfs diff=lfs merge=lfs -text
34
  *tfevents* filter=lfs diff=lfs merge=lfs -text
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
32
  *.zip filter=lfs diff=lfs merge=lfs -text
33
  *.zst filter=lfs diff=lfs merge=lfs -text
34
  *tfevents* filter=lfs diff=lfs merge=lfs -text
35
+ finetuned_models/decapoda_research/llama-13b-hf/adapter_model.bin filter=lfs diff=lfs merge=lfs -text
36
+ adapter_model.bin filter=lfs diff=lfs merge=lfs -text
37
+ pytorch_model-00011-of-00041.bin filter=lfs diff=lfs merge=lfs -text
38
+ pytorch_model-00018-of-00041.bin filter=lfs diff=lfs merge=lfs -text
39
+ pytorch_model-00036-of-00041.bin filter=lfs diff=lfs merge=lfs -text
40
+ pytorch_model-00023-of-00041.bin filter=lfs diff=lfs merge=lfs -text
41
+ pytorch_model-00024-of-00041.bin filter=lfs diff=lfs merge=lfs -text
42
+ pytorch_model-00041-of-00041.bin filter=lfs diff=lfs merge=lfs -text
43
+ pytorch_model-00008-of-00041.bin filter=lfs diff=lfs merge=lfs -text
44
+ pytorch_model-00025-of-00041.bin filter=lfs diff=lfs merge=lfs -text
45
+ pytorch_model-00030-of-00041.bin filter=lfs diff=lfs merge=lfs -text
46
+ pytorch_model-00005-of-00041.bin filter=lfs diff=lfs merge=lfs -text
47
+ pytorch_model-00012-of-00041.bin filter=lfs diff=lfs merge=lfs -text
48
+ pytorch_model-00016-of-00041.bin filter=lfs diff=lfs merge=lfs -text
49
+ pytorch_model-00028-of-00041.bin filter=lfs diff=lfs merge=lfs -text
50
+ pytorch_model-00032-of-00041.bin filter=lfs diff=lfs merge=lfs -text
51
+ pytorch_model-00037-of-00041.bin filter=lfs diff=lfs merge=lfs -text
52
+ pytorch_model-00014-of-00041.bin filter=lfs diff=lfs merge=lfs -text
53
+ pytorch_model-00021-of-00041.bin filter=lfs diff=lfs merge=lfs -text
54
+ pytorch_model-00039-of-00041.bin filter=lfs diff=lfs merge=lfs -text
55
+ pytorch_model-00000-of-00041.bin filter=lfs diff=lfs merge=lfs -text
56
+ pytorch_model-00002-of-00041.bin filter=lfs diff=lfs merge=lfs -text
57
+ pytorch_model-00013-of-00041.bin filter=lfs diff=lfs merge=lfs -text
58
+ pytorch_model-00015-of-00041.bin filter=lfs diff=lfs merge=lfs -text
59
+ pytorch_model-00035-of-00041.bin filter=lfs diff=lfs merge=lfs -text
60
+ pytorch_model-00038-of-00041.bin filter=lfs diff=lfs merge=lfs -text
61
+ pytorch_model-00026-of-00041.bin filter=lfs diff=lfs merge=lfs -text
62
+ pytorch_model-00027-of-00041.bin filter=lfs diff=lfs merge=lfs -text
63
+ pytorch_model-00001-of-00041.bin filter=lfs diff=lfs merge=lfs -text
64
+ pytorch_model-00009-of-00041.bin filter=lfs diff=lfs merge=lfs -text
65
+ pytorch_model-00010-of-00041.bin filter=lfs diff=lfs merge=lfs -text
66
+ pytorch_model-00017-of-00041.bin filter=lfs diff=lfs merge=lfs -text
67
+ pytorch_model-00020-of-00041.bin filter=lfs diff=lfs merge=lfs -text
68
+ pytorch_model-00022-of-00041.bin filter=lfs diff=lfs merge=lfs -text
69
+ pytorch_model-00031-of-00041.bin filter=lfs diff=lfs merge=lfs -text
70
+ pytorch_model-00033-of-00041.bin filter=lfs diff=lfs merge=lfs -text
71
+ pytorch_model-00034-of-00041.bin filter=lfs diff=lfs merge=lfs -text
72
+ pytorch_model-00040-of-00041.bin filter=lfs diff=lfs merge=lfs -text
73
+ pytorch_model-00003-of-00041.bin filter=lfs diff=lfs merge=lfs -text
74
+ pytorch_model-00004-of-00041.bin filter=lfs diff=lfs merge=lfs -text
75
+ pytorch_model-00006-of-00041.bin filter=lfs diff=lfs merge=lfs -text
76
+ pytorch_model-00007-of-00041.bin filter=lfs diff=lfs merge=lfs -text
77
+ pytorch_model-00019-of-00041.bin filter=lfs diff=lfs merge=lfs -text
78
+ pytorch_model-00029-of-00041.bin filter=lfs diff=lfs merge=lfs -text
LICENSE.txt ADDED
@@ -0,0 +1,76 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ LLaMA LICENSE AGREEMENT
2
+ This License Agreement (as may be amended in accordance with this License Agreement, “License”), between you, or your employer or other entity (if you are entering into this agreement on behalf of your employer or other entity) (“Licensee” or “you”) and Meta Platforms, Inc. (“Meta” or “we”) applies to your use of any computer program, algorithm, source code, object code, or software that is made available by Meta under this License (“Software”) and any specifications, manuals, documentation, and other written information provided by Meta related to the Software (“Documentation”).
3
+
4
+ By clicking “I Accept” below or by using the Software, you agree to the terms of this License. If you do not agree to this License, then you do not have any rights to use the Software or Documentation (collectively, the “Software Products”), and you must immediately cease using the Software Products. If you are agreeing to be bound by the terms of this License on behalf of your employer or other entity, you represent and warrant to Meta that you have full legal authority to bind your employer or such entity to this License. If you do not have the requisite authority, you may not accept the License or access the Software Products on behalf of your employer or other entity.
5
+
6
+
7
+
8
+ LICENSE GRANT
9
+
10
+ a. Subject to your compliance with the Documentation and Sections 2, 3, and 5, Meta grants you a non-exclusive, worldwide, non-transferable, non-sublicensable, revocable, royalty free and limited license under Meta’s copyright interests to reproduce, distribute, and create derivative works of the Software solely for your non-commercial research purposes. The foregoing license is personal to you, and you may not assign or sublicense this License or any other rights or obligations under this License without Meta’s prior written consent; any such assignment or sublicense will be void and will automatically and immediately terminate this License.
11
+
12
+ b. You may make a reasonable number of copies of the Documentation solely for use in connection with the license to the Software granted above.
13
+
14
+ c. The grant of rights expressly set forth in this Section 1 (License Grant) are the complete grant of rights to you in the Software Products, and no other licenses are granted, whether by waiver, estoppel, implication, equity or otherwise. Meta and its licensors reserve all rights not expressly granted by this License.
15
+
16
+
17
+ RESTRICTIONS
18
+
19
+ You will not, and will not permit, assist or cause any third party to:
20
+
21
+ a. use, modify, copy, reproduce, create derivative works of, or distribute the Software Products (or any derivative works thereof, works incorporating the Software Products, or any data produced by the Software), in whole or in part, for (i) any commercial or production purposes, (ii) military purposes or in the service of nuclear technology, (iii) purposes of surveillance, including any research or development relating to surveillance, (iv) biometric processing, (v) in any manner that infringes, misappropriates, or otherwise violates any third-party rights, or (vi) in any manner that violates any applicable law, including accessing the Software Products from an embargoed country as prohibited by the U.S. government, and violating any privacy or security laws, rules, regulations, directives, or governmental requirements (including the General Data Privacy Regulation (Regulation (EU) 2016/679), the California Consumer Privacy Act, and any and all laws governing the processing of biometric information), as well as all amendments and successor laws to any of the foregoing;
22
+
23
+ b. alter or remove copyright and other proprietary notices which appear on or in the Software Products;
24
+
25
+ c. utilize any equipment, device, software, or other means to circumvent or remove any security or protection used by Meta in connection with the Software, or to circumvent or remove any usage restrictions, or to enable functionality disabled by Meta; or
26
+
27
+ d. offer or impose any terms on the Software Products that alter, restrict, or are inconsistent with the terms of this License.
28
+
29
+
30
+ ATTRIBUTION
31
+
32
+ Together with any copies of the Software Products (as well as derivative works thereof or works incorporating the Software Products) that you distribute, you must provide (i) a copy of this License, and (ii) the following attribution notice: “LLaMA is licensed under the LLaMA license, Copyright (c) Meta Platforms, Inc. All Rights Reserved.”
33
+
34
+
35
+ DISCLAIMERS
36
+
37
+ THE SOFTWARE PRODUCTS ARE PROVIDED “AS IS” and “WITH ALL FAULTS” WITH NO WARRANTY OF ANY KIND, EXPRESS OR IMPLIED. META EXPRESSLY DISCLAIMS ALL REPRESENTATIONS AND WARRANTIES, EXPRESS OR IMPLIED, WHETHER BY STATUTE, CUSTOM, USAGE OR OTHERWISE AS TO ANY MATTERS RELATED TO THE SOFTWARE PRODUCTS, INCLUDING BUT NOT LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE, TITLE, SATISFACTORY QUALITY, OR NON-INFRINGEMENT. META MAKES NO WARRANTIES OR REPRESENTATIONS THAT THE SOFTWARE PRODUCTS WILL BE ERROR FREE OR FREE OF VIRUSES OR OTHER HARMFUL COMPONENTS, OR PRODUCE ANY PARTICULAR RESULTS.
38
+
39
+
40
+ LIMITATION OF LIABILITY
41
+
42
+ TO THE FULLEST EXTENT PERMITTED BY LAW, IN NO EVENT WILL META BE LIABLE TO YOU (A) UNDER ANY THEORY OF LIABILITY, WHETHER BASED IN CONTRACT, TORT, NEGLIGENCE, STRICT LIABILITY, WARRANTY, OR OTHERWISE UNDER THIS LICENSE, OR (B) FOR ANY INDIRECT, CONSEQUENTIAL, EXEMPLARY, INCIDENTAL, PUNITIVE OR SPECIAL DAMAGES OR LOST PROFITS, EVEN IF META HAS BEEN ADVISED OF THE POSSIBILITY OF SUCH DAMAGES. THE SOFTWARE PRODUCTS, THEIR CONSTITUENT COMPONENTS, AND ANY OUTPUT (COLLECTIVELY, “SOFTWARE MATERIALS”) ARE NOT DESIGNED OR INTENDED FOR USE IN ANY APPLICATION OR SITUATION WHERE FAILURE OR FAULT OF THE SOFTWARE MATERIALS COULD REASONABLY BE ANTICIPATED TO LEAD TO SERIOUS INJURY OF ANY PERSON, INCLUDING POTENTIAL DISCRIMINATION OR VIOLATION OF AN INDIVIDUAL’S PRIVACY RIGHTS, OR TO SEVERE PHYSICAL, PROPERTY, OR ENVIRONMENTAL DAMAGE (EACH, A “HIGH-RISK USE”). IF YOU ELECT TO USE ANY OF THE SOFTWARE MATERIALS FOR A HIGH-RISK USE, YOU DO SO AT YOUR OWN RISK. YOU AGREE TO DESIGN AND IMPLEMENT APPROPRIATE DECISION-MAKING AND RISK-MITIGATION PROCEDURES AND POLICIES IN CONNECTION WITH A HIGH-RISK USE SUCH THAT EVEN IF THERE IS A FAILURE OR FAULT IN ANY OF THE SOFTWARE MATERIALS, THE SAFETY OF PERSONS OR PROPERTY AFFECTED BY THE ACTIVITY STAYS AT A LEVEL THAT IS REASONABLE, APPROPRIATE, AND LAWFUL FOR THE FIELD OF THE HIGH-RISK USE.
43
+
44
+
45
+ INDEMNIFICATION
46
+
47
+ You will indemnify, defend and hold harmless Meta and our subsidiaries and affiliates, and each of our respective shareholders, directors, officers, employees, agents, successors, and assigns (collectively, the “Meta Parties”) from and against any losses, liabilities, damages, fines, penalties, and expenses (including reasonable attorneys’ fees) incurred by any Meta Party in connection with any claim, demand, allegation, lawsuit, proceeding, or investigation (collectively, “Claims”) arising out of or related to: (a) your access to or use of the Software Products (as well as any results or data generated from such access or use), including any High-Risk Use (defined below); (b) your violation of this License; or (c) your violation, misappropriation or infringement of any rights of another (including intellectual property or other proprietary rights and privacy rights). You will promptly notify the Meta Parties of any such Claims, and cooperate with Meta Parties in defending such Claims. You will also grant the Meta Parties sole control of the defense or settlement, at Meta’s sole option, of any Claims. This indemnity is in addition to, and not in lieu of, any other indemnities or remedies set forth in a written agreement between you and Meta or the other Meta Parties.
48
+
49
+
50
+ TERMINATION; SURVIVAL
51
+
52
+ a. This License will automatically terminate upon any breach by you of the terms of this License.
53
+
54
+ b. We may terminate this License, in whole or in part, at any time upon notice (including electronic) to you.
55
+
56
+ c. The following sections survive termination of this License: 2 (Restrictions), 3 (Attribution), 4 (Disclaimers), 5 (Limitation on Liability), 6 (Indemnification) 7 (Termination; Survival), 8 (Third Party Materials), 9 (Trademarks), 10 (Applicable Law; Dispute Resolution), and 11 (Miscellaneous).
57
+
58
+
59
+ THIRD PARTY MATERIALS
60
+
61
+ The Software Products may contain third-party software or other components (including free and open source software) (all of the foregoing, “Third Party Materials”), which are subject to the license terms of the respective third-party licensors. Your dealings or correspondence with third parties and your use of or interaction with any Third Party Materials are solely between you and the third party. Meta does not control or endorse, and makes no representations or warranties regarding, any Third Party Materials, and your access to and use of such Third Party Materials are at your own risk.
62
+
63
+
64
+ TRADEMARKS
65
+
66
+ Licensee has not been granted any trademark license as part of this License and may not use any name or mark associated with Meta without the prior written permission of Meta, except to the extent necessary to make the reference required by the “ATTRIBUTION” section of this Agreement.
67
+
68
+
69
+ APPLICABLE LAW; DISPUTE RESOLUTION
70
+
71
+ This License will be governed and construed under the laws of the State of California without regard to conflicts of law provisions. Any suit or proceeding arising out of or relating to this License will be brought in the federal or state courts, as applicable, in San Mateo County, California, and each party irrevocably submits to the jurisdiction and venue of such courts.
72
+
73
+
74
+ MISCELLANEOUS
75
+
76
+ If any provision or part of a provision of this License is unlawful, void or unenforceable, that provision or part of the provision is deemed severed from this License, and will not affect the validity and enforceability of any remaining provisions. The failure of Meta to exercise or enforce any right or provision of this License will not operate as a waiver of such right or provision. This License does not confer any third-party beneficiary rights upon any other person or entity. This License, together with the Documentation, contains the entire understanding between you and Meta regarding the subject matter of this License, and supersedes all other written or oral agreements and understandings between you and Meta regarding such subject matter. No change or addition to any provision of this License will be binding unless it is in writing and signed by an authorized representative of both you and Meta.
config.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"architectures": ["LLaMAForCausalLM"], "bos_token_id": 0, "eos_token_id": 1, "hidden_act": "silu", "hidden_size": 5120, "intermediate_size": 13824, "initializer_range": 0.02, "max_sequence_length": 2048, "model_type": "llama", "num_attention_heads": 40, "num_hidden_layers": 40, "pad_token_id": -1, "rms_norm_eps": 1e-06, "torch_dtype": "float16", "transformers_version": "4.27.0.dev0", "use_cache": true, "vocab_size": 32000}
generation_config.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"_from_model_config": true, "bos_token_id": 0, "eos_token_id": 1, "pad_token_id": 0, "transformers_version": "4.27.0.dev0"}
gpt-training-Trainer-conversations-Inference.ipynb ADDED
The diff for this file is too large to render. See raw diff
 
huggingface_notebook.ipynb ADDED
@@ -0,0 +1,68 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "cells": [
3
+ {
4
+ "cell_type": "code",
5
+ "execution_count": 1,
6
+ "id": "ebc5bd34",
7
+ "metadata": {},
8
+ "outputs": [],
9
+ "source": [
10
+ "from huggingface_hub import notebook_login"
11
+ ]
12
+ },
13
+ {
14
+ "cell_type": "code",
15
+ "execution_count": 2,
16
+ "id": "0b65fd16",
17
+ "metadata": {},
18
+ "outputs": [
19
+ {
20
+ "data": {
21
+ "application/vnd.jupyter.widget-view+json": {
22
+ "model_id": "bc3f54bfcc314a5bb12197f36e82a747",
23
+ "version_major": 2,
24
+ "version_minor": 0
25
+ },
26
+ "text/plain": [
27
+ "VBox(children=(HTML(value='<center> <img\\nsrc=https://huggingface.co/front/assets/huggingface_logo-noborder.sv…"
28
+ ]
29
+ },
30
+ "metadata": {},
31
+ "output_type": "display_data"
32
+ }
33
+ ],
34
+ "source": [
35
+ "notebook_login()"
36
+ ]
37
+ },
38
+ {
39
+ "cell_type": "code",
40
+ "execution_count": null,
41
+ "id": "960257a9",
42
+ "metadata": {},
43
+ "outputs": [],
44
+ "source": []
45
+ }
46
+ ],
47
+ "metadata": {
48
+ "kernelspec": {
49
+ "display_name": "Python 3 (ipykernel)",
50
+ "language": "python",
51
+ "name": "python3"
52
+ },
53
+ "language_info": {
54
+ "codemirror_mode": {
55
+ "name": "ipython",
56
+ "version": 3
57
+ },
58
+ "file_extension": ".py",
59
+ "mimetype": "text/x-python",
60
+ "name": "python",
61
+ "nbconvert_exporter": "python",
62
+ "pygments_lexer": "ipython3",
63
+ "version": "3.8.16"
64
+ }
65
+ },
66
+ "nbformat": 4,
67
+ "nbformat_minor": 5
68
+ }
pytorch_model-00000-of-00041.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6a78c977ed7f090a13703f5d9f69034c0ae405462f3b4302b42b7e0633bc02eb
3
+ size 951607026
pytorch_model-00001-of-00041.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c4f647c10dbb6aeb24cbf73af5527515b3c7ac7101a7301b09113506aa81bb20
3
+ size 951607026
pytorch_model-00002-of-00041.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a25f7e8a4459e98f42a34eefbba27996e6577469f01c659d77caadf9d77a838c
3
+ size 951607026
pytorch_model-00003-of-00041.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:25a8d928e11d1fc54ee829db98b72b08aabde986bb51fcd17bfe56511b0e4af9
3
+ size 951607026
pytorch_model-00004-of-00041.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1b9bb0f7433f7bddae99f0bc2436a752a0d21b8578f152da303d9278773503cd
3
+ size 951607026
pytorch_model-00005-of-00041.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a43bf6ab505435e4c34fcfceb185f47186dbb8a9d548e59e35835e6cbc6ee900
3
+ size 951607026
pytorch_model-00006-of-00041.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0d7506b1ec596c27855b2b256222b02754c31cd4a5e1fb4ae6f10568ed094163
3
+ size 951607026
pytorch_model-00007-of-00041.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e08679b387a3b402833e520c2f0551d4635d9340180cc41bc6bf33fcb6075767
3
+ size 951607026
pytorch_model-00008-of-00041.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b3e40e49eb70584e7a966f4032f088598209da81bb2e84282c85db69b30848ca
3
+ size 951607026
pytorch_model-00009-of-00041.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8c67dcda52cc13ee5255e3d888fa1bd6208e15705b86714eeb0dd671831c5dd2
3
+ size 951607026
pytorch_model-00010-of-00041.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:87637aeb35f554a41e3d16f7e5053f6d606fd2e797d008afae6fbede99faa6b4
3
+ size 951607026
pytorch_model-00011-of-00041.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6b9399c878c036fe69b012dad6b4d762bd5897bd90beeb910e24b7229d58c423
3
+ size 951607026
pytorch_model-00012-of-00041.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d74f25a944504be4abfac1bc4707258d66f2d46447eb332ebeee898451a13f6a
3
+ size 951607026
pytorch_model-00013-of-00041.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ec20f0543c04a0591aaf8dec77c01b91f103e5be09a8043431acf75fdb702cc3
3
+ size 951607026
pytorch_model-00014-of-00041.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1e448e96a139c901fab5bda37512241619cfd11e8872ca61e4c262e9101ecede
3
+ size 951607026
pytorch_model-00015-of-00041.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:44a9e6b066f52111e7bf9610ca11cd03bc30a2cc2e22602c1c295c4955f613a3
3
+ size 951607026
pytorch_model-00016-of-00041.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:eb00d18ae8e1180968ffebd3d0694deb8317d8d66f30354577cd8687bdb20d19
3
+ size 951607026
pytorch_model-00017-of-00041.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5aca85ed3e6bf8306e82daf5a3cb9b30c6fb198dfc8624e1b6914ddb5def466a
3
+ size 951607026
pytorch_model-00018-of-00041.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f82a8894bc807cf1cf48e83a5495f014d1bc6d7521596b2f62e44e0724a2cbc3
3
+ size 951607026
pytorch_model-00019-of-00041.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:21299e2588f945f08b878806d4938c9afaffb74d88bb6fc5bb135274f33af115
3
+ size 951607026
pytorch_model-00020-of-00041.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f4e9736d6bb618e2a87b8b32ac70429e3cdfb02af6c9a1e7ff6d6b2fc2c67b0f
3
+ size 951607026
pytorch_model-00021-of-00041.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:977dc581618aa467a92fd714652ea011cbcc5f31340cffd7a9b5985344b35d82
3
+ size 951607026
pytorch_model-00022-of-00041.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f4df87a3a8f51a56c2e35bcbc5dd7af8c48071fce198e727eedf92983cc48aa4
3
+ size 951607026
pytorch_model-00023-of-00041.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a06d3914925ac67e7d54cff5f09769aa8ff2ca802ad10ab6d67ac9931db52c71
3
+ size 951607026
pytorch_model-00024-of-00041.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2f6e1759670743dc9f7ab1d623a3770bea93cdfd4b90bc3b905ee3641255b639
3
+ size 951607026
pytorch_model-00025-of-00041.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:30c31f6bb567f55091a608a9826ff0f5902639321fe23d044e1b0aaefdd1c2e2
3
+ size 951607026
pytorch_model-00026-of-00041.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:22473c2581ffb15e07e731496c37e7ad733c0839ef0c87e0dd4b053b6c0f22d8
3
+ size 951607026
pytorch_model-00027-of-00041.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b9dd1fdf1e1ce6a8383163aee4c671e6e8c54fa02a6035aaf3c8ba17bb883c7e
3
+ size 951607026
pytorch_model-00028-of-00041.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:94fa6d8aa75d911cdc3f8ac16c328d021b7b0708bf66a00bb42eee83dc95cd00
3
+ size 951607026
pytorch_model-00029-of-00041.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ede04db4d006d4a6fa7d4df376de5640fbc81cb55a5030e5481a3f31e885e8c1
3
+ size 951607026
pytorch_model-00030-of-00041.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:edbe47ab7df79ad6621dd3d58632ef9d43489807fc78c5d330e039b457e73d3b
3
+ size 951607026
pytorch_model-00031-of-00041.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:48583ebe38227bd581f1c4f9f1d6e8bdb6a5d81d9539be8cc878cc6d3f8a171d
3
+ size 951607026
pytorch_model-00032-of-00041.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4dac2926314a8b2f8b62ef0b3834251419f3b28d262825658c076b41631b55b4
3
+ size 951607026
pytorch_model-00033-of-00041.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ded1034b1065bf2b80c65c3ef7588c18420423b135ad8cb588449ac5e7b0b5ae
3
+ size 951607026
pytorch_model-00034-of-00041.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e2fbdeca2fb4636cf4a9c61082f0ee6cb9ead581e6a9b366b25a41bd03ca11ab
3
+ size 951607026
pytorch_model-00035-of-00041.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:92aec1ceb44f4fd08614bf53e5b9d7a5b36240bd5d5413415aa1b722a619c56a
3
+ size 951607026
pytorch_model-00036-of-00041.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3b74725ed4f12af512053d149ed9e4924428d79b32d9d7870c8a5c10178798ae
3
+ size 951607026
pytorch_model-00037-of-00041.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:fafa213ed3a26d8113e1d88fe6cb900f94a192e88970edf0f01af7544546782b
3
+ size 951607026
pytorch_model-00038-of-00041.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f58ed89257a18ebac6568e102e1034d6652f14ec7dbaba0ab5df5b61fc5849eb
3
+ size 951607026
pytorch_model-00039-of-00041.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:61c11d72e93e8d841b098648d5943f9884733883999e9f2376060a4c07c54411
3
+ size 951607026
pytorch_model-00040-of-00041.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8919b8580dc1b5b22f38a5e5b524e5c019466c54c8fb941d8c19c824be2c3c3b
3
+ size 951607026
pytorch_model-00041-of-00041.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9436f4a24cb5ecbecab55cee0495902bf10bcc51534cceff2439d3f75e4cc507
3
+ size 983051724
pytorch_model.bin.index.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"weight_map": {"model.layers.0.input_layernorm.weight": "pytorch_model-00001-of-00041.bin", "model.layers.0.post_attention_layernorm.weight": "pytorch_model-00001-of-00041.bin", "model.layers.0.self_attn.q_proj.weight": "pytorch_model-00001-of-00041.bin", "model.layers.0.self_attn.k_proj.weight": "pytorch_model-00001-of-00041.bin", "model.layers.0.self_attn.v_proj.weight": "pytorch_model-00001-of-00041.bin", "model.layers.0.self_attn.o_proj.weight": "pytorch_model-00001-of-00041.bin", "model.layers.0.mlp.gate_proj.weight": "pytorch_model-00001-of-00041.bin", "model.layers.0.mlp.down_proj.weight": "pytorch_model-00001-of-00041.bin", "model.layers.0.mlp.up_proj.weight": "pytorch_model-00001-of-00041.bin", "model.layers.0.self_attn.rotary_emb.inv_freq": "pytorch_model-00001-of-00041.bin", "model.layers.1.input_layernorm.weight": "pytorch_model-00002-of-00041.bin", "model.layers.1.post_attention_layernorm.weight": "pytorch_model-00002-of-00041.bin", "model.layers.1.self_attn.q_proj.weight": "pytorch_model-00002-of-00041.bin", "model.layers.1.self_attn.k_proj.weight": "pytorch_model-00002-of-00041.bin", "model.layers.1.self_attn.v_proj.weight": "pytorch_model-00002-of-00041.bin", "model.layers.1.self_attn.o_proj.weight": "pytorch_model-00002-of-00041.bin", "model.layers.1.mlp.gate_proj.weight": "pytorch_model-00002-of-00041.bin", "model.layers.1.mlp.down_proj.weight": "pytorch_model-00002-of-00041.bin", "model.layers.1.mlp.up_proj.weight": "pytorch_model-00002-of-00041.bin", "model.layers.1.self_attn.rotary_emb.inv_freq": "pytorch_model-00002-of-00041.bin", "model.layers.2.input_layernorm.weight": "pytorch_model-00003-of-00041.bin", "model.layers.2.post_attention_layernorm.weight": "pytorch_model-00003-of-00041.bin", "model.layers.2.self_attn.q_proj.weight": "pytorch_model-00003-of-00041.bin", "model.layers.2.self_attn.k_proj.weight": "pytorch_model-00003-of-00041.bin", "model.layers.2.self_attn.v_proj.weight": "pytorch_model-00003-of-00041.bin", "model.layers.2.self_attn.o_proj.weight": "pytorch_model-00003-of-00041.bin", "model.layers.2.mlp.gate_proj.weight": "pytorch_model-00003-of-00041.bin", "model.layers.2.mlp.down_proj.weight": "pytorch_model-00003-of-00041.bin", "model.layers.2.mlp.up_proj.weight": "pytorch_model-00003-of-00041.bin", "model.layers.2.self_attn.rotary_emb.inv_freq": "pytorch_model-00003-of-00041.bin", "model.layers.3.input_layernorm.weight": "pytorch_model-00004-of-00041.bin", "model.layers.3.post_attention_layernorm.weight": "pytorch_model-00004-of-00041.bin", "model.layers.3.self_attn.q_proj.weight": "pytorch_model-00004-of-00041.bin", "model.layers.3.self_attn.k_proj.weight": "pytorch_model-00004-of-00041.bin", "model.layers.3.self_attn.v_proj.weight": "pytorch_model-00004-of-00041.bin", "model.layers.3.self_attn.o_proj.weight": "pytorch_model-00004-of-00041.bin", "model.layers.3.mlp.gate_proj.weight": "pytorch_model-00004-of-00041.bin", "model.layers.3.mlp.down_proj.weight": "pytorch_model-00004-of-00041.bin", "model.layers.3.mlp.up_proj.weight": "pytorch_model-00004-of-00041.bin", "model.layers.3.self_attn.rotary_emb.inv_freq": "pytorch_model-00004-of-00041.bin", "model.layers.4.input_layernorm.weight": "pytorch_model-00005-of-00041.bin", "model.layers.4.post_attention_layernorm.weight": "pytorch_model-00005-of-00041.bin", "model.layers.4.self_attn.q_proj.weight": "pytorch_model-00005-of-00041.bin", "model.layers.4.self_attn.k_proj.weight": "pytorch_model-00005-of-00041.bin", "model.layers.4.self_attn.v_proj.weight": "pytorch_model-00005-of-00041.bin", "model.layers.4.self_attn.o_proj.weight": "pytorch_model-00005-of-00041.bin", "model.layers.4.mlp.gate_proj.weight": "pytorch_model-00005-of-00041.bin", "model.layers.4.mlp.down_proj.weight": "pytorch_model-00005-of-00041.bin", "model.layers.4.mlp.up_proj.weight": "pytorch_model-00005-of-00041.bin", "model.layers.4.self_attn.rotary_emb.inv_freq": "pytorch_model-00005-of-00041.bin", "model.layers.5.input_layernorm.weight": "pytorch_model-00006-of-00041.bin", "model.layers.5.post_attention_layernorm.weight": "pytorch_model-00006-of-00041.bin", "model.layers.5.self_attn.q_proj.weight": "pytorch_model-00006-of-00041.bin", "model.layers.5.self_attn.k_proj.weight": "pytorch_model-00006-of-00041.bin", "model.layers.5.self_attn.v_proj.weight": "pytorch_model-00006-of-00041.bin", "model.layers.5.self_attn.o_proj.weight": "pytorch_model-00006-of-00041.bin", "model.layers.5.mlp.gate_proj.weight": "pytorch_model-00006-of-00041.bin", "model.layers.5.mlp.down_proj.weight": "pytorch_model-00006-of-00041.bin", "model.layers.5.mlp.up_proj.weight": "pytorch_model-00006-of-00041.bin", "model.layers.5.self_attn.rotary_emb.inv_freq": "pytorch_model-00006-of-00041.bin", "model.layers.6.input_layernorm.weight": "pytorch_model-00007-of-00041.bin", "model.layers.6.post_attention_layernorm.weight": "pytorch_model-00007-of-00041.bin", "model.layers.6.self_attn.q_proj.weight": "pytorch_model-00007-of-00041.bin", "model.layers.6.self_attn.k_proj.weight": "pytorch_model-00007-of-00041.bin", "model.layers.6.self_attn.v_proj.weight": "pytorch_model-00007-of-00041.bin", "model.layers.6.self_attn.o_proj.weight": "pytorch_model-00007-of-00041.bin", "model.layers.6.mlp.gate_proj.weight": "pytorch_model-00007-of-00041.bin", "model.layers.6.mlp.down_proj.weight": "pytorch_model-00007-of-00041.bin", "model.layers.6.mlp.up_proj.weight": "pytorch_model-00007-of-00041.bin", "model.layers.6.self_attn.rotary_emb.inv_freq": "pytorch_model-00007-of-00041.bin", "model.layers.7.input_layernorm.weight": "pytorch_model-00008-of-00041.bin", "model.layers.7.post_attention_layernorm.weight": "pytorch_model-00008-of-00041.bin", "model.layers.7.self_attn.q_proj.weight": "pytorch_model-00008-of-00041.bin", "model.layers.7.self_attn.k_proj.weight": "pytorch_model-00008-of-00041.bin", "model.layers.7.self_attn.v_proj.weight": "pytorch_model-00008-of-00041.bin", "model.layers.7.self_attn.o_proj.weight": "pytorch_model-00008-of-00041.bin", "model.layers.7.mlp.gate_proj.weight": "pytorch_model-00008-of-00041.bin", "model.layers.7.mlp.down_proj.weight": "pytorch_model-00008-of-00041.bin", "model.layers.7.mlp.up_proj.weight": "pytorch_model-00008-of-00041.bin", "model.layers.7.self_attn.rotary_emb.inv_freq": "pytorch_model-00008-of-00041.bin", "model.layers.8.input_layernorm.weight": "pytorch_model-00009-of-00041.bin", "model.layers.8.post_attention_layernorm.weight": "pytorch_model-00009-of-00041.bin", "model.layers.8.self_attn.q_proj.weight": "pytorch_model-00009-of-00041.bin", "model.layers.8.self_attn.k_proj.weight": "pytorch_model-00009-of-00041.bin", "model.layers.8.self_attn.v_proj.weight": "pytorch_model-00009-of-00041.bin", "model.layers.8.self_attn.o_proj.weight": "pytorch_model-00009-of-00041.bin", "model.layers.8.mlp.gate_proj.weight": "pytorch_model-00009-of-00041.bin", "model.layers.8.mlp.down_proj.weight": "pytorch_model-00009-of-00041.bin", "model.layers.8.mlp.up_proj.weight": "pytorch_model-00009-of-00041.bin", "model.layers.8.self_attn.rotary_emb.inv_freq": "pytorch_model-00009-of-00041.bin", "model.layers.9.input_layernorm.weight": "pytorch_model-00010-of-00041.bin", "model.layers.9.post_attention_layernorm.weight": "pytorch_model-00010-of-00041.bin", "model.layers.9.self_attn.q_proj.weight": "pytorch_model-00010-of-00041.bin", "model.layers.9.self_attn.k_proj.weight": "pytorch_model-00010-of-00041.bin", "model.layers.9.self_attn.v_proj.weight": "pytorch_model-00010-of-00041.bin", "model.layers.9.self_attn.o_proj.weight": "pytorch_model-00010-of-00041.bin", "model.layers.9.mlp.gate_proj.weight": "pytorch_model-00010-of-00041.bin", "model.layers.9.mlp.down_proj.weight": "pytorch_model-00010-of-00041.bin", "model.layers.9.mlp.up_proj.weight": "pytorch_model-00010-of-00041.bin", "model.layers.9.self_attn.rotary_emb.inv_freq": "pytorch_model-00010-of-00041.bin", "model.layers.10.input_layernorm.weight": "pytorch_model-00011-of-00041.bin", "model.layers.10.post_attention_layernorm.weight": "pytorch_model-00011-of-00041.bin", "model.layers.10.self_attn.q_proj.weight": "pytorch_model-00011-of-00041.bin", "model.layers.10.self_attn.k_proj.weight": "pytorch_model-00011-of-00041.bin", "model.layers.10.self_attn.v_proj.weight": "pytorch_model-00011-of-00041.bin", "model.layers.10.self_attn.o_proj.weight": "pytorch_model-00011-of-00041.bin", "model.layers.10.mlp.gate_proj.weight": "pytorch_model-00011-of-00041.bin", "model.layers.10.mlp.down_proj.weight": "pytorch_model-00011-of-00041.bin", "model.layers.10.mlp.up_proj.weight": "pytorch_model-00011-of-00041.bin", "model.layers.10.self_attn.rotary_emb.inv_freq": "pytorch_model-00011-of-00041.bin", "model.layers.11.input_layernorm.weight": "pytorch_model-00012-of-00041.bin", "model.layers.11.post_attention_layernorm.weight": "pytorch_model-00012-of-00041.bin", "model.layers.11.self_attn.q_proj.weight": "pytorch_model-00012-of-00041.bin", "model.layers.11.self_attn.k_proj.weight": "pytorch_model-00012-of-00041.bin", "model.layers.11.self_attn.v_proj.weight": "pytorch_model-00012-of-00041.bin", "model.layers.11.self_attn.o_proj.weight": "pytorch_model-00012-of-00041.bin", "model.layers.11.mlp.gate_proj.weight": "pytorch_model-00012-of-00041.bin", "model.layers.11.mlp.down_proj.weight": "pytorch_model-00012-of-00041.bin", "model.layers.11.mlp.up_proj.weight": "pytorch_model-00012-of-00041.bin", "model.layers.11.self_attn.rotary_emb.inv_freq": "pytorch_model-00012-of-00041.bin", "model.layers.12.input_layernorm.weight": "pytorch_model-00013-of-00041.bin", "model.layers.12.post_attention_layernorm.weight": "pytorch_model-00013-of-00041.bin", "model.layers.12.self_attn.q_proj.weight": "pytorch_model-00013-of-00041.bin", "model.layers.12.self_attn.k_proj.weight": "pytorch_model-00013-of-00041.bin", "model.layers.12.self_attn.v_proj.weight": "pytorch_model-00013-of-00041.bin", "model.layers.12.self_attn.o_proj.weight": "pytorch_model-00013-of-00041.bin", "model.layers.12.mlp.gate_proj.weight": "pytorch_model-00013-of-00041.bin", "model.layers.12.mlp.down_proj.weight": "pytorch_model-00013-of-00041.bin", "model.layers.12.mlp.up_proj.weight": "pytorch_model-00013-of-00041.bin", "model.layers.12.self_attn.rotary_emb.inv_freq": "pytorch_model-00013-of-00041.bin", "model.layers.13.input_layernorm.weight": "pytorch_model-00014-of-00041.bin", "model.layers.13.post_attention_layernorm.weight": "pytorch_model-00014-of-00041.bin", "model.layers.13.self_attn.q_proj.weight": "pytorch_model-00014-of-00041.bin", "model.layers.13.self_attn.k_proj.weight": "pytorch_model-00014-of-00041.bin", "model.layers.13.self_attn.v_proj.weight": "pytorch_model-00014-of-00041.bin", "model.layers.13.self_attn.o_proj.weight": "pytorch_model-00014-of-00041.bin", "model.layers.13.mlp.gate_proj.weight": "pytorch_model-00014-of-00041.bin", "model.layers.13.mlp.down_proj.weight": "pytorch_model-00014-of-00041.bin", "model.layers.13.mlp.up_proj.weight": "pytorch_model-00014-of-00041.bin", "model.layers.13.self_attn.rotary_emb.inv_freq": "pytorch_model-00014-of-00041.bin", "model.layers.14.input_layernorm.weight": "pytorch_model-00015-of-00041.bin", "model.layers.14.post_attention_layernorm.weight": "pytorch_model-00015-of-00041.bin", "model.layers.14.self_attn.q_proj.weight": "pytorch_model-00015-of-00041.bin", "model.layers.14.self_attn.k_proj.weight": "pytorch_model-00015-of-00041.bin", "model.layers.14.self_attn.v_proj.weight": "pytorch_model-00015-of-00041.bin", "model.layers.14.self_attn.o_proj.weight": "pytorch_model-00015-of-00041.bin", "model.layers.14.mlp.gate_proj.weight": "pytorch_model-00015-of-00041.bin", "model.layers.14.mlp.down_proj.weight": "pytorch_model-00015-of-00041.bin", "model.layers.14.mlp.up_proj.weight": "pytorch_model-00015-of-00041.bin", "model.layers.14.self_attn.rotary_emb.inv_freq": "pytorch_model-00015-of-00041.bin", "model.layers.15.input_layernorm.weight": "pytorch_model-00016-of-00041.bin", "model.layers.15.post_attention_layernorm.weight": "pytorch_model-00016-of-00041.bin", "model.layers.15.self_attn.q_proj.weight": "pytorch_model-00016-of-00041.bin", "model.layers.15.self_attn.k_proj.weight": "pytorch_model-00016-of-00041.bin", "model.layers.15.self_attn.v_proj.weight": "pytorch_model-00016-of-00041.bin", "model.layers.15.self_attn.o_proj.weight": "pytorch_model-00016-of-00041.bin", "model.layers.15.mlp.gate_proj.weight": "pytorch_model-00016-of-00041.bin", "model.layers.15.mlp.down_proj.weight": "pytorch_model-00016-of-00041.bin", "model.layers.15.mlp.up_proj.weight": "pytorch_model-00016-of-00041.bin", "model.layers.15.self_attn.rotary_emb.inv_freq": "pytorch_model-00016-of-00041.bin", "model.layers.16.input_layernorm.weight": "pytorch_model-00017-of-00041.bin", "model.layers.16.post_attention_layernorm.weight": "pytorch_model-00017-of-00041.bin", "model.layers.16.self_attn.q_proj.weight": "pytorch_model-00017-of-00041.bin", "model.layers.16.self_attn.k_proj.weight": "pytorch_model-00017-of-00041.bin", "model.layers.16.self_attn.v_proj.weight": "pytorch_model-00017-of-00041.bin", "model.layers.16.self_attn.o_proj.weight": "pytorch_model-00017-of-00041.bin", "model.layers.16.mlp.gate_proj.weight": "pytorch_model-00017-of-00041.bin", "model.layers.16.mlp.down_proj.weight": "pytorch_model-00017-of-00041.bin", "model.layers.16.mlp.up_proj.weight": "pytorch_model-00017-of-00041.bin", "model.layers.16.self_attn.rotary_emb.inv_freq": "pytorch_model-00017-of-00041.bin", "model.layers.17.input_layernorm.weight": "pytorch_model-00018-of-00041.bin", "model.layers.17.post_attention_layernorm.weight": "pytorch_model-00018-of-00041.bin", "model.layers.17.self_attn.q_proj.weight": "pytorch_model-00018-of-00041.bin", "model.layers.17.self_attn.k_proj.weight": "pytorch_model-00018-of-00041.bin", "model.layers.17.self_attn.v_proj.weight": "pytorch_model-00018-of-00041.bin", "model.layers.17.self_attn.o_proj.weight": "pytorch_model-00018-of-00041.bin", "model.layers.17.mlp.gate_proj.weight": "pytorch_model-00018-of-00041.bin", "model.layers.17.mlp.down_proj.weight": "pytorch_model-00018-of-00041.bin", "model.layers.17.mlp.up_proj.weight": "pytorch_model-00018-of-00041.bin", "model.layers.17.self_attn.rotary_emb.inv_freq": "pytorch_model-00018-of-00041.bin", "model.layers.18.input_layernorm.weight": "pytorch_model-00019-of-00041.bin", "model.layers.18.post_attention_layernorm.weight": "pytorch_model-00019-of-00041.bin", "model.layers.18.self_attn.q_proj.weight": "pytorch_model-00019-of-00041.bin", "model.layers.18.self_attn.k_proj.weight": "pytorch_model-00019-of-00041.bin", "model.layers.18.self_attn.v_proj.weight": "pytorch_model-00019-of-00041.bin", "model.layers.18.self_attn.o_proj.weight": "pytorch_model-00019-of-00041.bin", "model.layers.18.mlp.gate_proj.weight": "pytorch_model-00019-of-00041.bin", "model.layers.18.mlp.down_proj.weight": "pytorch_model-00019-of-00041.bin", "model.layers.18.mlp.up_proj.weight": "pytorch_model-00019-of-00041.bin", "model.layers.18.self_attn.rotary_emb.inv_freq": "pytorch_model-00019-of-00041.bin", "model.layers.19.input_layernorm.weight": "pytorch_model-00020-of-00041.bin", "model.layers.19.post_attention_layernorm.weight": "pytorch_model-00020-of-00041.bin", "model.layers.19.self_attn.q_proj.weight": "pytorch_model-00020-of-00041.bin", "model.layers.19.self_attn.k_proj.weight": "pytorch_model-00020-of-00041.bin", "model.layers.19.self_attn.v_proj.weight": "pytorch_model-00020-of-00041.bin", "model.layers.19.self_attn.o_proj.weight": "pytorch_model-00020-of-00041.bin", "model.layers.19.mlp.gate_proj.weight": "pytorch_model-00020-of-00041.bin", "model.layers.19.mlp.down_proj.weight": "pytorch_model-00020-of-00041.bin", "model.layers.19.mlp.up_proj.weight": "pytorch_model-00020-of-00041.bin", "model.layers.19.self_attn.rotary_emb.inv_freq": "pytorch_model-00020-of-00041.bin", "model.layers.20.input_layernorm.weight": "pytorch_model-00021-of-00041.bin", "model.layers.20.post_attention_layernorm.weight": "pytorch_model-00021-of-00041.bin", "model.layers.20.self_attn.q_proj.weight": "pytorch_model-00021-of-00041.bin", "model.layers.20.self_attn.k_proj.weight": "pytorch_model-00021-of-00041.bin", "model.layers.20.self_attn.v_proj.weight": "pytorch_model-00021-of-00041.bin", "model.layers.20.self_attn.o_proj.weight": "pytorch_model-00021-of-00041.bin", "model.layers.20.mlp.gate_proj.weight": "pytorch_model-00021-of-00041.bin", "model.layers.20.mlp.down_proj.weight": "pytorch_model-00021-of-00041.bin", "model.layers.20.mlp.up_proj.weight": "pytorch_model-00021-of-00041.bin", "model.layers.20.self_attn.rotary_emb.inv_freq": "pytorch_model-00021-of-00041.bin", "model.layers.21.input_layernorm.weight": "pytorch_model-00022-of-00041.bin", "model.layers.21.post_attention_layernorm.weight": "pytorch_model-00022-of-00041.bin", "model.layers.21.self_attn.q_proj.weight": "pytorch_model-00022-of-00041.bin", "model.layers.21.self_attn.k_proj.weight": "pytorch_model-00022-of-00041.bin", "model.layers.21.self_attn.v_proj.weight": "pytorch_model-00022-of-00041.bin", "model.layers.21.self_attn.o_proj.weight": "pytorch_model-00022-of-00041.bin", "model.layers.21.mlp.gate_proj.weight": "pytorch_model-00022-of-00041.bin", "model.layers.21.mlp.down_proj.weight": "pytorch_model-00022-of-00041.bin", "model.layers.21.mlp.up_proj.weight": "pytorch_model-00022-of-00041.bin", "model.layers.21.self_attn.rotary_emb.inv_freq": "pytorch_model-00022-of-00041.bin", "model.layers.22.input_layernorm.weight": "pytorch_model-00023-of-00041.bin", "model.layers.22.post_attention_layernorm.weight": "pytorch_model-00023-of-00041.bin", "model.layers.22.self_attn.q_proj.weight": "pytorch_model-00023-of-00041.bin", "model.layers.22.self_attn.k_proj.weight": "pytorch_model-00023-of-00041.bin", "model.layers.22.self_attn.v_proj.weight": "pytorch_model-00023-of-00041.bin", "model.layers.22.self_attn.o_proj.weight": "pytorch_model-00023-of-00041.bin", "model.layers.22.mlp.gate_proj.weight": "pytorch_model-00023-of-00041.bin", "model.layers.22.mlp.down_proj.weight": "pytorch_model-00023-of-00041.bin", "model.layers.22.mlp.up_proj.weight": "pytorch_model-00023-of-00041.bin", "model.layers.22.self_attn.rotary_emb.inv_freq": "pytorch_model-00023-of-00041.bin", "model.layers.23.input_layernorm.weight": "pytorch_model-00024-of-00041.bin", "model.layers.23.post_attention_layernorm.weight": "pytorch_model-00024-of-00041.bin", "model.layers.23.self_attn.q_proj.weight": "pytorch_model-00024-of-00041.bin", "model.layers.23.self_attn.k_proj.weight": "pytorch_model-00024-of-00041.bin", "model.layers.23.self_attn.v_proj.weight": "pytorch_model-00024-of-00041.bin", "model.layers.23.self_attn.o_proj.weight": "pytorch_model-00024-of-00041.bin", "model.layers.23.mlp.gate_proj.weight": "pytorch_model-00024-of-00041.bin", "model.layers.23.mlp.down_proj.weight": "pytorch_model-00024-of-00041.bin", "model.layers.23.mlp.up_proj.weight": "pytorch_model-00024-of-00041.bin", "model.layers.23.self_attn.rotary_emb.inv_freq": "pytorch_model-00024-of-00041.bin", "model.layers.24.input_layernorm.weight": "pytorch_model-00025-of-00041.bin", "model.layers.24.post_attention_layernorm.weight": "pytorch_model-00025-of-00041.bin", "model.layers.24.self_attn.q_proj.weight": "pytorch_model-00025-of-00041.bin", "model.layers.24.self_attn.k_proj.weight": "pytorch_model-00025-of-00041.bin", "model.layers.24.self_attn.v_proj.weight": "pytorch_model-00025-of-00041.bin", "model.layers.24.self_attn.o_proj.weight": "pytorch_model-00025-of-00041.bin", "model.layers.24.mlp.gate_proj.weight": "pytorch_model-00025-of-00041.bin", "model.layers.24.mlp.down_proj.weight": "pytorch_model-00025-of-00041.bin", "model.layers.24.mlp.up_proj.weight": "pytorch_model-00025-of-00041.bin", "model.layers.24.self_attn.rotary_emb.inv_freq": "pytorch_model-00025-of-00041.bin", "model.layers.25.input_layernorm.weight": "pytorch_model-00026-of-00041.bin", "model.layers.25.post_attention_layernorm.weight": "pytorch_model-00026-of-00041.bin", "model.layers.25.self_attn.q_proj.weight": "pytorch_model-00026-of-00041.bin", "model.layers.25.self_attn.k_proj.weight": "pytorch_model-00026-of-00041.bin", "model.layers.25.self_attn.v_proj.weight": "pytorch_model-00026-of-00041.bin", "model.layers.25.self_attn.o_proj.weight": "pytorch_model-00026-of-00041.bin", "model.layers.25.mlp.gate_proj.weight": "pytorch_model-00026-of-00041.bin", "model.layers.25.mlp.down_proj.weight": "pytorch_model-00026-of-00041.bin", "model.layers.25.mlp.up_proj.weight": "pytorch_model-00026-of-00041.bin", "model.layers.25.self_attn.rotary_emb.inv_freq": "pytorch_model-00026-of-00041.bin", "model.layers.26.input_layernorm.weight": "pytorch_model-00027-of-00041.bin", "model.layers.26.post_attention_layernorm.weight": "pytorch_model-00027-of-00041.bin", "model.layers.26.self_attn.q_proj.weight": "pytorch_model-00027-of-00041.bin", "model.layers.26.self_attn.k_proj.weight": "pytorch_model-00027-of-00041.bin", "model.layers.26.self_attn.v_proj.weight": "pytorch_model-00027-of-00041.bin", "model.layers.26.self_attn.o_proj.weight": "pytorch_model-00027-of-00041.bin", "model.layers.26.mlp.gate_proj.weight": "pytorch_model-00027-of-00041.bin", "model.layers.26.mlp.down_proj.weight": "pytorch_model-00027-of-00041.bin", "model.layers.26.mlp.up_proj.weight": "pytorch_model-00027-of-00041.bin", "model.layers.26.self_attn.rotary_emb.inv_freq": "pytorch_model-00027-of-00041.bin", "model.layers.27.input_layernorm.weight": "pytorch_model-00028-of-00041.bin", "model.layers.27.post_attention_layernorm.weight": "pytorch_model-00028-of-00041.bin", "model.layers.27.self_attn.q_proj.weight": "pytorch_model-00028-of-00041.bin", "model.layers.27.self_attn.k_proj.weight": "pytorch_model-00028-of-00041.bin", "model.layers.27.self_attn.v_proj.weight": "pytorch_model-00028-of-00041.bin", "model.layers.27.self_attn.o_proj.weight": "pytorch_model-00028-of-00041.bin", "model.layers.27.mlp.gate_proj.weight": "pytorch_model-00028-of-00041.bin", "model.layers.27.mlp.down_proj.weight": "pytorch_model-00028-of-00041.bin", "model.layers.27.mlp.up_proj.weight": "pytorch_model-00028-of-00041.bin", "model.layers.27.self_attn.rotary_emb.inv_freq": "pytorch_model-00028-of-00041.bin", "model.layers.28.input_layernorm.weight": "pytorch_model-00029-of-00041.bin", "model.layers.28.post_attention_layernorm.weight": "pytorch_model-00029-of-00041.bin", "model.layers.28.self_attn.q_proj.weight": "pytorch_model-00029-of-00041.bin", "model.layers.28.self_attn.k_proj.weight": "pytorch_model-00029-of-00041.bin", "model.layers.28.self_attn.v_proj.weight": "pytorch_model-00029-of-00041.bin", "model.layers.28.self_attn.o_proj.weight": "pytorch_model-00029-of-00041.bin", "model.layers.28.mlp.gate_proj.weight": "pytorch_model-00029-of-00041.bin", "model.layers.28.mlp.down_proj.weight": "pytorch_model-00029-of-00041.bin", "model.layers.28.mlp.up_proj.weight": "pytorch_model-00029-of-00041.bin", "model.layers.28.self_attn.rotary_emb.inv_freq": "pytorch_model-00029-of-00041.bin", "model.layers.29.input_layernorm.weight": "pytorch_model-00030-of-00041.bin", "model.layers.29.post_attention_layernorm.weight": "pytorch_model-00030-of-00041.bin", "model.layers.29.self_attn.q_proj.weight": "pytorch_model-00030-of-00041.bin", "model.layers.29.self_attn.k_proj.weight": "pytorch_model-00030-of-00041.bin", "model.layers.29.self_attn.v_proj.weight": "pytorch_model-00030-of-00041.bin", "model.layers.29.self_attn.o_proj.weight": "pytorch_model-00030-of-00041.bin", "model.layers.29.mlp.gate_proj.weight": "pytorch_model-00030-of-00041.bin", "model.layers.29.mlp.down_proj.weight": "pytorch_model-00030-of-00041.bin", "model.layers.29.mlp.up_proj.weight": "pytorch_model-00030-of-00041.bin", "model.layers.29.self_attn.rotary_emb.inv_freq": "pytorch_model-00030-of-00041.bin", "model.layers.30.input_layernorm.weight": "pytorch_model-00031-of-00041.bin", "model.layers.30.post_attention_layernorm.weight": "pytorch_model-00031-of-00041.bin", "model.layers.30.self_attn.q_proj.weight": "pytorch_model-00031-of-00041.bin", "model.layers.30.self_attn.k_proj.weight": "pytorch_model-00031-of-00041.bin", "model.layers.30.self_attn.v_proj.weight": "pytorch_model-00031-of-00041.bin", "model.layers.30.self_attn.o_proj.weight": "pytorch_model-00031-of-00041.bin", "model.layers.30.mlp.gate_proj.weight": "pytorch_model-00031-of-00041.bin", "model.layers.30.mlp.down_proj.weight": "pytorch_model-00031-of-00041.bin", "model.layers.30.mlp.up_proj.weight": "pytorch_model-00031-of-00041.bin", "model.layers.30.self_attn.rotary_emb.inv_freq": "pytorch_model-00031-of-00041.bin", "model.layers.31.input_layernorm.weight": "pytorch_model-00032-of-00041.bin", "model.layers.31.post_attention_layernorm.weight": "pytorch_model-00032-of-00041.bin", "model.layers.31.self_attn.q_proj.weight": "pytorch_model-00032-of-00041.bin", "model.layers.31.self_attn.k_proj.weight": "pytorch_model-00032-of-00041.bin", "model.layers.31.self_attn.v_proj.weight": "pytorch_model-00032-of-00041.bin", "model.layers.31.self_attn.o_proj.weight": "pytorch_model-00032-of-00041.bin", "model.layers.31.mlp.gate_proj.weight": "pytorch_model-00032-of-00041.bin", "model.layers.31.mlp.down_proj.weight": "pytorch_model-00032-of-00041.bin", "model.layers.31.mlp.up_proj.weight": "pytorch_model-00032-of-00041.bin", "model.layers.31.self_attn.rotary_emb.inv_freq": "pytorch_model-00032-of-00041.bin", "model.layers.32.input_layernorm.weight": "pytorch_model-00033-of-00041.bin", "model.layers.32.post_attention_layernorm.weight": "pytorch_model-00033-of-00041.bin", "model.layers.32.self_attn.q_proj.weight": "pytorch_model-00033-of-00041.bin", "model.layers.32.self_attn.k_proj.weight": "pytorch_model-00033-of-00041.bin", "model.layers.32.self_attn.v_proj.weight": "pytorch_model-00033-of-00041.bin", "model.layers.32.self_attn.o_proj.weight": "pytorch_model-00033-of-00041.bin", "model.layers.32.mlp.gate_proj.weight": "pytorch_model-00033-of-00041.bin", "model.layers.32.mlp.down_proj.weight": "pytorch_model-00033-of-00041.bin", "model.layers.32.mlp.up_proj.weight": "pytorch_model-00033-of-00041.bin", "model.layers.32.self_attn.rotary_emb.inv_freq": "pytorch_model-00033-of-00041.bin", "model.layers.33.input_layernorm.weight": "pytorch_model-00034-of-00041.bin", "model.layers.33.post_attention_layernorm.weight": "pytorch_model-00034-of-00041.bin", "model.layers.33.self_attn.q_proj.weight": "pytorch_model-00034-of-00041.bin", "model.layers.33.self_attn.k_proj.weight": "pytorch_model-00034-of-00041.bin", "model.layers.33.self_attn.v_proj.weight": "pytorch_model-00034-of-00041.bin", "model.layers.33.self_attn.o_proj.weight": "pytorch_model-00034-of-00041.bin", "model.layers.33.mlp.gate_proj.weight": "pytorch_model-00034-of-00041.bin", "model.layers.33.mlp.down_proj.weight": "pytorch_model-00034-of-00041.bin", "model.layers.33.mlp.up_proj.weight": "pytorch_model-00034-of-00041.bin", "model.layers.33.self_attn.rotary_emb.inv_freq": "pytorch_model-00034-of-00041.bin", "model.layers.34.input_layernorm.weight": "pytorch_model-00035-of-00041.bin", "model.layers.34.post_attention_layernorm.weight": "pytorch_model-00035-of-00041.bin", "model.layers.34.self_attn.q_proj.weight": "pytorch_model-00035-of-00041.bin", "model.layers.34.self_attn.k_proj.weight": "pytorch_model-00035-of-00041.bin", "model.layers.34.self_attn.v_proj.weight": "pytorch_model-00035-of-00041.bin", "model.layers.34.self_attn.o_proj.weight": "pytorch_model-00035-of-00041.bin", "model.layers.34.mlp.gate_proj.weight": "pytorch_model-00035-of-00041.bin", "model.layers.34.mlp.down_proj.weight": "pytorch_model-00035-of-00041.bin", "model.layers.34.mlp.up_proj.weight": "pytorch_model-00035-of-00041.bin", "model.layers.34.self_attn.rotary_emb.inv_freq": "pytorch_model-00035-of-00041.bin", "model.layers.35.input_layernorm.weight": "pytorch_model-00036-of-00041.bin", "model.layers.35.post_attention_layernorm.weight": "pytorch_model-00036-of-00041.bin", "model.layers.35.self_attn.q_proj.weight": "pytorch_model-00036-of-00041.bin", "model.layers.35.self_attn.k_proj.weight": "pytorch_model-00036-of-00041.bin", "model.layers.35.self_attn.v_proj.weight": "pytorch_model-00036-of-00041.bin", "model.layers.35.self_attn.o_proj.weight": "pytorch_model-00036-of-00041.bin", "model.layers.35.mlp.gate_proj.weight": "pytorch_model-00036-of-00041.bin", "model.layers.35.mlp.down_proj.weight": "pytorch_model-00036-of-00041.bin", "model.layers.35.mlp.up_proj.weight": "pytorch_model-00036-of-00041.bin", "model.layers.35.self_attn.rotary_emb.inv_freq": "pytorch_model-00036-of-00041.bin", "model.layers.36.input_layernorm.weight": "pytorch_model-00037-of-00041.bin", "model.layers.36.post_attention_layernorm.weight": "pytorch_model-00037-of-00041.bin", "model.layers.36.self_attn.q_proj.weight": "pytorch_model-00037-of-00041.bin", "model.layers.36.self_attn.k_proj.weight": "pytorch_model-00037-of-00041.bin", "model.layers.36.self_attn.v_proj.weight": "pytorch_model-00037-of-00041.bin", "model.layers.36.self_attn.o_proj.weight": "pytorch_model-00037-of-00041.bin", "model.layers.36.mlp.gate_proj.weight": "pytorch_model-00037-of-00041.bin", "model.layers.36.mlp.down_proj.weight": "pytorch_model-00037-of-00041.bin", "model.layers.36.mlp.up_proj.weight": "pytorch_model-00037-of-00041.bin", "model.layers.36.self_attn.rotary_emb.inv_freq": "pytorch_model-00037-of-00041.bin", "model.layers.37.input_layernorm.weight": "pytorch_model-00038-of-00041.bin", "model.layers.37.post_attention_layernorm.weight": "pytorch_model-00038-of-00041.bin", "model.layers.37.self_attn.q_proj.weight": "pytorch_model-00038-of-00041.bin", "model.layers.37.self_attn.k_proj.weight": "pytorch_model-00038-of-00041.bin", "model.layers.37.self_attn.v_proj.weight": "pytorch_model-00038-of-00041.bin", "model.layers.37.self_attn.o_proj.weight": "pytorch_model-00038-of-00041.bin", "model.layers.37.mlp.gate_proj.weight": "pytorch_model-00038-of-00041.bin", "model.layers.37.mlp.down_proj.weight": "pytorch_model-00038-of-00041.bin", "model.layers.37.mlp.up_proj.weight": "pytorch_model-00038-of-00041.bin", "model.layers.37.self_attn.rotary_emb.inv_freq": "pytorch_model-00038-of-00041.bin", "model.layers.38.input_layernorm.weight": "pytorch_model-00039-of-00041.bin", "model.layers.38.post_attention_layernorm.weight": "pytorch_model-00039-of-00041.bin", "model.layers.38.self_attn.q_proj.weight": "pytorch_model-00039-of-00041.bin", "model.layers.38.self_attn.k_proj.weight": "pytorch_model-00039-of-00041.bin", "model.layers.38.self_attn.v_proj.weight": "pytorch_model-00039-of-00041.bin", "model.layers.38.self_attn.o_proj.weight": "pytorch_model-00039-of-00041.bin", "model.layers.38.mlp.gate_proj.weight": "pytorch_model-00039-of-00041.bin", "model.layers.38.mlp.down_proj.weight": "pytorch_model-00039-of-00041.bin", "model.layers.38.mlp.up_proj.weight": "pytorch_model-00039-of-00041.bin", "model.layers.38.self_attn.rotary_emb.inv_freq": "pytorch_model-00039-of-00041.bin", "model.layers.39.input_layernorm.weight": "pytorch_model-00040-of-00041.bin", "model.layers.39.post_attention_layernorm.weight": "pytorch_model-00040-of-00041.bin", "model.layers.39.self_attn.q_proj.weight": "pytorch_model-00040-of-00041.bin", "model.layers.39.self_attn.k_proj.weight": "pytorch_model-00040-of-00041.bin", "model.layers.39.self_attn.v_proj.weight": "pytorch_model-00040-of-00041.bin", "model.layers.39.self_attn.o_proj.weight": "pytorch_model-00040-of-00041.bin", "model.layers.39.mlp.gate_proj.weight": "pytorch_model-00040-of-00041.bin", "model.layers.39.mlp.down_proj.weight": "pytorch_model-00040-of-00041.bin", "model.layers.39.mlp.up_proj.weight": "pytorch_model-00040-of-00041.bin", "model.layers.39.self_attn.rotary_emb.inv_freq": "pytorch_model-00040-of-00041.bin", "model.norm.weight": "pytorch_model-00041-of-00041.bin", "model.embed_tokens.weight": "pytorch_model-00041-of-00041.bin", "lm_head.weight": "pytorch_model-00041-of-00041.bin"}, "metadata": {"total_size": 26031733760}}
special_tokens_map.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {}