sharpenb commited on
Commit
55424c3
·
verified ·
1 Parent(s): ee40e2f

Upload folder using huggingface_hub (#3)

Browse files

- 85d915d7878324eb59a67628c1209e23c1322d10e19f6ad681cc4a79f3591f7c (12cab331b686ed2c50156a9919b2e41379b86f8d)
- c59e15490e11479cf687c25902a3f517e098af1367d3e2d55a11c4e7f68190e3 (86b73252ccf978342944bec061f8f907e8b72f3b)
- c79dd03db625f10edc3b00811d3888fe5cd41b25598125accc3c1653c35019ee (97dfff44f44445dc3c0d3260c2c07c5897b34b38)
- 3fe323d4120cd3375673453998ab105206ab094cd6f2908ba67a2fe9228057c1 (d9f7a4707252fbfad87ac156fb1025796e0f89ea)
- 32a67da953b8c160569cbcff906becc58cb641bdb51b9ce7deb9916b11ffdb6e (79d7c21aab2eddffc9cc6f318939041850ddf46a)
- 69e958e9f211affa9e70263c567c47e6bb4921bd2e4f9d354a7edcd1f26a21ce (bf75b12df892de7ac6717d12d60c4c8f6408b439)
- 347c3e87f6b06bed1123bae751e04e74131819a3feb6d2a896eae35da53be620 (7078c45d87a8305a5b8f2fc45472c766dcf5af33)
- e475021b2fc4d9e209f0a902d98fbf5461e4c343aba15ff995055c1cfa780fa8 (f300f5a5adca0b40aec6a925d01471da3be29047)
- f53dc08b38ebfd6534ddf627bbad4aee45089cb573f723abda428e83f25657a3 (eed66ae0f101fd79406e7c364f1a56220661547e)
- a80bd4923cb01ddbbb94473baeea41240e0b114ffee98473d143a2d1fcdcaa79 (3f4a0f76a4b4b4c8d3e35276813475449eb6d7e5)
- 87c4b1f661ac695a5431fb281e924520e15e6d72d9ba7477b42a2b7ca7f40043 (4783c1bba0cafc4c8b3c711c83d8e1772135aba0)
- 4bb369a0c183bb678a44bd63a1f3fa5ad27779c84ffa59b8ea203534af067c99 (e22345fa81be134f93de54eba2de31b59fa635bf)
- eeb0be3c0a612b41fd89aee994c5c67311e294a11e16f862a47c48f1c8285c26 (0862f82b608d3b96d75d02ab1e292feff58e339a)
- 088a1b13cbac33bafa445b55cde1c479c6cfff99a47fc736e9850a75c248b043 (f7704f2b7d92a7b02233319d1af1832ffc39724a)

config.json CHANGED
@@ -23,23 +23,21 @@
23
  "quantization_config": {
24
  "_load_in_4bit": true,
25
  "_load_in_8bit": false,
26
- "bnb_4bit_compute_dtype": "bfloat16",
27
  "bnb_4bit_quant_storage": "uint8",
28
  "bnb_4bit_quant_type": "fp4",
29
  "bnb_4bit_use_double_quant": false,
30
  "llm_int8_enable_fp32_cpu_offload": false,
31
  "llm_int8_has_fp16_weight": false,
32
- "llm_int8_skip_modules": [
33
- "lm_head"
34
- ],
35
  "llm_int8_threshold": 6.0,
36
  "load_in_4bit": true,
37
  "load_in_8bit": false,
38
  "quant_method": "bitsandbytes"
39
  },
40
  "rope_theta": 75000000.0,
41
- "torch_dtype": "bfloat16",
42
- "transformers_version": "4.39.3",
43
  "use_cache": true,
44
  "use_qk_norm": true,
45
  "vocab_size": 256000
 
23
  "quantization_config": {
24
  "_load_in_4bit": true,
25
  "_load_in_8bit": false,
26
+ "bnb_4bit_compute_dtype": "float32",
27
  "bnb_4bit_quant_storage": "uint8",
28
  "bnb_4bit_quant_type": "fp4",
29
  "bnb_4bit_use_double_quant": false,
30
  "llm_int8_enable_fp32_cpu_offload": false,
31
  "llm_int8_has_fp16_weight": false,
32
+ "llm_int8_skip_modules": null,
 
 
33
  "llm_int8_threshold": 6.0,
34
  "load_in_4bit": true,
35
  "load_in_8bit": false,
36
  "quant_method": "bitsandbytes"
37
  },
38
  "rope_theta": 75000000.0,
39
+ "torch_dtype": "float16",
40
+ "transformers_version": "4.40.0.dev0",
41
  "use_cache": true,
42
  "use_qk_norm": true,
43
  "vocab_size": 256000
generation_config.json CHANGED
@@ -3,5 +3,5 @@
3
  "bos_token_id": 5,
4
  "eos_token_id": 255001,
5
  "pad_token_id": 0,
6
- "transformers_version": "4.39.3"
7
  }
 
3
  "bos_token_id": 5,
4
  "eos_token_id": 255001,
5
  "pad_token_id": 0,
6
+ "transformers_version": "4.40.0.dev0"
7
  }
model-00001-of-00013.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:24556465c06abb47e07225bf653535e59081099da5319299a294a2cfa03e0af3
3
  size 6291456144
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ee321057203258a9356a5fe30deab557b0b7070d1da93153b5dddd66328cf000
3
  size 6291456144
model-00002-of-00013.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:29e178fbfbfe94571750c29e9b86567aa020cc9d358eeca4c7303aed6c7ddc8a
3
- size 4841424092
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:bc6e0fbf34d0556f2493cfef9af9028f34022d43f93656b809318fef15fc3912
3
+ size 4841585116
model-00003-of-00013.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:73047ab2f7ea8f7c37a879c1d825cb25ed4962809bf32ba5512d8a2858a24820
3
- size 4990084444
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a9fe7f614fdd5f5c89e073f7ca007a787a6716ef292968b864dda9ec46daf7dc
3
+ size 4990245468
model-00004-of-00013.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:f8f2ec69c369998b4ca039735bc2a0d7b8857a0806d40d42fb9c495731ae077c
3
- size 4975902838
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b40ea27cb60fdaa9929e9a5ca9e6d3778ce2116125303952ea363e45280048a3
3
+ size 4976037032
model-00005-of-00013.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:4c856cfa7343ad282452f103bd19798c20a223313d99399f467c128245e46e66
3
- size 4841448956
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d9fd066feceee09ae3d635a82a5d2e9185c4fbbabe0fc617c7044dd1c6cfc49b
3
+ size 4841609996
model-00006-of-00013.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:91d13810aabccf6566ca9f781103ed8dc302d31faf02469342e66ce1327dbf84
3
- size 4890967546
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b8f2feab12ad1bb6f33139d770a395cbe54e4487707aa120c3e3308d50fc9c89
3
+ size 4891101741
model-00007-of-00013.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:933a6f9722b4ce31e4a4449146af7d267f2eb7e9b4756e662ac46886a92d5879
3
- size 4841448956
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7b13d297600dd703c65d8761ca5f3179f3c46cd1d5635bd08768f4dabbf763fa
3
+ size 4841609996
model-00008-of-00013.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:ad477af24261d70a0215b7ec6aa022c04bce81f704abc98dff88cc6781adbe48
3
- size 4890967546
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2917c4ad99a6ecfbb1c2f2115ca641a89fef4d3c22eced0a3e2fb1e290eb7744
3
+ size 4891101741
model-00009-of-00013.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:ca7824fee63f98bdd3f4d474055851792683c70bfeaa5e81f7ea1db0712b9852
3
- size 4841448956
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b9ea840650e66a38915d6b11411f871d4576b9937abd8d2f1ec1ab2d2cf06a0f
3
+ size 4841609996
model-00010-of-00013.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:e2a538ce4cba9a4e5db72968a6bc792ae1ef893f8c727a48d00f7c527121df08
3
- size 4890967546
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e65ff8d242b12d9604ec233e2d3bd46d5de5cd3e1b9f637a1363fffae57f0026
3
+ size 4891101741
model-00011-of-00013.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:18832ba8fc377cd101182bc1375e1733765b872e7e9ae9545eef4d4516879fe1
3
- size 4841448956
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:097002aa2fc8ce2d7e2e088a81d7c4e203ec739673c12a34c5a234849270d410
3
+ size 4841609996
model-00012-of-00013.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:9b3fd6a1c3ed2297cabc3674d8168fbeee07c3e706a5433a8b4cc49967786a01
3
- size 4890967546
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f4df3d4a2d7e70b27b27eaae3883198af66c75c099f49f5bd372e1486b3e9939
3
+ size 4891101741
model-00013-of-00013.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:d56000379065b81ac28d6407242dddae5d7a2478db48fca3e59168def0e7cbf9
3
- size 2887915698
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:352a5e44a0c7257724cf7cbca26f30fe84dbe89b932c67630adef1e07d22bd8d
3
+ size 2887996212
model.safetensors.index.json CHANGED
@@ -1,6 +1,6 @@
1
  {
2
  "metadata": {
3
- "total_size": 62916223616
4
  },
5
  "weight_map": {
6
  "model.embed_tokens.weight": "model-00001-of-00013.safetensors",
@@ -17,6 +17,7 @@
17
  "model.layers.0.mlp.up_proj.weight.absmax": "model-00002-of-00013.safetensors",
18
  "model.layers.0.mlp.up_proj.weight.quant_map": "model-00002-of-00013.safetensors",
19
  "model.layers.0.mlp.up_proj.weight.quant_state.bitsandbytes__fp4": "model-00002-of-00013.safetensors",
 
20
  "model.layers.0.self_attn.k_proj.weight": "model-00002-of-00013.safetensors",
21
  "model.layers.0.self_attn.k_proj.weight.absmax": "model-00002-of-00013.safetensors",
22
  "model.layers.0.self_attn.k_proj.weight.quant_map": "model-00002-of-00013.safetensors",
@@ -25,6 +26,7 @@
25
  "model.layers.0.self_attn.o_proj.weight.absmax": "model-00002-of-00013.safetensors",
26
  "model.layers.0.self_attn.o_proj.weight.quant_map": "model-00002-of-00013.safetensors",
27
  "model.layers.0.self_attn.o_proj.weight.quant_state.bitsandbytes__fp4": "model-00002-of-00013.safetensors",
 
28
  "model.layers.0.self_attn.q_proj.weight": "model-00002-of-00013.safetensors",
29
  "model.layers.0.self_attn.q_proj.weight.absmax": "model-00002-of-00013.safetensors",
30
  "model.layers.0.self_attn.q_proj.weight.quant_map": "model-00002-of-00013.safetensors",
@@ -46,6 +48,7 @@
46
  "model.layers.1.mlp.up_proj.weight.absmax": "model-00002-of-00013.safetensors",
47
  "model.layers.1.mlp.up_proj.weight.quant_map": "model-00002-of-00013.safetensors",
48
  "model.layers.1.mlp.up_proj.weight.quant_state.bitsandbytes__fp4": "model-00002-of-00013.safetensors",
 
49
  "model.layers.1.self_attn.k_proj.weight": "model-00002-of-00013.safetensors",
50
  "model.layers.1.self_attn.k_proj.weight.absmax": "model-00002-of-00013.safetensors",
51
  "model.layers.1.self_attn.k_proj.weight.quant_map": "model-00002-of-00013.safetensors",
@@ -54,6 +57,7 @@
54
  "model.layers.1.self_attn.o_proj.weight.absmax": "model-00002-of-00013.safetensors",
55
  "model.layers.1.self_attn.o_proj.weight.quant_map": "model-00002-of-00013.safetensors",
56
  "model.layers.1.self_attn.o_proj.weight.quant_state.bitsandbytes__fp4": "model-00002-of-00013.safetensors",
 
57
  "model.layers.1.self_attn.q_proj.weight": "model-00002-of-00013.safetensors",
58
  "model.layers.1.self_attn.q_proj.weight.absmax": "model-00002-of-00013.safetensors",
59
  "model.layers.1.self_attn.q_proj.weight.quant_map": "model-00002-of-00013.safetensors",
@@ -75,6 +79,7 @@
75
  "model.layers.10.mlp.up_proj.weight.absmax": "model-00003-of-00013.safetensors",
76
  "model.layers.10.mlp.up_proj.weight.quant_map": "model-00003-of-00013.safetensors",
77
  "model.layers.10.mlp.up_proj.weight.quant_state.bitsandbytes__fp4": "model-00003-of-00013.safetensors",
 
78
  "model.layers.10.self_attn.k_proj.weight": "model-00003-of-00013.safetensors",
79
  "model.layers.10.self_attn.k_proj.weight.absmax": "model-00003-of-00013.safetensors",
80
  "model.layers.10.self_attn.k_proj.weight.quant_map": "model-00003-of-00013.safetensors",
@@ -83,6 +88,7 @@
83
  "model.layers.10.self_attn.o_proj.weight.absmax": "model-00003-of-00013.safetensors",
84
  "model.layers.10.self_attn.o_proj.weight.quant_map": "model-00003-of-00013.safetensors",
85
  "model.layers.10.self_attn.o_proj.weight.quant_state.bitsandbytes__fp4": "model-00003-of-00013.safetensors",
 
86
  "model.layers.10.self_attn.q_proj.weight": "model-00003-of-00013.safetensors",
87
  "model.layers.10.self_attn.q_proj.weight.absmax": "model-00003-of-00013.safetensors",
88
  "model.layers.10.self_attn.q_proj.weight.quant_map": "model-00003-of-00013.safetensors",
@@ -104,6 +110,7 @@
104
  "model.layers.11.mlp.up_proj.weight.absmax": "model-00004-of-00013.safetensors",
105
  "model.layers.11.mlp.up_proj.weight.quant_map": "model-00004-of-00013.safetensors",
106
  "model.layers.11.mlp.up_proj.weight.quant_state.bitsandbytes__fp4": "model-00004-of-00013.safetensors",
 
107
  "model.layers.11.self_attn.k_proj.weight": "model-00003-of-00013.safetensors",
108
  "model.layers.11.self_attn.k_proj.weight.absmax": "model-00003-of-00013.safetensors",
109
  "model.layers.11.self_attn.k_proj.weight.quant_map": "model-00003-of-00013.safetensors",
@@ -112,6 +119,7 @@
112
  "model.layers.11.self_attn.o_proj.weight.absmax": "model-00004-of-00013.safetensors",
113
  "model.layers.11.self_attn.o_proj.weight.quant_map": "model-00004-of-00013.safetensors",
114
  "model.layers.11.self_attn.o_proj.weight.quant_state.bitsandbytes__fp4": "model-00004-of-00013.safetensors",
 
115
  "model.layers.11.self_attn.q_proj.weight": "model-00003-of-00013.safetensors",
116
  "model.layers.11.self_attn.q_proj.weight.absmax": "model-00003-of-00013.safetensors",
117
  "model.layers.11.self_attn.q_proj.weight.quant_map": "model-00003-of-00013.safetensors",
@@ -133,6 +141,7 @@
133
  "model.layers.12.mlp.up_proj.weight.absmax": "model-00004-of-00013.safetensors",
134
  "model.layers.12.mlp.up_proj.weight.quant_map": "model-00004-of-00013.safetensors",
135
  "model.layers.12.mlp.up_proj.weight.quant_state.bitsandbytes__fp4": "model-00004-of-00013.safetensors",
 
136
  "model.layers.12.self_attn.k_proj.weight": "model-00004-of-00013.safetensors",
137
  "model.layers.12.self_attn.k_proj.weight.absmax": "model-00004-of-00013.safetensors",
138
  "model.layers.12.self_attn.k_proj.weight.quant_map": "model-00004-of-00013.safetensors",
@@ -141,6 +150,7 @@
141
  "model.layers.12.self_attn.o_proj.weight.absmax": "model-00004-of-00013.safetensors",
142
  "model.layers.12.self_attn.o_proj.weight.quant_map": "model-00004-of-00013.safetensors",
143
  "model.layers.12.self_attn.o_proj.weight.quant_state.bitsandbytes__fp4": "model-00004-of-00013.safetensors",
 
144
  "model.layers.12.self_attn.q_proj.weight": "model-00004-of-00013.safetensors",
145
  "model.layers.12.self_attn.q_proj.weight.absmax": "model-00004-of-00013.safetensors",
146
  "model.layers.12.self_attn.q_proj.weight.quant_map": "model-00004-of-00013.safetensors",
@@ -162,6 +172,7 @@
162
  "model.layers.13.mlp.up_proj.weight.absmax": "model-00004-of-00013.safetensors",
163
  "model.layers.13.mlp.up_proj.weight.quant_map": "model-00004-of-00013.safetensors",
164
  "model.layers.13.mlp.up_proj.weight.quant_state.bitsandbytes__fp4": "model-00004-of-00013.safetensors",
 
165
  "model.layers.13.self_attn.k_proj.weight": "model-00004-of-00013.safetensors",
166
  "model.layers.13.self_attn.k_proj.weight.absmax": "model-00004-of-00013.safetensors",
167
  "model.layers.13.self_attn.k_proj.weight.quant_map": "model-00004-of-00013.safetensors",
@@ -170,6 +181,7 @@
170
  "model.layers.13.self_attn.o_proj.weight.absmax": "model-00004-of-00013.safetensors",
171
  "model.layers.13.self_attn.o_proj.weight.quant_map": "model-00004-of-00013.safetensors",
172
  "model.layers.13.self_attn.o_proj.weight.quant_state.bitsandbytes__fp4": "model-00004-of-00013.safetensors",
 
173
  "model.layers.13.self_attn.q_proj.weight": "model-00004-of-00013.safetensors",
174
  "model.layers.13.self_attn.q_proj.weight.absmax": "model-00004-of-00013.safetensors",
175
  "model.layers.13.self_attn.q_proj.weight.quant_map": "model-00004-of-00013.safetensors",
@@ -191,6 +203,7 @@
191
  "model.layers.14.mlp.up_proj.weight.absmax": "model-00004-of-00013.safetensors",
192
  "model.layers.14.mlp.up_proj.weight.quant_map": "model-00004-of-00013.safetensors",
193
  "model.layers.14.mlp.up_proj.weight.quant_state.bitsandbytes__fp4": "model-00004-of-00013.safetensors",
 
194
  "model.layers.14.self_attn.k_proj.weight": "model-00004-of-00013.safetensors",
195
  "model.layers.14.self_attn.k_proj.weight.absmax": "model-00004-of-00013.safetensors",
196
  "model.layers.14.self_attn.k_proj.weight.quant_map": "model-00004-of-00013.safetensors",
@@ -199,6 +212,7 @@
199
  "model.layers.14.self_attn.o_proj.weight.absmax": "model-00004-of-00013.safetensors",
200
  "model.layers.14.self_attn.o_proj.weight.quant_map": "model-00004-of-00013.safetensors",
201
  "model.layers.14.self_attn.o_proj.weight.quant_state.bitsandbytes__fp4": "model-00004-of-00013.safetensors",
 
202
  "model.layers.14.self_attn.q_proj.weight": "model-00004-of-00013.safetensors",
203
  "model.layers.14.self_attn.q_proj.weight.absmax": "model-00004-of-00013.safetensors",
204
  "model.layers.14.self_attn.q_proj.weight.quant_map": "model-00004-of-00013.safetensors",
@@ -220,6 +234,7 @@
220
  "model.layers.15.mlp.up_proj.weight.absmax": "model-00004-of-00013.safetensors",
221
  "model.layers.15.mlp.up_proj.weight.quant_map": "model-00004-of-00013.safetensors",
222
  "model.layers.15.mlp.up_proj.weight.quant_state.bitsandbytes__fp4": "model-00004-of-00013.safetensors",
 
223
  "model.layers.15.self_attn.k_proj.weight": "model-00004-of-00013.safetensors",
224
  "model.layers.15.self_attn.k_proj.weight.absmax": "model-00004-of-00013.safetensors",
225
  "model.layers.15.self_attn.k_proj.weight.quant_map": "model-00004-of-00013.safetensors",
@@ -228,6 +243,7 @@
228
  "model.layers.15.self_attn.o_proj.weight.absmax": "model-00004-of-00013.safetensors",
229
  "model.layers.15.self_attn.o_proj.weight.quant_map": "model-00004-of-00013.safetensors",
230
  "model.layers.15.self_attn.o_proj.weight.quant_state.bitsandbytes__fp4": "model-00004-of-00013.safetensors",
 
231
  "model.layers.15.self_attn.q_proj.weight": "model-00004-of-00013.safetensors",
232
  "model.layers.15.self_attn.q_proj.weight.absmax": "model-00004-of-00013.safetensors",
233
  "model.layers.15.self_attn.q_proj.weight.quant_map": "model-00004-of-00013.safetensors",
@@ -249,6 +265,7 @@
249
  "model.layers.16.mlp.up_proj.weight.absmax": "model-00004-of-00013.safetensors",
250
  "model.layers.16.mlp.up_proj.weight.quant_map": "model-00004-of-00013.safetensors",
251
  "model.layers.16.mlp.up_proj.weight.quant_state.bitsandbytes__fp4": "model-00004-of-00013.safetensors",
 
252
  "model.layers.16.self_attn.k_proj.weight": "model-00004-of-00013.safetensors",
253
  "model.layers.16.self_attn.k_proj.weight.absmax": "model-00004-of-00013.safetensors",
254
  "model.layers.16.self_attn.k_proj.weight.quant_map": "model-00004-of-00013.safetensors",
@@ -257,6 +274,7 @@
257
  "model.layers.16.self_attn.o_proj.weight.absmax": "model-00004-of-00013.safetensors",
258
  "model.layers.16.self_attn.o_proj.weight.quant_map": "model-00004-of-00013.safetensors",
259
  "model.layers.16.self_attn.o_proj.weight.quant_state.bitsandbytes__fp4": "model-00004-of-00013.safetensors",
 
260
  "model.layers.16.self_attn.q_proj.weight": "model-00004-of-00013.safetensors",
261
  "model.layers.16.self_attn.q_proj.weight.absmax": "model-00004-of-00013.safetensors",
262
  "model.layers.16.self_attn.q_proj.weight.quant_map": "model-00004-of-00013.safetensors",
@@ -278,6 +296,7 @@
278
  "model.layers.17.mlp.up_proj.weight.absmax": "model-00005-of-00013.safetensors",
279
  "model.layers.17.mlp.up_proj.weight.quant_map": "model-00005-of-00013.safetensors",
280
  "model.layers.17.mlp.up_proj.weight.quant_state.bitsandbytes__fp4": "model-00005-of-00013.safetensors",
 
281
  "model.layers.17.self_attn.k_proj.weight": "model-00005-of-00013.safetensors",
282
  "model.layers.17.self_attn.k_proj.weight.absmax": "model-00005-of-00013.safetensors",
283
  "model.layers.17.self_attn.k_proj.weight.quant_map": "model-00005-of-00013.safetensors",
@@ -286,6 +305,7 @@
286
  "model.layers.17.self_attn.o_proj.weight.absmax": "model-00005-of-00013.safetensors",
287
  "model.layers.17.self_attn.o_proj.weight.quant_map": "model-00005-of-00013.safetensors",
288
  "model.layers.17.self_attn.o_proj.weight.quant_state.bitsandbytes__fp4": "model-00005-of-00013.safetensors",
 
289
  "model.layers.17.self_attn.q_proj.weight": "model-00005-of-00013.safetensors",
290
  "model.layers.17.self_attn.q_proj.weight.absmax": "model-00005-of-00013.safetensors",
291
  "model.layers.17.self_attn.q_proj.weight.quant_map": "model-00005-of-00013.safetensors",
@@ -307,6 +327,7 @@
307
  "model.layers.18.mlp.up_proj.weight.absmax": "model-00005-of-00013.safetensors",
308
  "model.layers.18.mlp.up_proj.weight.quant_map": "model-00005-of-00013.safetensors",
309
  "model.layers.18.mlp.up_proj.weight.quant_state.bitsandbytes__fp4": "model-00005-of-00013.safetensors",
 
310
  "model.layers.18.self_attn.k_proj.weight": "model-00005-of-00013.safetensors",
311
  "model.layers.18.self_attn.k_proj.weight.absmax": "model-00005-of-00013.safetensors",
312
  "model.layers.18.self_attn.k_proj.weight.quant_map": "model-00005-of-00013.safetensors",
@@ -315,6 +336,7 @@
315
  "model.layers.18.self_attn.o_proj.weight.absmax": "model-00005-of-00013.safetensors",
316
  "model.layers.18.self_attn.o_proj.weight.quant_map": "model-00005-of-00013.safetensors",
317
  "model.layers.18.self_attn.o_proj.weight.quant_state.bitsandbytes__fp4": "model-00005-of-00013.safetensors",
 
318
  "model.layers.18.self_attn.q_proj.weight": "model-00005-of-00013.safetensors",
319
  "model.layers.18.self_attn.q_proj.weight.absmax": "model-00005-of-00013.safetensors",
320
  "model.layers.18.self_attn.q_proj.weight.quant_map": "model-00005-of-00013.safetensors",
@@ -336,6 +358,7 @@
336
  "model.layers.19.mlp.up_proj.weight.absmax": "model-00005-of-00013.safetensors",
337
  "model.layers.19.mlp.up_proj.weight.quant_map": "model-00005-of-00013.safetensors",
338
  "model.layers.19.mlp.up_proj.weight.quant_state.bitsandbytes__fp4": "model-00005-of-00013.safetensors",
 
339
  "model.layers.19.self_attn.k_proj.weight": "model-00005-of-00013.safetensors",
340
  "model.layers.19.self_attn.k_proj.weight.absmax": "model-00005-of-00013.safetensors",
341
  "model.layers.19.self_attn.k_proj.weight.quant_map": "model-00005-of-00013.safetensors",
@@ -344,6 +367,7 @@
344
  "model.layers.19.self_attn.o_proj.weight.absmax": "model-00005-of-00013.safetensors",
345
  "model.layers.19.self_attn.o_proj.weight.quant_map": "model-00005-of-00013.safetensors",
346
  "model.layers.19.self_attn.o_proj.weight.quant_state.bitsandbytes__fp4": "model-00005-of-00013.safetensors",
 
347
  "model.layers.19.self_attn.q_proj.weight": "model-00005-of-00013.safetensors",
348
  "model.layers.19.self_attn.q_proj.weight.absmax": "model-00005-of-00013.safetensors",
349
  "model.layers.19.self_attn.q_proj.weight.quant_map": "model-00005-of-00013.safetensors",
@@ -365,6 +389,7 @@
365
  "model.layers.2.mlp.up_proj.weight.absmax": "model-00002-of-00013.safetensors",
366
  "model.layers.2.mlp.up_proj.weight.quant_map": "model-00002-of-00013.safetensors",
367
  "model.layers.2.mlp.up_proj.weight.quant_state.bitsandbytes__fp4": "model-00002-of-00013.safetensors",
 
368
  "model.layers.2.self_attn.k_proj.weight": "model-00002-of-00013.safetensors",
369
  "model.layers.2.self_attn.k_proj.weight.absmax": "model-00002-of-00013.safetensors",
370
  "model.layers.2.self_attn.k_proj.weight.quant_map": "model-00002-of-00013.safetensors",
@@ -373,6 +398,7 @@
373
  "model.layers.2.self_attn.o_proj.weight.absmax": "model-00002-of-00013.safetensors",
374
  "model.layers.2.self_attn.o_proj.weight.quant_map": "model-00002-of-00013.safetensors",
375
  "model.layers.2.self_attn.o_proj.weight.quant_state.bitsandbytes__fp4": "model-00002-of-00013.safetensors",
 
376
  "model.layers.2.self_attn.q_proj.weight": "model-00002-of-00013.safetensors",
377
  "model.layers.2.self_attn.q_proj.weight.absmax": "model-00002-of-00013.safetensors",
378
  "model.layers.2.self_attn.q_proj.weight.quant_map": "model-00002-of-00013.safetensors",
@@ -394,6 +420,7 @@
394
  "model.layers.20.mlp.up_proj.weight.absmax": "model-00005-of-00013.safetensors",
395
  "model.layers.20.mlp.up_proj.weight.quant_map": "model-00005-of-00013.safetensors",
396
  "model.layers.20.mlp.up_proj.weight.quant_state.bitsandbytes__fp4": "model-00005-of-00013.safetensors",
 
397
  "model.layers.20.self_attn.k_proj.weight": "model-00005-of-00013.safetensors",
398
  "model.layers.20.self_attn.k_proj.weight.absmax": "model-00005-of-00013.safetensors",
399
  "model.layers.20.self_attn.k_proj.weight.quant_map": "model-00005-of-00013.safetensors",
@@ -402,6 +429,7 @@
402
  "model.layers.20.self_attn.o_proj.weight.absmax": "model-00005-of-00013.safetensors",
403
  "model.layers.20.self_attn.o_proj.weight.quant_map": "model-00005-of-00013.safetensors",
404
  "model.layers.20.self_attn.o_proj.weight.quant_state.bitsandbytes__fp4": "model-00005-of-00013.safetensors",
 
405
  "model.layers.20.self_attn.q_proj.weight": "model-00005-of-00013.safetensors",
406
  "model.layers.20.self_attn.q_proj.weight.absmax": "model-00005-of-00013.safetensors",
407
  "model.layers.20.self_attn.q_proj.weight.quant_map": "model-00005-of-00013.safetensors",
@@ -423,6 +451,7 @@
423
  "model.layers.21.mlp.up_proj.weight.absmax": "model-00005-of-00013.safetensors",
424
  "model.layers.21.mlp.up_proj.weight.quant_map": "model-00005-of-00013.safetensors",
425
  "model.layers.21.mlp.up_proj.weight.quant_state.bitsandbytes__fp4": "model-00005-of-00013.safetensors",
 
426
  "model.layers.21.self_attn.k_proj.weight": "model-00005-of-00013.safetensors",
427
  "model.layers.21.self_attn.k_proj.weight.absmax": "model-00005-of-00013.safetensors",
428
  "model.layers.21.self_attn.k_proj.weight.quant_map": "model-00005-of-00013.safetensors",
@@ -431,6 +460,7 @@
431
  "model.layers.21.self_attn.o_proj.weight.absmax": "model-00005-of-00013.safetensors",
432
  "model.layers.21.self_attn.o_proj.weight.quant_map": "model-00005-of-00013.safetensors",
433
  "model.layers.21.self_attn.o_proj.weight.quant_state.bitsandbytes__fp4": "model-00005-of-00013.safetensors",
 
434
  "model.layers.21.self_attn.q_proj.weight": "model-00005-of-00013.safetensors",
435
  "model.layers.21.self_attn.q_proj.weight.absmax": "model-00005-of-00013.safetensors",
436
  "model.layers.21.self_attn.q_proj.weight.quant_map": "model-00005-of-00013.safetensors",
@@ -452,6 +482,7 @@
452
  "model.layers.22.mlp.up_proj.weight.absmax": "model-00006-of-00013.safetensors",
453
  "model.layers.22.mlp.up_proj.weight.quant_map": "model-00006-of-00013.safetensors",
454
  "model.layers.22.mlp.up_proj.weight.quant_state.bitsandbytes__fp4": "model-00006-of-00013.safetensors",
 
455
  "model.layers.22.self_attn.k_proj.weight": "model-00005-of-00013.safetensors",
456
  "model.layers.22.self_attn.k_proj.weight.absmax": "model-00005-of-00013.safetensors",
457
  "model.layers.22.self_attn.k_proj.weight.quant_map": "model-00005-of-00013.safetensors",
@@ -460,6 +491,7 @@
460
  "model.layers.22.self_attn.o_proj.weight.absmax": "model-00005-of-00013.safetensors",
461
  "model.layers.22.self_attn.o_proj.weight.quant_map": "model-00005-of-00013.safetensors",
462
  "model.layers.22.self_attn.o_proj.weight.quant_state.bitsandbytes__fp4": "model-00005-of-00013.safetensors",
 
463
  "model.layers.22.self_attn.q_proj.weight": "model-00005-of-00013.safetensors",
464
  "model.layers.22.self_attn.q_proj.weight.absmax": "model-00005-of-00013.safetensors",
465
  "model.layers.22.self_attn.q_proj.weight.quant_map": "model-00005-of-00013.safetensors",
@@ -481,6 +513,7 @@
481
  "model.layers.23.mlp.up_proj.weight.absmax": "model-00006-of-00013.safetensors",
482
  "model.layers.23.mlp.up_proj.weight.quant_map": "model-00006-of-00013.safetensors",
483
  "model.layers.23.mlp.up_proj.weight.quant_state.bitsandbytes__fp4": "model-00006-of-00013.safetensors",
 
484
  "model.layers.23.self_attn.k_proj.weight": "model-00006-of-00013.safetensors",
485
  "model.layers.23.self_attn.k_proj.weight.absmax": "model-00006-of-00013.safetensors",
486
  "model.layers.23.self_attn.k_proj.weight.quant_map": "model-00006-of-00013.safetensors",
@@ -489,6 +522,7 @@
489
  "model.layers.23.self_attn.o_proj.weight.absmax": "model-00006-of-00013.safetensors",
490
  "model.layers.23.self_attn.o_proj.weight.quant_map": "model-00006-of-00013.safetensors",
491
  "model.layers.23.self_attn.o_proj.weight.quant_state.bitsandbytes__fp4": "model-00006-of-00013.safetensors",
 
492
  "model.layers.23.self_attn.q_proj.weight": "model-00006-of-00013.safetensors",
493
  "model.layers.23.self_attn.q_proj.weight.absmax": "model-00006-of-00013.safetensors",
494
  "model.layers.23.self_attn.q_proj.weight.quant_map": "model-00006-of-00013.safetensors",
@@ -510,6 +544,7 @@
510
  "model.layers.24.mlp.up_proj.weight.absmax": "model-00006-of-00013.safetensors",
511
  "model.layers.24.mlp.up_proj.weight.quant_map": "model-00006-of-00013.safetensors",
512
  "model.layers.24.mlp.up_proj.weight.quant_state.bitsandbytes__fp4": "model-00006-of-00013.safetensors",
 
513
  "model.layers.24.self_attn.k_proj.weight": "model-00006-of-00013.safetensors",
514
  "model.layers.24.self_attn.k_proj.weight.absmax": "model-00006-of-00013.safetensors",
515
  "model.layers.24.self_attn.k_proj.weight.quant_map": "model-00006-of-00013.safetensors",
@@ -518,6 +553,7 @@
518
  "model.layers.24.self_attn.o_proj.weight.absmax": "model-00006-of-00013.safetensors",
519
  "model.layers.24.self_attn.o_proj.weight.quant_map": "model-00006-of-00013.safetensors",
520
  "model.layers.24.self_attn.o_proj.weight.quant_state.bitsandbytes__fp4": "model-00006-of-00013.safetensors",
 
521
  "model.layers.24.self_attn.q_proj.weight": "model-00006-of-00013.safetensors",
522
  "model.layers.24.self_attn.q_proj.weight.absmax": "model-00006-of-00013.safetensors",
523
  "model.layers.24.self_attn.q_proj.weight.quant_map": "model-00006-of-00013.safetensors",
@@ -539,6 +575,7 @@
539
  "model.layers.25.mlp.up_proj.weight.absmax": "model-00006-of-00013.safetensors",
540
  "model.layers.25.mlp.up_proj.weight.quant_map": "model-00006-of-00013.safetensors",
541
  "model.layers.25.mlp.up_proj.weight.quant_state.bitsandbytes__fp4": "model-00006-of-00013.safetensors",
 
542
  "model.layers.25.self_attn.k_proj.weight": "model-00006-of-00013.safetensors",
543
  "model.layers.25.self_attn.k_proj.weight.absmax": "model-00006-of-00013.safetensors",
544
  "model.layers.25.self_attn.k_proj.weight.quant_map": "model-00006-of-00013.safetensors",
@@ -547,6 +584,7 @@
547
  "model.layers.25.self_attn.o_proj.weight.absmax": "model-00006-of-00013.safetensors",
548
  "model.layers.25.self_attn.o_proj.weight.quant_map": "model-00006-of-00013.safetensors",
549
  "model.layers.25.self_attn.o_proj.weight.quant_state.bitsandbytes__fp4": "model-00006-of-00013.safetensors",
 
550
  "model.layers.25.self_attn.q_proj.weight": "model-00006-of-00013.safetensors",
551
  "model.layers.25.self_attn.q_proj.weight.absmax": "model-00006-of-00013.safetensors",
552
  "model.layers.25.self_attn.q_proj.weight.quant_map": "model-00006-of-00013.safetensors",
@@ -568,6 +606,7 @@
568
  "model.layers.26.mlp.up_proj.weight.absmax": "model-00006-of-00013.safetensors",
569
  "model.layers.26.mlp.up_proj.weight.quant_map": "model-00006-of-00013.safetensors",
570
  "model.layers.26.mlp.up_proj.weight.quant_state.bitsandbytes__fp4": "model-00006-of-00013.safetensors",
 
571
  "model.layers.26.self_attn.k_proj.weight": "model-00006-of-00013.safetensors",
572
  "model.layers.26.self_attn.k_proj.weight.absmax": "model-00006-of-00013.safetensors",
573
  "model.layers.26.self_attn.k_proj.weight.quant_map": "model-00006-of-00013.safetensors",
@@ -576,6 +615,7 @@
576
  "model.layers.26.self_attn.o_proj.weight.absmax": "model-00006-of-00013.safetensors",
577
  "model.layers.26.self_attn.o_proj.weight.quant_map": "model-00006-of-00013.safetensors",
578
  "model.layers.26.self_attn.o_proj.weight.quant_state.bitsandbytes__fp4": "model-00006-of-00013.safetensors",
 
579
  "model.layers.26.self_attn.q_proj.weight": "model-00006-of-00013.safetensors",
580
  "model.layers.26.self_attn.q_proj.weight.absmax": "model-00006-of-00013.safetensors",
581
  "model.layers.26.self_attn.q_proj.weight.quant_map": "model-00006-of-00013.safetensors",
@@ -597,6 +637,7 @@
597
  "model.layers.27.mlp.up_proj.weight.absmax": "model-00006-of-00013.safetensors",
598
  "model.layers.27.mlp.up_proj.weight.quant_map": "model-00006-of-00013.safetensors",
599
  "model.layers.27.mlp.up_proj.weight.quant_state.bitsandbytes__fp4": "model-00006-of-00013.safetensors",
 
600
  "model.layers.27.self_attn.k_proj.weight": "model-00006-of-00013.safetensors",
601
  "model.layers.27.self_attn.k_proj.weight.absmax": "model-00006-of-00013.safetensors",
602
  "model.layers.27.self_attn.k_proj.weight.quant_map": "model-00006-of-00013.safetensors",
@@ -605,6 +646,7 @@
605
  "model.layers.27.self_attn.o_proj.weight.absmax": "model-00006-of-00013.safetensors",
606
  "model.layers.27.self_attn.o_proj.weight.quant_map": "model-00006-of-00013.safetensors",
607
  "model.layers.27.self_attn.o_proj.weight.quant_state.bitsandbytes__fp4": "model-00006-of-00013.safetensors",
 
608
  "model.layers.27.self_attn.q_proj.weight": "model-00006-of-00013.safetensors",
609
  "model.layers.27.self_attn.q_proj.weight.absmax": "model-00006-of-00013.safetensors",
610
  "model.layers.27.self_attn.q_proj.weight.quant_map": "model-00006-of-00013.safetensors",
@@ -626,6 +668,7 @@
626
  "model.layers.28.mlp.up_proj.weight.absmax": "model-00007-of-00013.safetensors",
627
  "model.layers.28.mlp.up_proj.weight.quant_map": "model-00007-of-00013.safetensors",
628
  "model.layers.28.mlp.up_proj.weight.quant_state.bitsandbytes__fp4": "model-00007-of-00013.safetensors",
 
629
  "model.layers.28.self_attn.k_proj.weight": "model-00007-of-00013.safetensors",
630
  "model.layers.28.self_attn.k_proj.weight.absmax": "model-00007-of-00013.safetensors",
631
  "model.layers.28.self_attn.k_proj.weight.quant_map": "model-00007-of-00013.safetensors",
@@ -634,6 +677,7 @@
634
  "model.layers.28.self_attn.o_proj.weight.absmax": "model-00007-of-00013.safetensors",
635
  "model.layers.28.self_attn.o_proj.weight.quant_map": "model-00007-of-00013.safetensors",
636
  "model.layers.28.self_attn.o_proj.weight.quant_state.bitsandbytes__fp4": "model-00007-of-00013.safetensors",
 
637
  "model.layers.28.self_attn.q_proj.weight": "model-00007-of-00013.safetensors",
638
  "model.layers.28.self_attn.q_proj.weight.absmax": "model-00007-of-00013.safetensors",
639
  "model.layers.28.self_attn.q_proj.weight.quant_map": "model-00007-of-00013.safetensors",
@@ -655,6 +699,7 @@
655
  "model.layers.29.mlp.up_proj.weight.absmax": "model-00007-of-00013.safetensors",
656
  "model.layers.29.mlp.up_proj.weight.quant_map": "model-00007-of-00013.safetensors",
657
  "model.layers.29.mlp.up_proj.weight.quant_state.bitsandbytes__fp4": "model-00007-of-00013.safetensors",
 
658
  "model.layers.29.self_attn.k_proj.weight": "model-00007-of-00013.safetensors",
659
  "model.layers.29.self_attn.k_proj.weight.absmax": "model-00007-of-00013.safetensors",
660
  "model.layers.29.self_attn.k_proj.weight.quant_map": "model-00007-of-00013.safetensors",
@@ -663,6 +708,7 @@
663
  "model.layers.29.self_attn.o_proj.weight.absmax": "model-00007-of-00013.safetensors",
664
  "model.layers.29.self_attn.o_proj.weight.quant_map": "model-00007-of-00013.safetensors",
665
  "model.layers.29.self_attn.o_proj.weight.quant_state.bitsandbytes__fp4": "model-00007-of-00013.safetensors",
 
666
  "model.layers.29.self_attn.q_proj.weight": "model-00007-of-00013.safetensors",
667
  "model.layers.29.self_attn.q_proj.weight.absmax": "model-00007-of-00013.safetensors",
668
  "model.layers.29.self_attn.q_proj.weight.quant_map": "model-00007-of-00013.safetensors",
@@ -684,6 +730,7 @@
684
  "model.layers.3.mlp.up_proj.weight.absmax": "model-00002-of-00013.safetensors",
685
  "model.layers.3.mlp.up_proj.weight.quant_map": "model-00002-of-00013.safetensors",
686
  "model.layers.3.mlp.up_proj.weight.quant_state.bitsandbytes__fp4": "model-00002-of-00013.safetensors",
 
687
  "model.layers.3.self_attn.k_proj.weight": "model-00002-of-00013.safetensors",
688
  "model.layers.3.self_attn.k_proj.weight.absmax": "model-00002-of-00013.safetensors",
689
  "model.layers.3.self_attn.k_proj.weight.quant_map": "model-00002-of-00013.safetensors",
@@ -692,6 +739,7 @@
692
  "model.layers.3.self_attn.o_proj.weight.absmax": "model-00002-of-00013.safetensors",
693
  "model.layers.3.self_attn.o_proj.weight.quant_map": "model-00002-of-00013.safetensors",
694
  "model.layers.3.self_attn.o_proj.weight.quant_state.bitsandbytes__fp4": "model-00002-of-00013.safetensors",
 
695
  "model.layers.3.self_attn.q_proj.weight": "model-00002-of-00013.safetensors",
696
  "model.layers.3.self_attn.q_proj.weight.absmax": "model-00002-of-00013.safetensors",
697
  "model.layers.3.self_attn.q_proj.weight.quant_map": "model-00002-of-00013.safetensors",
@@ -713,6 +761,7 @@
713
  "model.layers.30.mlp.up_proj.weight.absmax": "model-00007-of-00013.safetensors",
714
  "model.layers.30.mlp.up_proj.weight.quant_map": "model-00007-of-00013.safetensors",
715
  "model.layers.30.mlp.up_proj.weight.quant_state.bitsandbytes__fp4": "model-00007-of-00013.safetensors",
 
716
  "model.layers.30.self_attn.k_proj.weight": "model-00007-of-00013.safetensors",
717
  "model.layers.30.self_attn.k_proj.weight.absmax": "model-00007-of-00013.safetensors",
718
  "model.layers.30.self_attn.k_proj.weight.quant_map": "model-00007-of-00013.safetensors",
@@ -721,6 +770,7 @@
721
  "model.layers.30.self_attn.o_proj.weight.absmax": "model-00007-of-00013.safetensors",
722
  "model.layers.30.self_attn.o_proj.weight.quant_map": "model-00007-of-00013.safetensors",
723
  "model.layers.30.self_attn.o_proj.weight.quant_state.bitsandbytes__fp4": "model-00007-of-00013.safetensors",
 
724
  "model.layers.30.self_attn.q_proj.weight": "model-00007-of-00013.safetensors",
725
  "model.layers.30.self_attn.q_proj.weight.absmax": "model-00007-of-00013.safetensors",
726
  "model.layers.30.self_attn.q_proj.weight.quant_map": "model-00007-of-00013.safetensors",
@@ -742,6 +792,7 @@
742
  "model.layers.31.mlp.up_proj.weight.absmax": "model-00007-of-00013.safetensors",
743
  "model.layers.31.mlp.up_proj.weight.quant_map": "model-00007-of-00013.safetensors",
744
  "model.layers.31.mlp.up_proj.weight.quant_state.bitsandbytes__fp4": "model-00007-of-00013.safetensors",
 
745
  "model.layers.31.self_attn.k_proj.weight": "model-00007-of-00013.safetensors",
746
  "model.layers.31.self_attn.k_proj.weight.absmax": "model-00007-of-00013.safetensors",
747
  "model.layers.31.self_attn.k_proj.weight.quant_map": "model-00007-of-00013.safetensors",
@@ -750,6 +801,7 @@
750
  "model.layers.31.self_attn.o_proj.weight.absmax": "model-00007-of-00013.safetensors",
751
  "model.layers.31.self_attn.o_proj.weight.quant_map": "model-00007-of-00013.safetensors",
752
  "model.layers.31.self_attn.o_proj.weight.quant_state.bitsandbytes__fp4": "model-00007-of-00013.safetensors",
 
753
  "model.layers.31.self_attn.q_proj.weight": "model-00007-of-00013.safetensors",
754
  "model.layers.31.self_attn.q_proj.weight.absmax": "model-00007-of-00013.safetensors",
755
  "model.layers.31.self_attn.q_proj.weight.quant_map": "model-00007-of-00013.safetensors",
@@ -771,6 +823,7 @@
771
  "model.layers.32.mlp.up_proj.weight.absmax": "model-00007-of-00013.safetensors",
772
  "model.layers.32.mlp.up_proj.weight.quant_map": "model-00007-of-00013.safetensors",
773
  "model.layers.32.mlp.up_proj.weight.quant_state.bitsandbytes__fp4": "model-00007-of-00013.safetensors",
 
774
  "model.layers.32.self_attn.k_proj.weight": "model-00007-of-00013.safetensors",
775
  "model.layers.32.self_attn.k_proj.weight.absmax": "model-00007-of-00013.safetensors",
776
  "model.layers.32.self_attn.k_proj.weight.quant_map": "model-00007-of-00013.safetensors",
@@ -779,6 +832,7 @@
779
  "model.layers.32.self_attn.o_proj.weight.absmax": "model-00007-of-00013.safetensors",
780
  "model.layers.32.self_attn.o_proj.weight.quant_map": "model-00007-of-00013.safetensors",
781
  "model.layers.32.self_attn.o_proj.weight.quant_state.bitsandbytes__fp4": "model-00007-of-00013.safetensors",
 
782
  "model.layers.32.self_attn.q_proj.weight": "model-00007-of-00013.safetensors",
783
  "model.layers.32.self_attn.q_proj.weight.absmax": "model-00007-of-00013.safetensors",
784
  "model.layers.32.self_attn.q_proj.weight.quant_map": "model-00007-of-00013.safetensors",
@@ -800,6 +854,7 @@
800
  "model.layers.33.mlp.up_proj.weight.absmax": "model-00008-of-00013.safetensors",
801
  "model.layers.33.mlp.up_proj.weight.quant_map": "model-00008-of-00013.safetensors",
802
  "model.layers.33.mlp.up_proj.weight.quant_state.bitsandbytes__fp4": "model-00008-of-00013.safetensors",
 
803
  "model.layers.33.self_attn.k_proj.weight": "model-00007-of-00013.safetensors",
804
  "model.layers.33.self_attn.k_proj.weight.absmax": "model-00007-of-00013.safetensors",
805
  "model.layers.33.self_attn.k_proj.weight.quant_map": "model-00007-of-00013.safetensors",
@@ -808,6 +863,7 @@
808
  "model.layers.33.self_attn.o_proj.weight.absmax": "model-00007-of-00013.safetensors",
809
  "model.layers.33.self_attn.o_proj.weight.quant_map": "model-00007-of-00013.safetensors",
810
  "model.layers.33.self_attn.o_proj.weight.quant_state.bitsandbytes__fp4": "model-00007-of-00013.safetensors",
 
811
  "model.layers.33.self_attn.q_proj.weight": "model-00007-of-00013.safetensors",
812
  "model.layers.33.self_attn.q_proj.weight.absmax": "model-00007-of-00013.safetensors",
813
  "model.layers.33.self_attn.q_proj.weight.quant_map": "model-00007-of-00013.safetensors",
@@ -829,6 +885,7 @@
829
  "model.layers.34.mlp.up_proj.weight.absmax": "model-00008-of-00013.safetensors",
830
  "model.layers.34.mlp.up_proj.weight.quant_map": "model-00008-of-00013.safetensors",
831
  "model.layers.34.mlp.up_proj.weight.quant_state.bitsandbytes__fp4": "model-00008-of-00013.safetensors",
 
832
  "model.layers.34.self_attn.k_proj.weight": "model-00008-of-00013.safetensors",
833
  "model.layers.34.self_attn.k_proj.weight.absmax": "model-00008-of-00013.safetensors",
834
  "model.layers.34.self_attn.k_proj.weight.quant_map": "model-00008-of-00013.safetensors",
@@ -837,6 +894,7 @@
837
  "model.layers.34.self_attn.o_proj.weight.absmax": "model-00008-of-00013.safetensors",
838
  "model.layers.34.self_attn.o_proj.weight.quant_map": "model-00008-of-00013.safetensors",
839
  "model.layers.34.self_attn.o_proj.weight.quant_state.bitsandbytes__fp4": "model-00008-of-00013.safetensors",
 
840
  "model.layers.34.self_attn.q_proj.weight": "model-00008-of-00013.safetensors",
841
  "model.layers.34.self_attn.q_proj.weight.absmax": "model-00008-of-00013.safetensors",
842
  "model.layers.34.self_attn.q_proj.weight.quant_map": "model-00008-of-00013.safetensors",
@@ -858,6 +916,7 @@
858
  "model.layers.35.mlp.up_proj.weight.absmax": "model-00008-of-00013.safetensors",
859
  "model.layers.35.mlp.up_proj.weight.quant_map": "model-00008-of-00013.safetensors",
860
  "model.layers.35.mlp.up_proj.weight.quant_state.bitsandbytes__fp4": "model-00008-of-00013.safetensors",
 
861
  "model.layers.35.self_attn.k_proj.weight": "model-00008-of-00013.safetensors",
862
  "model.layers.35.self_attn.k_proj.weight.absmax": "model-00008-of-00013.safetensors",
863
  "model.layers.35.self_attn.k_proj.weight.quant_map": "model-00008-of-00013.safetensors",
@@ -866,6 +925,7 @@
866
  "model.layers.35.self_attn.o_proj.weight.absmax": "model-00008-of-00013.safetensors",
867
  "model.layers.35.self_attn.o_proj.weight.quant_map": "model-00008-of-00013.safetensors",
868
  "model.layers.35.self_attn.o_proj.weight.quant_state.bitsandbytes__fp4": "model-00008-of-00013.safetensors",
 
869
  "model.layers.35.self_attn.q_proj.weight": "model-00008-of-00013.safetensors",
870
  "model.layers.35.self_attn.q_proj.weight.absmax": "model-00008-of-00013.safetensors",
871
  "model.layers.35.self_attn.q_proj.weight.quant_map": "model-00008-of-00013.safetensors",
@@ -887,6 +947,7 @@
887
  "model.layers.36.mlp.up_proj.weight.absmax": "model-00008-of-00013.safetensors",
888
  "model.layers.36.mlp.up_proj.weight.quant_map": "model-00008-of-00013.safetensors",
889
  "model.layers.36.mlp.up_proj.weight.quant_state.bitsandbytes__fp4": "model-00008-of-00013.safetensors",
 
890
  "model.layers.36.self_attn.k_proj.weight": "model-00008-of-00013.safetensors",
891
  "model.layers.36.self_attn.k_proj.weight.absmax": "model-00008-of-00013.safetensors",
892
  "model.layers.36.self_attn.k_proj.weight.quant_map": "model-00008-of-00013.safetensors",
@@ -895,6 +956,7 @@
895
  "model.layers.36.self_attn.o_proj.weight.absmax": "model-00008-of-00013.safetensors",
896
  "model.layers.36.self_attn.o_proj.weight.quant_map": "model-00008-of-00013.safetensors",
897
  "model.layers.36.self_attn.o_proj.weight.quant_state.bitsandbytes__fp4": "model-00008-of-00013.safetensors",
 
898
  "model.layers.36.self_attn.q_proj.weight": "model-00008-of-00013.safetensors",
899
  "model.layers.36.self_attn.q_proj.weight.absmax": "model-00008-of-00013.safetensors",
900
  "model.layers.36.self_attn.q_proj.weight.quant_map": "model-00008-of-00013.safetensors",
@@ -916,6 +978,7 @@
916
  "model.layers.37.mlp.up_proj.weight.absmax": "model-00008-of-00013.safetensors",
917
  "model.layers.37.mlp.up_proj.weight.quant_map": "model-00008-of-00013.safetensors",
918
  "model.layers.37.mlp.up_proj.weight.quant_state.bitsandbytes__fp4": "model-00008-of-00013.safetensors",
 
919
  "model.layers.37.self_attn.k_proj.weight": "model-00008-of-00013.safetensors",
920
  "model.layers.37.self_attn.k_proj.weight.absmax": "model-00008-of-00013.safetensors",
921
  "model.layers.37.self_attn.k_proj.weight.quant_map": "model-00008-of-00013.safetensors",
@@ -924,6 +987,7 @@
924
  "model.layers.37.self_attn.o_proj.weight.absmax": "model-00008-of-00013.safetensors",
925
  "model.layers.37.self_attn.o_proj.weight.quant_map": "model-00008-of-00013.safetensors",
926
  "model.layers.37.self_attn.o_proj.weight.quant_state.bitsandbytes__fp4": "model-00008-of-00013.safetensors",
 
927
  "model.layers.37.self_attn.q_proj.weight": "model-00008-of-00013.safetensors",
928
  "model.layers.37.self_attn.q_proj.weight.absmax": "model-00008-of-00013.safetensors",
929
  "model.layers.37.self_attn.q_proj.weight.quant_map": "model-00008-of-00013.safetensors",
@@ -945,6 +1009,7 @@
945
  "model.layers.38.mlp.up_proj.weight.absmax": "model-00008-of-00013.safetensors",
946
  "model.layers.38.mlp.up_proj.weight.quant_map": "model-00008-of-00013.safetensors",
947
  "model.layers.38.mlp.up_proj.weight.quant_state.bitsandbytes__fp4": "model-00008-of-00013.safetensors",
 
948
  "model.layers.38.self_attn.k_proj.weight": "model-00008-of-00013.safetensors",
949
  "model.layers.38.self_attn.k_proj.weight.absmax": "model-00008-of-00013.safetensors",
950
  "model.layers.38.self_attn.k_proj.weight.quant_map": "model-00008-of-00013.safetensors",
@@ -953,6 +1018,7 @@
953
  "model.layers.38.self_attn.o_proj.weight.absmax": "model-00008-of-00013.safetensors",
954
  "model.layers.38.self_attn.o_proj.weight.quant_map": "model-00008-of-00013.safetensors",
955
  "model.layers.38.self_attn.o_proj.weight.quant_state.bitsandbytes__fp4": "model-00008-of-00013.safetensors",
 
956
  "model.layers.38.self_attn.q_proj.weight": "model-00008-of-00013.safetensors",
957
  "model.layers.38.self_attn.q_proj.weight.absmax": "model-00008-of-00013.safetensors",
958
  "model.layers.38.self_attn.q_proj.weight.quant_map": "model-00008-of-00013.safetensors",
@@ -974,6 +1040,7 @@
974
  "model.layers.39.mlp.up_proj.weight.absmax": "model-00009-of-00013.safetensors",
975
  "model.layers.39.mlp.up_proj.weight.quant_map": "model-00009-of-00013.safetensors",
976
  "model.layers.39.mlp.up_proj.weight.quant_state.bitsandbytes__fp4": "model-00009-of-00013.safetensors",
 
977
  "model.layers.39.self_attn.k_proj.weight": "model-00009-of-00013.safetensors",
978
  "model.layers.39.self_attn.k_proj.weight.absmax": "model-00009-of-00013.safetensors",
979
  "model.layers.39.self_attn.k_proj.weight.quant_map": "model-00009-of-00013.safetensors",
@@ -982,6 +1049,7 @@
982
  "model.layers.39.self_attn.o_proj.weight.absmax": "model-00009-of-00013.safetensors",
983
  "model.layers.39.self_attn.o_proj.weight.quant_map": "model-00009-of-00013.safetensors",
984
  "model.layers.39.self_attn.o_proj.weight.quant_state.bitsandbytes__fp4": "model-00009-of-00013.safetensors",
 
985
  "model.layers.39.self_attn.q_proj.weight": "model-00009-of-00013.safetensors",
986
  "model.layers.39.self_attn.q_proj.weight.absmax": "model-00009-of-00013.safetensors",
987
  "model.layers.39.self_attn.q_proj.weight.quant_map": "model-00009-of-00013.safetensors",
@@ -1003,6 +1071,7 @@
1003
  "model.layers.4.mlp.up_proj.weight.absmax": "model-00002-of-00013.safetensors",
1004
  "model.layers.4.mlp.up_proj.weight.quant_map": "model-00002-of-00013.safetensors",
1005
  "model.layers.4.mlp.up_proj.weight.quant_state.bitsandbytes__fp4": "model-00002-of-00013.safetensors",
 
1006
  "model.layers.4.self_attn.k_proj.weight": "model-00002-of-00013.safetensors",
1007
  "model.layers.4.self_attn.k_proj.weight.absmax": "model-00002-of-00013.safetensors",
1008
  "model.layers.4.self_attn.k_proj.weight.quant_map": "model-00002-of-00013.safetensors",
@@ -1011,6 +1080,7 @@
1011
  "model.layers.4.self_attn.o_proj.weight.absmax": "model-00002-of-00013.safetensors",
1012
  "model.layers.4.self_attn.o_proj.weight.quant_map": "model-00002-of-00013.safetensors",
1013
  "model.layers.4.self_attn.o_proj.weight.quant_state.bitsandbytes__fp4": "model-00002-of-00013.safetensors",
 
1014
  "model.layers.4.self_attn.q_proj.weight": "model-00002-of-00013.safetensors",
1015
  "model.layers.4.self_attn.q_proj.weight.absmax": "model-00002-of-00013.safetensors",
1016
  "model.layers.4.self_attn.q_proj.weight.quant_map": "model-00002-of-00013.safetensors",
@@ -1032,6 +1102,7 @@
1032
  "model.layers.40.mlp.up_proj.weight.absmax": "model-00009-of-00013.safetensors",
1033
  "model.layers.40.mlp.up_proj.weight.quant_map": "model-00009-of-00013.safetensors",
1034
  "model.layers.40.mlp.up_proj.weight.quant_state.bitsandbytes__fp4": "model-00009-of-00013.safetensors",
 
1035
  "model.layers.40.self_attn.k_proj.weight": "model-00009-of-00013.safetensors",
1036
  "model.layers.40.self_attn.k_proj.weight.absmax": "model-00009-of-00013.safetensors",
1037
  "model.layers.40.self_attn.k_proj.weight.quant_map": "model-00009-of-00013.safetensors",
@@ -1040,6 +1111,7 @@
1040
  "model.layers.40.self_attn.o_proj.weight.absmax": "model-00009-of-00013.safetensors",
1041
  "model.layers.40.self_attn.o_proj.weight.quant_map": "model-00009-of-00013.safetensors",
1042
  "model.layers.40.self_attn.o_proj.weight.quant_state.bitsandbytes__fp4": "model-00009-of-00013.safetensors",
 
1043
  "model.layers.40.self_attn.q_proj.weight": "model-00009-of-00013.safetensors",
1044
  "model.layers.40.self_attn.q_proj.weight.absmax": "model-00009-of-00013.safetensors",
1045
  "model.layers.40.self_attn.q_proj.weight.quant_map": "model-00009-of-00013.safetensors",
@@ -1061,6 +1133,7 @@
1061
  "model.layers.41.mlp.up_proj.weight.absmax": "model-00009-of-00013.safetensors",
1062
  "model.layers.41.mlp.up_proj.weight.quant_map": "model-00009-of-00013.safetensors",
1063
  "model.layers.41.mlp.up_proj.weight.quant_state.bitsandbytes__fp4": "model-00009-of-00013.safetensors",
 
1064
  "model.layers.41.self_attn.k_proj.weight": "model-00009-of-00013.safetensors",
1065
  "model.layers.41.self_attn.k_proj.weight.absmax": "model-00009-of-00013.safetensors",
1066
  "model.layers.41.self_attn.k_proj.weight.quant_map": "model-00009-of-00013.safetensors",
@@ -1069,6 +1142,7 @@
1069
  "model.layers.41.self_attn.o_proj.weight.absmax": "model-00009-of-00013.safetensors",
1070
  "model.layers.41.self_attn.o_proj.weight.quant_map": "model-00009-of-00013.safetensors",
1071
  "model.layers.41.self_attn.o_proj.weight.quant_state.bitsandbytes__fp4": "model-00009-of-00013.safetensors",
 
1072
  "model.layers.41.self_attn.q_proj.weight": "model-00009-of-00013.safetensors",
1073
  "model.layers.41.self_attn.q_proj.weight.absmax": "model-00009-of-00013.safetensors",
1074
  "model.layers.41.self_attn.q_proj.weight.quant_map": "model-00009-of-00013.safetensors",
@@ -1090,6 +1164,7 @@
1090
  "model.layers.42.mlp.up_proj.weight.absmax": "model-00009-of-00013.safetensors",
1091
  "model.layers.42.mlp.up_proj.weight.quant_map": "model-00009-of-00013.safetensors",
1092
  "model.layers.42.mlp.up_proj.weight.quant_state.bitsandbytes__fp4": "model-00009-of-00013.safetensors",
 
1093
  "model.layers.42.self_attn.k_proj.weight": "model-00009-of-00013.safetensors",
1094
  "model.layers.42.self_attn.k_proj.weight.absmax": "model-00009-of-00013.safetensors",
1095
  "model.layers.42.self_attn.k_proj.weight.quant_map": "model-00009-of-00013.safetensors",
@@ -1098,6 +1173,7 @@
1098
  "model.layers.42.self_attn.o_proj.weight.absmax": "model-00009-of-00013.safetensors",
1099
  "model.layers.42.self_attn.o_proj.weight.quant_map": "model-00009-of-00013.safetensors",
1100
  "model.layers.42.self_attn.o_proj.weight.quant_state.bitsandbytes__fp4": "model-00009-of-00013.safetensors",
 
1101
  "model.layers.42.self_attn.q_proj.weight": "model-00009-of-00013.safetensors",
1102
  "model.layers.42.self_attn.q_proj.weight.absmax": "model-00009-of-00013.safetensors",
1103
  "model.layers.42.self_attn.q_proj.weight.quant_map": "model-00009-of-00013.safetensors",
@@ -1119,6 +1195,7 @@
1119
  "model.layers.43.mlp.up_proj.weight.absmax": "model-00009-of-00013.safetensors",
1120
  "model.layers.43.mlp.up_proj.weight.quant_map": "model-00009-of-00013.safetensors",
1121
  "model.layers.43.mlp.up_proj.weight.quant_state.bitsandbytes__fp4": "model-00009-of-00013.safetensors",
 
1122
  "model.layers.43.self_attn.k_proj.weight": "model-00009-of-00013.safetensors",
1123
  "model.layers.43.self_attn.k_proj.weight.absmax": "model-00009-of-00013.safetensors",
1124
  "model.layers.43.self_attn.k_proj.weight.quant_map": "model-00009-of-00013.safetensors",
@@ -1127,6 +1204,7 @@
1127
  "model.layers.43.self_attn.o_proj.weight.absmax": "model-00009-of-00013.safetensors",
1128
  "model.layers.43.self_attn.o_proj.weight.quant_map": "model-00009-of-00013.safetensors",
1129
  "model.layers.43.self_attn.o_proj.weight.quant_state.bitsandbytes__fp4": "model-00009-of-00013.safetensors",
 
1130
  "model.layers.43.self_attn.q_proj.weight": "model-00009-of-00013.safetensors",
1131
  "model.layers.43.self_attn.q_proj.weight.absmax": "model-00009-of-00013.safetensors",
1132
  "model.layers.43.self_attn.q_proj.weight.quant_map": "model-00009-of-00013.safetensors",
@@ -1148,6 +1226,7 @@
1148
  "model.layers.44.mlp.up_proj.weight.absmax": "model-00010-of-00013.safetensors",
1149
  "model.layers.44.mlp.up_proj.weight.quant_map": "model-00010-of-00013.safetensors",
1150
  "model.layers.44.mlp.up_proj.weight.quant_state.bitsandbytes__fp4": "model-00010-of-00013.safetensors",
 
1151
  "model.layers.44.self_attn.k_proj.weight": "model-00009-of-00013.safetensors",
1152
  "model.layers.44.self_attn.k_proj.weight.absmax": "model-00009-of-00013.safetensors",
1153
  "model.layers.44.self_attn.k_proj.weight.quant_map": "model-00009-of-00013.safetensors",
@@ -1156,6 +1235,7 @@
1156
  "model.layers.44.self_attn.o_proj.weight.absmax": "model-00009-of-00013.safetensors",
1157
  "model.layers.44.self_attn.o_proj.weight.quant_map": "model-00009-of-00013.safetensors",
1158
  "model.layers.44.self_attn.o_proj.weight.quant_state.bitsandbytes__fp4": "model-00009-of-00013.safetensors",
 
1159
  "model.layers.44.self_attn.q_proj.weight": "model-00009-of-00013.safetensors",
1160
  "model.layers.44.self_attn.q_proj.weight.absmax": "model-00009-of-00013.safetensors",
1161
  "model.layers.44.self_attn.q_proj.weight.quant_map": "model-00009-of-00013.safetensors",
@@ -1177,6 +1257,7 @@
1177
  "model.layers.45.mlp.up_proj.weight.absmax": "model-00010-of-00013.safetensors",
1178
  "model.layers.45.mlp.up_proj.weight.quant_map": "model-00010-of-00013.safetensors",
1179
  "model.layers.45.mlp.up_proj.weight.quant_state.bitsandbytes__fp4": "model-00010-of-00013.safetensors",
 
1180
  "model.layers.45.self_attn.k_proj.weight": "model-00010-of-00013.safetensors",
1181
  "model.layers.45.self_attn.k_proj.weight.absmax": "model-00010-of-00013.safetensors",
1182
  "model.layers.45.self_attn.k_proj.weight.quant_map": "model-00010-of-00013.safetensors",
@@ -1185,6 +1266,7 @@
1185
  "model.layers.45.self_attn.o_proj.weight.absmax": "model-00010-of-00013.safetensors",
1186
  "model.layers.45.self_attn.o_proj.weight.quant_map": "model-00010-of-00013.safetensors",
1187
  "model.layers.45.self_attn.o_proj.weight.quant_state.bitsandbytes__fp4": "model-00010-of-00013.safetensors",
 
1188
  "model.layers.45.self_attn.q_proj.weight": "model-00010-of-00013.safetensors",
1189
  "model.layers.45.self_attn.q_proj.weight.absmax": "model-00010-of-00013.safetensors",
1190
  "model.layers.45.self_attn.q_proj.weight.quant_map": "model-00010-of-00013.safetensors",
@@ -1206,6 +1288,7 @@
1206
  "model.layers.46.mlp.up_proj.weight.absmax": "model-00010-of-00013.safetensors",
1207
  "model.layers.46.mlp.up_proj.weight.quant_map": "model-00010-of-00013.safetensors",
1208
  "model.layers.46.mlp.up_proj.weight.quant_state.bitsandbytes__fp4": "model-00010-of-00013.safetensors",
 
1209
  "model.layers.46.self_attn.k_proj.weight": "model-00010-of-00013.safetensors",
1210
  "model.layers.46.self_attn.k_proj.weight.absmax": "model-00010-of-00013.safetensors",
1211
  "model.layers.46.self_attn.k_proj.weight.quant_map": "model-00010-of-00013.safetensors",
@@ -1214,6 +1297,7 @@
1214
  "model.layers.46.self_attn.o_proj.weight.absmax": "model-00010-of-00013.safetensors",
1215
  "model.layers.46.self_attn.o_proj.weight.quant_map": "model-00010-of-00013.safetensors",
1216
  "model.layers.46.self_attn.o_proj.weight.quant_state.bitsandbytes__fp4": "model-00010-of-00013.safetensors",
 
1217
  "model.layers.46.self_attn.q_proj.weight": "model-00010-of-00013.safetensors",
1218
  "model.layers.46.self_attn.q_proj.weight.absmax": "model-00010-of-00013.safetensors",
1219
  "model.layers.46.self_attn.q_proj.weight.quant_map": "model-00010-of-00013.safetensors",
@@ -1235,6 +1319,7 @@
1235
  "model.layers.47.mlp.up_proj.weight.absmax": "model-00010-of-00013.safetensors",
1236
  "model.layers.47.mlp.up_proj.weight.quant_map": "model-00010-of-00013.safetensors",
1237
  "model.layers.47.mlp.up_proj.weight.quant_state.bitsandbytes__fp4": "model-00010-of-00013.safetensors",
 
1238
  "model.layers.47.self_attn.k_proj.weight": "model-00010-of-00013.safetensors",
1239
  "model.layers.47.self_attn.k_proj.weight.absmax": "model-00010-of-00013.safetensors",
1240
  "model.layers.47.self_attn.k_proj.weight.quant_map": "model-00010-of-00013.safetensors",
@@ -1243,6 +1328,7 @@
1243
  "model.layers.47.self_attn.o_proj.weight.absmax": "model-00010-of-00013.safetensors",
1244
  "model.layers.47.self_attn.o_proj.weight.quant_map": "model-00010-of-00013.safetensors",
1245
  "model.layers.47.self_attn.o_proj.weight.quant_state.bitsandbytes__fp4": "model-00010-of-00013.safetensors",
 
1246
  "model.layers.47.self_attn.q_proj.weight": "model-00010-of-00013.safetensors",
1247
  "model.layers.47.self_attn.q_proj.weight.absmax": "model-00010-of-00013.safetensors",
1248
  "model.layers.47.self_attn.q_proj.weight.quant_map": "model-00010-of-00013.safetensors",
@@ -1264,6 +1350,7 @@
1264
  "model.layers.48.mlp.up_proj.weight.absmax": "model-00010-of-00013.safetensors",
1265
  "model.layers.48.mlp.up_proj.weight.quant_map": "model-00010-of-00013.safetensors",
1266
  "model.layers.48.mlp.up_proj.weight.quant_state.bitsandbytes__fp4": "model-00010-of-00013.safetensors",
 
1267
  "model.layers.48.self_attn.k_proj.weight": "model-00010-of-00013.safetensors",
1268
  "model.layers.48.self_attn.k_proj.weight.absmax": "model-00010-of-00013.safetensors",
1269
  "model.layers.48.self_attn.k_proj.weight.quant_map": "model-00010-of-00013.safetensors",
@@ -1272,6 +1359,7 @@
1272
  "model.layers.48.self_attn.o_proj.weight.absmax": "model-00010-of-00013.safetensors",
1273
  "model.layers.48.self_attn.o_proj.weight.quant_map": "model-00010-of-00013.safetensors",
1274
  "model.layers.48.self_attn.o_proj.weight.quant_state.bitsandbytes__fp4": "model-00010-of-00013.safetensors",
 
1275
  "model.layers.48.self_attn.q_proj.weight": "model-00010-of-00013.safetensors",
1276
  "model.layers.48.self_attn.q_proj.weight.absmax": "model-00010-of-00013.safetensors",
1277
  "model.layers.48.self_attn.q_proj.weight.quant_map": "model-00010-of-00013.safetensors",
@@ -1293,6 +1381,7 @@
1293
  "model.layers.49.mlp.up_proj.weight.absmax": "model-00010-of-00013.safetensors",
1294
  "model.layers.49.mlp.up_proj.weight.quant_map": "model-00010-of-00013.safetensors",
1295
  "model.layers.49.mlp.up_proj.weight.quant_state.bitsandbytes__fp4": "model-00010-of-00013.safetensors",
 
1296
  "model.layers.49.self_attn.k_proj.weight": "model-00010-of-00013.safetensors",
1297
  "model.layers.49.self_attn.k_proj.weight.absmax": "model-00010-of-00013.safetensors",
1298
  "model.layers.49.self_attn.k_proj.weight.quant_map": "model-00010-of-00013.safetensors",
@@ -1301,6 +1390,7 @@
1301
  "model.layers.49.self_attn.o_proj.weight.absmax": "model-00010-of-00013.safetensors",
1302
  "model.layers.49.self_attn.o_proj.weight.quant_map": "model-00010-of-00013.safetensors",
1303
  "model.layers.49.self_attn.o_proj.weight.quant_state.bitsandbytes__fp4": "model-00010-of-00013.safetensors",
 
1304
  "model.layers.49.self_attn.q_proj.weight": "model-00010-of-00013.safetensors",
1305
  "model.layers.49.self_attn.q_proj.weight.absmax": "model-00010-of-00013.safetensors",
1306
  "model.layers.49.self_attn.q_proj.weight.quant_map": "model-00010-of-00013.safetensors",
@@ -1322,6 +1412,7 @@
1322
  "model.layers.5.mlp.up_proj.weight.absmax": "model-00003-of-00013.safetensors",
1323
  "model.layers.5.mlp.up_proj.weight.quant_map": "model-00003-of-00013.safetensors",
1324
  "model.layers.5.mlp.up_proj.weight.quant_state.bitsandbytes__fp4": "model-00003-of-00013.safetensors",
 
1325
  "model.layers.5.self_attn.k_proj.weight": "model-00002-of-00013.safetensors",
1326
  "model.layers.5.self_attn.k_proj.weight.absmax": "model-00002-of-00013.safetensors",
1327
  "model.layers.5.self_attn.k_proj.weight.quant_map": "model-00002-of-00013.safetensors",
@@ -1330,6 +1421,7 @@
1330
  "model.layers.5.self_attn.o_proj.weight.absmax": "model-00002-of-00013.safetensors",
1331
  "model.layers.5.self_attn.o_proj.weight.quant_map": "model-00002-of-00013.safetensors",
1332
  "model.layers.5.self_attn.o_proj.weight.quant_state.bitsandbytes__fp4": "model-00002-of-00013.safetensors",
 
1333
  "model.layers.5.self_attn.q_proj.weight": "model-00002-of-00013.safetensors",
1334
  "model.layers.5.self_attn.q_proj.weight.absmax": "model-00002-of-00013.safetensors",
1335
  "model.layers.5.self_attn.q_proj.weight.quant_map": "model-00002-of-00013.safetensors",
@@ -1351,6 +1443,7 @@
1351
  "model.layers.50.mlp.up_proj.weight.absmax": "model-00011-of-00013.safetensors",
1352
  "model.layers.50.mlp.up_proj.weight.quant_map": "model-00011-of-00013.safetensors",
1353
  "model.layers.50.mlp.up_proj.weight.quant_state.bitsandbytes__fp4": "model-00011-of-00013.safetensors",
 
1354
  "model.layers.50.self_attn.k_proj.weight": "model-00011-of-00013.safetensors",
1355
  "model.layers.50.self_attn.k_proj.weight.absmax": "model-00011-of-00013.safetensors",
1356
  "model.layers.50.self_attn.k_proj.weight.quant_map": "model-00011-of-00013.safetensors",
@@ -1359,6 +1452,7 @@
1359
  "model.layers.50.self_attn.o_proj.weight.absmax": "model-00011-of-00013.safetensors",
1360
  "model.layers.50.self_attn.o_proj.weight.quant_map": "model-00011-of-00013.safetensors",
1361
  "model.layers.50.self_attn.o_proj.weight.quant_state.bitsandbytes__fp4": "model-00011-of-00013.safetensors",
 
1362
  "model.layers.50.self_attn.q_proj.weight": "model-00011-of-00013.safetensors",
1363
  "model.layers.50.self_attn.q_proj.weight.absmax": "model-00011-of-00013.safetensors",
1364
  "model.layers.50.self_attn.q_proj.weight.quant_map": "model-00011-of-00013.safetensors",
@@ -1380,6 +1474,7 @@
1380
  "model.layers.51.mlp.up_proj.weight.absmax": "model-00011-of-00013.safetensors",
1381
  "model.layers.51.mlp.up_proj.weight.quant_map": "model-00011-of-00013.safetensors",
1382
  "model.layers.51.mlp.up_proj.weight.quant_state.bitsandbytes__fp4": "model-00011-of-00013.safetensors",
 
1383
  "model.layers.51.self_attn.k_proj.weight": "model-00011-of-00013.safetensors",
1384
  "model.layers.51.self_attn.k_proj.weight.absmax": "model-00011-of-00013.safetensors",
1385
  "model.layers.51.self_attn.k_proj.weight.quant_map": "model-00011-of-00013.safetensors",
@@ -1388,6 +1483,7 @@
1388
  "model.layers.51.self_attn.o_proj.weight.absmax": "model-00011-of-00013.safetensors",
1389
  "model.layers.51.self_attn.o_proj.weight.quant_map": "model-00011-of-00013.safetensors",
1390
  "model.layers.51.self_attn.o_proj.weight.quant_state.bitsandbytes__fp4": "model-00011-of-00013.safetensors",
 
1391
  "model.layers.51.self_attn.q_proj.weight": "model-00011-of-00013.safetensors",
1392
  "model.layers.51.self_attn.q_proj.weight.absmax": "model-00011-of-00013.safetensors",
1393
  "model.layers.51.self_attn.q_proj.weight.quant_map": "model-00011-of-00013.safetensors",
@@ -1409,6 +1505,7 @@
1409
  "model.layers.52.mlp.up_proj.weight.absmax": "model-00011-of-00013.safetensors",
1410
  "model.layers.52.mlp.up_proj.weight.quant_map": "model-00011-of-00013.safetensors",
1411
  "model.layers.52.mlp.up_proj.weight.quant_state.bitsandbytes__fp4": "model-00011-of-00013.safetensors",
 
1412
  "model.layers.52.self_attn.k_proj.weight": "model-00011-of-00013.safetensors",
1413
  "model.layers.52.self_attn.k_proj.weight.absmax": "model-00011-of-00013.safetensors",
1414
  "model.layers.52.self_attn.k_proj.weight.quant_map": "model-00011-of-00013.safetensors",
@@ -1417,6 +1514,7 @@
1417
  "model.layers.52.self_attn.o_proj.weight.absmax": "model-00011-of-00013.safetensors",
1418
  "model.layers.52.self_attn.o_proj.weight.quant_map": "model-00011-of-00013.safetensors",
1419
  "model.layers.52.self_attn.o_proj.weight.quant_state.bitsandbytes__fp4": "model-00011-of-00013.safetensors",
 
1420
  "model.layers.52.self_attn.q_proj.weight": "model-00011-of-00013.safetensors",
1421
  "model.layers.52.self_attn.q_proj.weight.absmax": "model-00011-of-00013.safetensors",
1422
  "model.layers.52.self_attn.q_proj.weight.quant_map": "model-00011-of-00013.safetensors",
@@ -1438,6 +1536,7 @@
1438
  "model.layers.53.mlp.up_proj.weight.absmax": "model-00011-of-00013.safetensors",
1439
  "model.layers.53.mlp.up_proj.weight.quant_map": "model-00011-of-00013.safetensors",
1440
  "model.layers.53.mlp.up_proj.weight.quant_state.bitsandbytes__fp4": "model-00011-of-00013.safetensors",
 
1441
  "model.layers.53.self_attn.k_proj.weight": "model-00011-of-00013.safetensors",
1442
  "model.layers.53.self_attn.k_proj.weight.absmax": "model-00011-of-00013.safetensors",
1443
  "model.layers.53.self_attn.k_proj.weight.quant_map": "model-00011-of-00013.safetensors",
@@ -1446,6 +1545,7 @@
1446
  "model.layers.53.self_attn.o_proj.weight.absmax": "model-00011-of-00013.safetensors",
1447
  "model.layers.53.self_attn.o_proj.weight.quant_map": "model-00011-of-00013.safetensors",
1448
  "model.layers.53.self_attn.o_proj.weight.quant_state.bitsandbytes__fp4": "model-00011-of-00013.safetensors",
 
1449
  "model.layers.53.self_attn.q_proj.weight": "model-00011-of-00013.safetensors",
1450
  "model.layers.53.self_attn.q_proj.weight.absmax": "model-00011-of-00013.safetensors",
1451
  "model.layers.53.self_attn.q_proj.weight.quant_map": "model-00011-of-00013.safetensors",
@@ -1467,6 +1567,7 @@
1467
  "model.layers.54.mlp.up_proj.weight.absmax": "model-00011-of-00013.safetensors",
1468
  "model.layers.54.mlp.up_proj.weight.quant_map": "model-00011-of-00013.safetensors",
1469
  "model.layers.54.mlp.up_proj.weight.quant_state.bitsandbytes__fp4": "model-00011-of-00013.safetensors",
 
1470
  "model.layers.54.self_attn.k_proj.weight": "model-00011-of-00013.safetensors",
1471
  "model.layers.54.self_attn.k_proj.weight.absmax": "model-00011-of-00013.safetensors",
1472
  "model.layers.54.self_attn.k_proj.weight.quant_map": "model-00011-of-00013.safetensors",
@@ -1475,6 +1576,7 @@
1475
  "model.layers.54.self_attn.o_proj.weight.absmax": "model-00011-of-00013.safetensors",
1476
  "model.layers.54.self_attn.o_proj.weight.quant_map": "model-00011-of-00013.safetensors",
1477
  "model.layers.54.self_attn.o_proj.weight.quant_state.bitsandbytes__fp4": "model-00011-of-00013.safetensors",
 
1478
  "model.layers.54.self_attn.q_proj.weight": "model-00011-of-00013.safetensors",
1479
  "model.layers.54.self_attn.q_proj.weight.absmax": "model-00011-of-00013.safetensors",
1480
  "model.layers.54.self_attn.q_proj.weight.quant_map": "model-00011-of-00013.safetensors",
@@ -1496,6 +1598,7 @@
1496
  "model.layers.55.mlp.up_proj.weight.absmax": "model-00012-of-00013.safetensors",
1497
  "model.layers.55.mlp.up_proj.weight.quant_map": "model-00012-of-00013.safetensors",
1498
  "model.layers.55.mlp.up_proj.weight.quant_state.bitsandbytes__fp4": "model-00012-of-00013.safetensors",
 
1499
  "model.layers.55.self_attn.k_proj.weight": "model-00011-of-00013.safetensors",
1500
  "model.layers.55.self_attn.k_proj.weight.absmax": "model-00011-of-00013.safetensors",
1501
  "model.layers.55.self_attn.k_proj.weight.quant_map": "model-00011-of-00013.safetensors",
@@ -1504,6 +1607,7 @@
1504
  "model.layers.55.self_attn.o_proj.weight.absmax": "model-00011-of-00013.safetensors",
1505
  "model.layers.55.self_attn.o_proj.weight.quant_map": "model-00011-of-00013.safetensors",
1506
  "model.layers.55.self_attn.o_proj.weight.quant_state.bitsandbytes__fp4": "model-00011-of-00013.safetensors",
 
1507
  "model.layers.55.self_attn.q_proj.weight": "model-00011-of-00013.safetensors",
1508
  "model.layers.55.self_attn.q_proj.weight.absmax": "model-00011-of-00013.safetensors",
1509
  "model.layers.55.self_attn.q_proj.weight.quant_map": "model-00011-of-00013.safetensors",
@@ -1525,6 +1629,7 @@
1525
  "model.layers.56.mlp.up_proj.weight.absmax": "model-00012-of-00013.safetensors",
1526
  "model.layers.56.mlp.up_proj.weight.quant_map": "model-00012-of-00013.safetensors",
1527
  "model.layers.56.mlp.up_proj.weight.quant_state.bitsandbytes__fp4": "model-00012-of-00013.safetensors",
 
1528
  "model.layers.56.self_attn.k_proj.weight": "model-00012-of-00013.safetensors",
1529
  "model.layers.56.self_attn.k_proj.weight.absmax": "model-00012-of-00013.safetensors",
1530
  "model.layers.56.self_attn.k_proj.weight.quant_map": "model-00012-of-00013.safetensors",
@@ -1533,6 +1638,7 @@
1533
  "model.layers.56.self_attn.o_proj.weight.absmax": "model-00012-of-00013.safetensors",
1534
  "model.layers.56.self_attn.o_proj.weight.quant_map": "model-00012-of-00013.safetensors",
1535
  "model.layers.56.self_attn.o_proj.weight.quant_state.bitsandbytes__fp4": "model-00012-of-00013.safetensors",
 
1536
  "model.layers.56.self_attn.q_proj.weight": "model-00012-of-00013.safetensors",
1537
  "model.layers.56.self_attn.q_proj.weight.absmax": "model-00012-of-00013.safetensors",
1538
  "model.layers.56.self_attn.q_proj.weight.quant_map": "model-00012-of-00013.safetensors",
@@ -1554,6 +1660,7 @@
1554
  "model.layers.57.mlp.up_proj.weight.absmax": "model-00012-of-00013.safetensors",
1555
  "model.layers.57.mlp.up_proj.weight.quant_map": "model-00012-of-00013.safetensors",
1556
  "model.layers.57.mlp.up_proj.weight.quant_state.bitsandbytes__fp4": "model-00012-of-00013.safetensors",
 
1557
  "model.layers.57.self_attn.k_proj.weight": "model-00012-of-00013.safetensors",
1558
  "model.layers.57.self_attn.k_proj.weight.absmax": "model-00012-of-00013.safetensors",
1559
  "model.layers.57.self_attn.k_proj.weight.quant_map": "model-00012-of-00013.safetensors",
@@ -1562,6 +1669,7 @@
1562
  "model.layers.57.self_attn.o_proj.weight.absmax": "model-00012-of-00013.safetensors",
1563
  "model.layers.57.self_attn.o_proj.weight.quant_map": "model-00012-of-00013.safetensors",
1564
  "model.layers.57.self_attn.o_proj.weight.quant_state.bitsandbytes__fp4": "model-00012-of-00013.safetensors",
 
1565
  "model.layers.57.self_attn.q_proj.weight": "model-00012-of-00013.safetensors",
1566
  "model.layers.57.self_attn.q_proj.weight.absmax": "model-00012-of-00013.safetensors",
1567
  "model.layers.57.self_attn.q_proj.weight.quant_map": "model-00012-of-00013.safetensors",
@@ -1583,6 +1691,7 @@
1583
  "model.layers.58.mlp.up_proj.weight.absmax": "model-00012-of-00013.safetensors",
1584
  "model.layers.58.mlp.up_proj.weight.quant_map": "model-00012-of-00013.safetensors",
1585
  "model.layers.58.mlp.up_proj.weight.quant_state.bitsandbytes__fp4": "model-00012-of-00013.safetensors",
 
1586
  "model.layers.58.self_attn.k_proj.weight": "model-00012-of-00013.safetensors",
1587
  "model.layers.58.self_attn.k_proj.weight.absmax": "model-00012-of-00013.safetensors",
1588
  "model.layers.58.self_attn.k_proj.weight.quant_map": "model-00012-of-00013.safetensors",
@@ -1591,6 +1700,7 @@
1591
  "model.layers.58.self_attn.o_proj.weight.absmax": "model-00012-of-00013.safetensors",
1592
  "model.layers.58.self_attn.o_proj.weight.quant_map": "model-00012-of-00013.safetensors",
1593
  "model.layers.58.self_attn.o_proj.weight.quant_state.bitsandbytes__fp4": "model-00012-of-00013.safetensors",
 
1594
  "model.layers.58.self_attn.q_proj.weight": "model-00012-of-00013.safetensors",
1595
  "model.layers.58.self_attn.q_proj.weight.absmax": "model-00012-of-00013.safetensors",
1596
  "model.layers.58.self_attn.q_proj.weight.quant_map": "model-00012-of-00013.safetensors",
@@ -1612,6 +1722,7 @@
1612
  "model.layers.59.mlp.up_proj.weight.absmax": "model-00012-of-00013.safetensors",
1613
  "model.layers.59.mlp.up_proj.weight.quant_map": "model-00012-of-00013.safetensors",
1614
  "model.layers.59.mlp.up_proj.weight.quant_state.bitsandbytes__fp4": "model-00012-of-00013.safetensors",
 
1615
  "model.layers.59.self_attn.k_proj.weight": "model-00012-of-00013.safetensors",
1616
  "model.layers.59.self_attn.k_proj.weight.absmax": "model-00012-of-00013.safetensors",
1617
  "model.layers.59.self_attn.k_proj.weight.quant_map": "model-00012-of-00013.safetensors",
@@ -1620,6 +1731,7 @@
1620
  "model.layers.59.self_attn.o_proj.weight.absmax": "model-00012-of-00013.safetensors",
1621
  "model.layers.59.self_attn.o_proj.weight.quant_map": "model-00012-of-00013.safetensors",
1622
  "model.layers.59.self_attn.o_proj.weight.quant_state.bitsandbytes__fp4": "model-00012-of-00013.safetensors",
 
1623
  "model.layers.59.self_attn.q_proj.weight": "model-00012-of-00013.safetensors",
1624
  "model.layers.59.self_attn.q_proj.weight.absmax": "model-00012-of-00013.safetensors",
1625
  "model.layers.59.self_attn.q_proj.weight.quant_map": "model-00012-of-00013.safetensors",
@@ -1641,6 +1753,7 @@
1641
  "model.layers.6.mlp.up_proj.weight.absmax": "model-00003-of-00013.safetensors",
1642
  "model.layers.6.mlp.up_proj.weight.quant_map": "model-00003-of-00013.safetensors",
1643
  "model.layers.6.mlp.up_proj.weight.quant_state.bitsandbytes__fp4": "model-00003-of-00013.safetensors",
 
1644
  "model.layers.6.self_attn.k_proj.weight": "model-00003-of-00013.safetensors",
1645
  "model.layers.6.self_attn.k_proj.weight.absmax": "model-00003-of-00013.safetensors",
1646
  "model.layers.6.self_attn.k_proj.weight.quant_map": "model-00003-of-00013.safetensors",
@@ -1649,6 +1762,7 @@
1649
  "model.layers.6.self_attn.o_proj.weight.absmax": "model-00003-of-00013.safetensors",
1650
  "model.layers.6.self_attn.o_proj.weight.quant_map": "model-00003-of-00013.safetensors",
1651
  "model.layers.6.self_attn.o_proj.weight.quant_state.bitsandbytes__fp4": "model-00003-of-00013.safetensors",
 
1652
  "model.layers.6.self_attn.q_proj.weight": "model-00003-of-00013.safetensors",
1653
  "model.layers.6.self_attn.q_proj.weight.absmax": "model-00003-of-00013.safetensors",
1654
  "model.layers.6.self_attn.q_proj.weight.quant_map": "model-00003-of-00013.safetensors",
@@ -1670,6 +1784,7 @@
1670
  "model.layers.60.mlp.up_proj.weight.absmax": "model-00012-of-00013.safetensors",
1671
  "model.layers.60.mlp.up_proj.weight.quant_map": "model-00012-of-00013.safetensors",
1672
  "model.layers.60.mlp.up_proj.weight.quant_state.bitsandbytes__fp4": "model-00012-of-00013.safetensors",
 
1673
  "model.layers.60.self_attn.k_proj.weight": "model-00012-of-00013.safetensors",
1674
  "model.layers.60.self_attn.k_proj.weight.absmax": "model-00012-of-00013.safetensors",
1675
  "model.layers.60.self_attn.k_proj.weight.quant_map": "model-00012-of-00013.safetensors",
@@ -1678,6 +1793,7 @@
1678
  "model.layers.60.self_attn.o_proj.weight.absmax": "model-00012-of-00013.safetensors",
1679
  "model.layers.60.self_attn.o_proj.weight.quant_map": "model-00012-of-00013.safetensors",
1680
  "model.layers.60.self_attn.o_proj.weight.quant_state.bitsandbytes__fp4": "model-00012-of-00013.safetensors",
 
1681
  "model.layers.60.self_attn.q_proj.weight": "model-00012-of-00013.safetensors",
1682
  "model.layers.60.self_attn.q_proj.weight.absmax": "model-00012-of-00013.safetensors",
1683
  "model.layers.60.self_attn.q_proj.weight.quant_map": "model-00012-of-00013.safetensors",
@@ -1699,6 +1815,7 @@
1699
  "model.layers.61.mlp.up_proj.weight.absmax": "model-00013-of-00013.safetensors",
1700
  "model.layers.61.mlp.up_proj.weight.quant_map": "model-00013-of-00013.safetensors",
1701
  "model.layers.61.mlp.up_proj.weight.quant_state.bitsandbytes__fp4": "model-00013-of-00013.safetensors",
 
1702
  "model.layers.61.self_attn.k_proj.weight": "model-00013-of-00013.safetensors",
1703
  "model.layers.61.self_attn.k_proj.weight.absmax": "model-00013-of-00013.safetensors",
1704
  "model.layers.61.self_attn.k_proj.weight.quant_map": "model-00013-of-00013.safetensors",
@@ -1707,6 +1824,7 @@
1707
  "model.layers.61.self_attn.o_proj.weight.absmax": "model-00013-of-00013.safetensors",
1708
  "model.layers.61.self_attn.o_proj.weight.quant_map": "model-00013-of-00013.safetensors",
1709
  "model.layers.61.self_attn.o_proj.weight.quant_state.bitsandbytes__fp4": "model-00013-of-00013.safetensors",
 
1710
  "model.layers.61.self_attn.q_proj.weight": "model-00013-of-00013.safetensors",
1711
  "model.layers.61.self_attn.q_proj.weight.absmax": "model-00013-of-00013.safetensors",
1712
  "model.layers.61.self_attn.q_proj.weight.quant_map": "model-00013-of-00013.safetensors",
@@ -1728,6 +1846,7 @@
1728
  "model.layers.62.mlp.up_proj.weight.absmax": "model-00013-of-00013.safetensors",
1729
  "model.layers.62.mlp.up_proj.weight.quant_map": "model-00013-of-00013.safetensors",
1730
  "model.layers.62.mlp.up_proj.weight.quant_state.bitsandbytes__fp4": "model-00013-of-00013.safetensors",
 
1731
  "model.layers.62.self_attn.k_proj.weight": "model-00013-of-00013.safetensors",
1732
  "model.layers.62.self_attn.k_proj.weight.absmax": "model-00013-of-00013.safetensors",
1733
  "model.layers.62.self_attn.k_proj.weight.quant_map": "model-00013-of-00013.safetensors",
@@ -1736,6 +1855,7 @@
1736
  "model.layers.62.self_attn.o_proj.weight.absmax": "model-00013-of-00013.safetensors",
1737
  "model.layers.62.self_attn.o_proj.weight.quant_map": "model-00013-of-00013.safetensors",
1738
  "model.layers.62.self_attn.o_proj.weight.quant_state.bitsandbytes__fp4": "model-00013-of-00013.safetensors",
 
1739
  "model.layers.62.self_attn.q_proj.weight": "model-00013-of-00013.safetensors",
1740
  "model.layers.62.self_attn.q_proj.weight.absmax": "model-00013-of-00013.safetensors",
1741
  "model.layers.62.self_attn.q_proj.weight.quant_map": "model-00013-of-00013.safetensors",
@@ -1757,6 +1877,7 @@
1757
  "model.layers.63.mlp.up_proj.weight.absmax": "model-00013-of-00013.safetensors",
1758
  "model.layers.63.mlp.up_proj.weight.quant_map": "model-00013-of-00013.safetensors",
1759
  "model.layers.63.mlp.up_proj.weight.quant_state.bitsandbytes__fp4": "model-00013-of-00013.safetensors",
 
1760
  "model.layers.63.self_attn.k_proj.weight": "model-00013-of-00013.safetensors",
1761
  "model.layers.63.self_attn.k_proj.weight.absmax": "model-00013-of-00013.safetensors",
1762
  "model.layers.63.self_attn.k_proj.weight.quant_map": "model-00013-of-00013.safetensors",
@@ -1765,6 +1886,7 @@
1765
  "model.layers.63.self_attn.o_proj.weight.absmax": "model-00013-of-00013.safetensors",
1766
  "model.layers.63.self_attn.o_proj.weight.quant_map": "model-00013-of-00013.safetensors",
1767
  "model.layers.63.self_attn.o_proj.weight.quant_state.bitsandbytes__fp4": "model-00013-of-00013.safetensors",
 
1768
  "model.layers.63.self_attn.q_proj.weight": "model-00013-of-00013.safetensors",
1769
  "model.layers.63.self_attn.q_proj.weight.absmax": "model-00013-of-00013.safetensors",
1770
  "model.layers.63.self_attn.q_proj.weight.quant_map": "model-00013-of-00013.safetensors",
@@ -1786,6 +1908,7 @@
1786
  "model.layers.7.mlp.up_proj.weight.absmax": "model-00003-of-00013.safetensors",
1787
  "model.layers.7.mlp.up_proj.weight.quant_map": "model-00003-of-00013.safetensors",
1788
  "model.layers.7.mlp.up_proj.weight.quant_state.bitsandbytes__fp4": "model-00003-of-00013.safetensors",
 
1789
  "model.layers.7.self_attn.k_proj.weight": "model-00003-of-00013.safetensors",
1790
  "model.layers.7.self_attn.k_proj.weight.absmax": "model-00003-of-00013.safetensors",
1791
  "model.layers.7.self_attn.k_proj.weight.quant_map": "model-00003-of-00013.safetensors",
@@ -1794,6 +1917,7 @@
1794
  "model.layers.7.self_attn.o_proj.weight.absmax": "model-00003-of-00013.safetensors",
1795
  "model.layers.7.self_attn.o_proj.weight.quant_map": "model-00003-of-00013.safetensors",
1796
  "model.layers.7.self_attn.o_proj.weight.quant_state.bitsandbytes__fp4": "model-00003-of-00013.safetensors",
 
1797
  "model.layers.7.self_attn.q_proj.weight": "model-00003-of-00013.safetensors",
1798
  "model.layers.7.self_attn.q_proj.weight.absmax": "model-00003-of-00013.safetensors",
1799
  "model.layers.7.self_attn.q_proj.weight.quant_map": "model-00003-of-00013.safetensors",
@@ -1815,6 +1939,7 @@
1815
  "model.layers.8.mlp.up_proj.weight.absmax": "model-00003-of-00013.safetensors",
1816
  "model.layers.8.mlp.up_proj.weight.quant_map": "model-00003-of-00013.safetensors",
1817
  "model.layers.8.mlp.up_proj.weight.quant_state.bitsandbytes__fp4": "model-00003-of-00013.safetensors",
 
1818
  "model.layers.8.self_attn.k_proj.weight": "model-00003-of-00013.safetensors",
1819
  "model.layers.8.self_attn.k_proj.weight.absmax": "model-00003-of-00013.safetensors",
1820
  "model.layers.8.self_attn.k_proj.weight.quant_map": "model-00003-of-00013.safetensors",
@@ -1823,6 +1948,7 @@
1823
  "model.layers.8.self_attn.o_proj.weight.absmax": "model-00003-of-00013.safetensors",
1824
  "model.layers.8.self_attn.o_proj.weight.quant_map": "model-00003-of-00013.safetensors",
1825
  "model.layers.8.self_attn.o_proj.weight.quant_state.bitsandbytes__fp4": "model-00003-of-00013.safetensors",
 
1826
  "model.layers.8.self_attn.q_proj.weight": "model-00003-of-00013.safetensors",
1827
  "model.layers.8.self_attn.q_proj.weight.absmax": "model-00003-of-00013.safetensors",
1828
  "model.layers.8.self_attn.q_proj.weight.quant_map": "model-00003-of-00013.safetensors",
@@ -1844,6 +1970,7 @@
1844
  "model.layers.9.mlp.up_proj.weight.absmax": "model-00003-of-00013.safetensors",
1845
  "model.layers.9.mlp.up_proj.weight.quant_map": "model-00003-of-00013.safetensors",
1846
  "model.layers.9.mlp.up_proj.weight.quant_state.bitsandbytes__fp4": "model-00003-of-00013.safetensors",
 
1847
  "model.layers.9.self_attn.k_proj.weight": "model-00003-of-00013.safetensors",
1848
  "model.layers.9.self_attn.k_proj.weight.absmax": "model-00003-of-00013.safetensors",
1849
  "model.layers.9.self_attn.k_proj.weight.quant_map": "model-00003-of-00013.safetensors",
@@ -1852,6 +1979,7 @@
1852
  "model.layers.9.self_attn.o_proj.weight.absmax": "model-00003-of-00013.safetensors",
1853
  "model.layers.9.self_attn.o_proj.weight.quant_map": "model-00003-of-00013.safetensors",
1854
  "model.layers.9.self_attn.o_proj.weight.quant_state.bitsandbytes__fp4": "model-00003-of-00013.safetensors",
 
1855
  "model.layers.9.self_attn.q_proj.weight": "model-00003-of-00013.safetensors",
1856
  "model.layers.9.self_attn.q_proj.weight.absmax": "model-00003-of-00013.safetensors",
1857
  "model.layers.9.self_attn.q_proj.weight.quant_map": "model-00003-of-00013.safetensors",
 
1
  {
2
  "metadata": {
3
+ "total_size": 62917927104
4
  },
5
  "weight_map": {
6
  "model.embed_tokens.weight": "model-00001-of-00013.safetensors",
 
17
  "model.layers.0.mlp.up_proj.weight.absmax": "model-00002-of-00013.safetensors",
18
  "model.layers.0.mlp.up_proj.weight.quant_map": "model-00002-of-00013.safetensors",
19
  "model.layers.0.mlp.up_proj.weight.quant_state.bitsandbytes__fp4": "model-00002-of-00013.safetensors",
20
+ "model.layers.0.self_attn.k_norm.weight": "model-00002-of-00013.safetensors",
21
  "model.layers.0.self_attn.k_proj.weight": "model-00002-of-00013.safetensors",
22
  "model.layers.0.self_attn.k_proj.weight.absmax": "model-00002-of-00013.safetensors",
23
  "model.layers.0.self_attn.k_proj.weight.quant_map": "model-00002-of-00013.safetensors",
 
26
  "model.layers.0.self_attn.o_proj.weight.absmax": "model-00002-of-00013.safetensors",
27
  "model.layers.0.self_attn.o_proj.weight.quant_map": "model-00002-of-00013.safetensors",
28
  "model.layers.0.self_attn.o_proj.weight.quant_state.bitsandbytes__fp4": "model-00002-of-00013.safetensors",
29
+ "model.layers.0.self_attn.q_norm.weight": "model-00002-of-00013.safetensors",
30
  "model.layers.0.self_attn.q_proj.weight": "model-00002-of-00013.safetensors",
31
  "model.layers.0.self_attn.q_proj.weight.absmax": "model-00002-of-00013.safetensors",
32
  "model.layers.0.self_attn.q_proj.weight.quant_map": "model-00002-of-00013.safetensors",
 
48
  "model.layers.1.mlp.up_proj.weight.absmax": "model-00002-of-00013.safetensors",
49
  "model.layers.1.mlp.up_proj.weight.quant_map": "model-00002-of-00013.safetensors",
50
  "model.layers.1.mlp.up_proj.weight.quant_state.bitsandbytes__fp4": "model-00002-of-00013.safetensors",
51
+ "model.layers.1.self_attn.k_norm.weight": "model-00002-of-00013.safetensors",
52
  "model.layers.1.self_attn.k_proj.weight": "model-00002-of-00013.safetensors",
53
  "model.layers.1.self_attn.k_proj.weight.absmax": "model-00002-of-00013.safetensors",
54
  "model.layers.1.self_attn.k_proj.weight.quant_map": "model-00002-of-00013.safetensors",
 
57
  "model.layers.1.self_attn.o_proj.weight.absmax": "model-00002-of-00013.safetensors",
58
  "model.layers.1.self_attn.o_proj.weight.quant_map": "model-00002-of-00013.safetensors",
59
  "model.layers.1.self_attn.o_proj.weight.quant_state.bitsandbytes__fp4": "model-00002-of-00013.safetensors",
60
+ "model.layers.1.self_attn.q_norm.weight": "model-00002-of-00013.safetensors",
61
  "model.layers.1.self_attn.q_proj.weight": "model-00002-of-00013.safetensors",
62
  "model.layers.1.self_attn.q_proj.weight.absmax": "model-00002-of-00013.safetensors",
63
  "model.layers.1.self_attn.q_proj.weight.quant_map": "model-00002-of-00013.safetensors",
 
79
  "model.layers.10.mlp.up_proj.weight.absmax": "model-00003-of-00013.safetensors",
80
  "model.layers.10.mlp.up_proj.weight.quant_map": "model-00003-of-00013.safetensors",
81
  "model.layers.10.mlp.up_proj.weight.quant_state.bitsandbytes__fp4": "model-00003-of-00013.safetensors",
82
+ "model.layers.10.self_attn.k_norm.weight": "model-00003-of-00013.safetensors",
83
  "model.layers.10.self_attn.k_proj.weight": "model-00003-of-00013.safetensors",
84
  "model.layers.10.self_attn.k_proj.weight.absmax": "model-00003-of-00013.safetensors",
85
  "model.layers.10.self_attn.k_proj.weight.quant_map": "model-00003-of-00013.safetensors",
 
88
  "model.layers.10.self_attn.o_proj.weight.absmax": "model-00003-of-00013.safetensors",
89
  "model.layers.10.self_attn.o_proj.weight.quant_map": "model-00003-of-00013.safetensors",
90
  "model.layers.10.self_attn.o_proj.weight.quant_state.bitsandbytes__fp4": "model-00003-of-00013.safetensors",
91
+ "model.layers.10.self_attn.q_norm.weight": "model-00003-of-00013.safetensors",
92
  "model.layers.10.self_attn.q_proj.weight": "model-00003-of-00013.safetensors",
93
  "model.layers.10.self_attn.q_proj.weight.absmax": "model-00003-of-00013.safetensors",
94
  "model.layers.10.self_attn.q_proj.weight.quant_map": "model-00003-of-00013.safetensors",
 
110
  "model.layers.11.mlp.up_proj.weight.absmax": "model-00004-of-00013.safetensors",
111
  "model.layers.11.mlp.up_proj.weight.quant_map": "model-00004-of-00013.safetensors",
112
  "model.layers.11.mlp.up_proj.weight.quant_state.bitsandbytes__fp4": "model-00004-of-00013.safetensors",
113
+ "model.layers.11.self_attn.k_norm.weight": "model-00003-of-00013.safetensors",
114
  "model.layers.11.self_attn.k_proj.weight": "model-00003-of-00013.safetensors",
115
  "model.layers.11.self_attn.k_proj.weight.absmax": "model-00003-of-00013.safetensors",
116
  "model.layers.11.self_attn.k_proj.weight.quant_map": "model-00003-of-00013.safetensors",
 
119
  "model.layers.11.self_attn.o_proj.weight.absmax": "model-00004-of-00013.safetensors",
120
  "model.layers.11.self_attn.o_proj.weight.quant_map": "model-00004-of-00013.safetensors",
121
  "model.layers.11.self_attn.o_proj.weight.quant_state.bitsandbytes__fp4": "model-00004-of-00013.safetensors",
122
+ "model.layers.11.self_attn.q_norm.weight": "model-00003-of-00013.safetensors",
123
  "model.layers.11.self_attn.q_proj.weight": "model-00003-of-00013.safetensors",
124
  "model.layers.11.self_attn.q_proj.weight.absmax": "model-00003-of-00013.safetensors",
125
  "model.layers.11.self_attn.q_proj.weight.quant_map": "model-00003-of-00013.safetensors",
 
141
  "model.layers.12.mlp.up_proj.weight.absmax": "model-00004-of-00013.safetensors",
142
  "model.layers.12.mlp.up_proj.weight.quant_map": "model-00004-of-00013.safetensors",
143
  "model.layers.12.mlp.up_proj.weight.quant_state.bitsandbytes__fp4": "model-00004-of-00013.safetensors",
144
+ "model.layers.12.self_attn.k_norm.weight": "model-00004-of-00013.safetensors",
145
  "model.layers.12.self_attn.k_proj.weight": "model-00004-of-00013.safetensors",
146
  "model.layers.12.self_attn.k_proj.weight.absmax": "model-00004-of-00013.safetensors",
147
  "model.layers.12.self_attn.k_proj.weight.quant_map": "model-00004-of-00013.safetensors",
 
150
  "model.layers.12.self_attn.o_proj.weight.absmax": "model-00004-of-00013.safetensors",
151
  "model.layers.12.self_attn.o_proj.weight.quant_map": "model-00004-of-00013.safetensors",
152
  "model.layers.12.self_attn.o_proj.weight.quant_state.bitsandbytes__fp4": "model-00004-of-00013.safetensors",
153
+ "model.layers.12.self_attn.q_norm.weight": "model-00004-of-00013.safetensors",
154
  "model.layers.12.self_attn.q_proj.weight": "model-00004-of-00013.safetensors",
155
  "model.layers.12.self_attn.q_proj.weight.absmax": "model-00004-of-00013.safetensors",
156
  "model.layers.12.self_attn.q_proj.weight.quant_map": "model-00004-of-00013.safetensors",
 
172
  "model.layers.13.mlp.up_proj.weight.absmax": "model-00004-of-00013.safetensors",
173
  "model.layers.13.mlp.up_proj.weight.quant_map": "model-00004-of-00013.safetensors",
174
  "model.layers.13.mlp.up_proj.weight.quant_state.bitsandbytes__fp4": "model-00004-of-00013.safetensors",
175
+ "model.layers.13.self_attn.k_norm.weight": "model-00004-of-00013.safetensors",
176
  "model.layers.13.self_attn.k_proj.weight": "model-00004-of-00013.safetensors",
177
  "model.layers.13.self_attn.k_proj.weight.absmax": "model-00004-of-00013.safetensors",
178
  "model.layers.13.self_attn.k_proj.weight.quant_map": "model-00004-of-00013.safetensors",
 
181
  "model.layers.13.self_attn.o_proj.weight.absmax": "model-00004-of-00013.safetensors",
182
  "model.layers.13.self_attn.o_proj.weight.quant_map": "model-00004-of-00013.safetensors",
183
  "model.layers.13.self_attn.o_proj.weight.quant_state.bitsandbytes__fp4": "model-00004-of-00013.safetensors",
184
+ "model.layers.13.self_attn.q_norm.weight": "model-00004-of-00013.safetensors",
185
  "model.layers.13.self_attn.q_proj.weight": "model-00004-of-00013.safetensors",
186
  "model.layers.13.self_attn.q_proj.weight.absmax": "model-00004-of-00013.safetensors",
187
  "model.layers.13.self_attn.q_proj.weight.quant_map": "model-00004-of-00013.safetensors",
 
203
  "model.layers.14.mlp.up_proj.weight.absmax": "model-00004-of-00013.safetensors",
204
  "model.layers.14.mlp.up_proj.weight.quant_map": "model-00004-of-00013.safetensors",
205
  "model.layers.14.mlp.up_proj.weight.quant_state.bitsandbytes__fp4": "model-00004-of-00013.safetensors",
206
+ "model.layers.14.self_attn.k_norm.weight": "model-00004-of-00013.safetensors",
207
  "model.layers.14.self_attn.k_proj.weight": "model-00004-of-00013.safetensors",
208
  "model.layers.14.self_attn.k_proj.weight.absmax": "model-00004-of-00013.safetensors",
209
  "model.layers.14.self_attn.k_proj.weight.quant_map": "model-00004-of-00013.safetensors",
 
212
  "model.layers.14.self_attn.o_proj.weight.absmax": "model-00004-of-00013.safetensors",
213
  "model.layers.14.self_attn.o_proj.weight.quant_map": "model-00004-of-00013.safetensors",
214
  "model.layers.14.self_attn.o_proj.weight.quant_state.bitsandbytes__fp4": "model-00004-of-00013.safetensors",
215
+ "model.layers.14.self_attn.q_norm.weight": "model-00004-of-00013.safetensors",
216
  "model.layers.14.self_attn.q_proj.weight": "model-00004-of-00013.safetensors",
217
  "model.layers.14.self_attn.q_proj.weight.absmax": "model-00004-of-00013.safetensors",
218
  "model.layers.14.self_attn.q_proj.weight.quant_map": "model-00004-of-00013.safetensors",
 
234
  "model.layers.15.mlp.up_proj.weight.absmax": "model-00004-of-00013.safetensors",
235
  "model.layers.15.mlp.up_proj.weight.quant_map": "model-00004-of-00013.safetensors",
236
  "model.layers.15.mlp.up_proj.weight.quant_state.bitsandbytes__fp4": "model-00004-of-00013.safetensors",
237
+ "model.layers.15.self_attn.k_norm.weight": "model-00004-of-00013.safetensors",
238
  "model.layers.15.self_attn.k_proj.weight": "model-00004-of-00013.safetensors",
239
  "model.layers.15.self_attn.k_proj.weight.absmax": "model-00004-of-00013.safetensors",
240
  "model.layers.15.self_attn.k_proj.weight.quant_map": "model-00004-of-00013.safetensors",
 
243
  "model.layers.15.self_attn.o_proj.weight.absmax": "model-00004-of-00013.safetensors",
244
  "model.layers.15.self_attn.o_proj.weight.quant_map": "model-00004-of-00013.safetensors",
245
  "model.layers.15.self_attn.o_proj.weight.quant_state.bitsandbytes__fp4": "model-00004-of-00013.safetensors",
246
+ "model.layers.15.self_attn.q_norm.weight": "model-00004-of-00013.safetensors",
247
  "model.layers.15.self_attn.q_proj.weight": "model-00004-of-00013.safetensors",
248
  "model.layers.15.self_attn.q_proj.weight.absmax": "model-00004-of-00013.safetensors",
249
  "model.layers.15.self_attn.q_proj.weight.quant_map": "model-00004-of-00013.safetensors",
 
265
  "model.layers.16.mlp.up_proj.weight.absmax": "model-00004-of-00013.safetensors",
266
  "model.layers.16.mlp.up_proj.weight.quant_map": "model-00004-of-00013.safetensors",
267
  "model.layers.16.mlp.up_proj.weight.quant_state.bitsandbytes__fp4": "model-00004-of-00013.safetensors",
268
+ "model.layers.16.self_attn.k_norm.weight": "model-00004-of-00013.safetensors",
269
  "model.layers.16.self_attn.k_proj.weight": "model-00004-of-00013.safetensors",
270
  "model.layers.16.self_attn.k_proj.weight.absmax": "model-00004-of-00013.safetensors",
271
  "model.layers.16.self_attn.k_proj.weight.quant_map": "model-00004-of-00013.safetensors",
 
274
  "model.layers.16.self_attn.o_proj.weight.absmax": "model-00004-of-00013.safetensors",
275
  "model.layers.16.self_attn.o_proj.weight.quant_map": "model-00004-of-00013.safetensors",
276
  "model.layers.16.self_attn.o_proj.weight.quant_state.bitsandbytes__fp4": "model-00004-of-00013.safetensors",
277
+ "model.layers.16.self_attn.q_norm.weight": "model-00004-of-00013.safetensors",
278
  "model.layers.16.self_attn.q_proj.weight": "model-00004-of-00013.safetensors",
279
  "model.layers.16.self_attn.q_proj.weight.absmax": "model-00004-of-00013.safetensors",
280
  "model.layers.16.self_attn.q_proj.weight.quant_map": "model-00004-of-00013.safetensors",
 
296
  "model.layers.17.mlp.up_proj.weight.absmax": "model-00005-of-00013.safetensors",
297
  "model.layers.17.mlp.up_proj.weight.quant_map": "model-00005-of-00013.safetensors",
298
  "model.layers.17.mlp.up_proj.weight.quant_state.bitsandbytes__fp4": "model-00005-of-00013.safetensors",
299
+ "model.layers.17.self_attn.k_norm.weight": "model-00005-of-00013.safetensors",
300
  "model.layers.17.self_attn.k_proj.weight": "model-00005-of-00013.safetensors",
301
  "model.layers.17.self_attn.k_proj.weight.absmax": "model-00005-of-00013.safetensors",
302
  "model.layers.17.self_attn.k_proj.weight.quant_map": "model-00005-of-00013.safetensors",
 
305
  "model.layers.17.self_attn.o_proj.weight.absmax": "model-00005-of-00013.safetensors",
306
  "model.layers.17.self_attn.o_proj.weight.quant_map": "model-00005-of-00013.safetensors",
307
  "model.layers.17.self_attn.o_proj.weight.quant_state.bitsandbytes__fp4": "model-00005-of-00013.safetensors",
308
+ "model.layers.17.self_attn.q_norm.weight": "model-00005-of-00013.safetensors",
309
  "model.layers.17.self_attn.q_proj.weight": "model-00005-of-00013.safetensors",
310
  "model.layers.17.self_attn.q_proj.weight.absmax": "model-00005-of-00013.safetensors",
311
  "model.layers.17.self_attn.q_proj.weight.quant_map": "model-00005-of-00013.safetensors",
 
327
  "model.layers.18.mlp.up_proj.weight.absmax": "model-00005-of-00013.safetensors",
328
  "model.layers.18.mlp.up_proj.weight.quant_map": "model-00005-of-00013.safetensors",
329
  "model.layers.18.mlp.up_proj.weight.quant_state.bitsandbytes__fp4": "model-00005-of-00013.safetensors",
330
+ "model.layers.18.self_attn.k_norm.weight": "model-00005-of-00013.safetensors",
331
  "model.layers.18.self_attn.k_proj.weight": "model-00005-of-00013.safetensors",
332
  "model.layers.18.self_attn.k_proj.weight.absmax": "model-00005-of-00013.safetensors",
333
  "model.layers.18.self_attn.k_proj.weight.quant_map": "model-00005-of-00013.safetensors",
 
336
  "model.layers.18.self_attn.o_proj.weight.absmax": "model-00005-of-00013.safetensors",
337
  "model.layers.18.self_attn.o_proj.weight.quant_map": "model-00005-of-00013.safetensors",
338
  "model.layers.18.self_attn.o_proj.weight.quant_state.bitsandbytes__fp4": "model-00005-of-00013.safetensors",
339
+ "model.layers.18.self_attn.q_norm.weight": "model-00005-of-00013.safetensors",
340
  "model.layers.18.self_attn.q_proj.weight": "model-00005-of-00013.safetensors",
341
  "model.layers.18.self_attn.q_proj.weight.absmax": "model-00005-of-00013.safetensors",
342
  "model.layers.18.self_attn.q_proj.weight.quant_map": "model-00005-of-00013.safetensors",
 
358
  "model.layers.19.mlp.up_proj.weight.absmax": "model-00005-of-00013.safetensors",
359
  "model.layers.19.mlp.up_proj.weight.quant_map": "model-00005-of-00013.safetensors",
360
  "model.layers.19.mlp.up_proj.weight.quant_state.bitsandbytes__fp4": "model-00005-of-00013.safetensors",
361
+ "model.layers.19.self_attn.k_norm.weight": "model-00005-of-00013.safetensors",
362
  "model.layers.19.self_attn.k_proj.weight": "model-00005-of-00013.safetensors",
363
  "model.layers.19.self_attn.k_proj.weight.absmax": "model-00005-of-00013.safetensors",
364
  "model.layers.19.self_attn.k_proj.weight.quant_map": "model-00005-of-00013.safetensors",
 
367
  "model.layers.19.self_attn.o_proj.weight.absmax": "model-00005-of-00013.safetensors",
368
  "model.layers.19.self_attn.o_proj.weight.quant_map": "model-00005-of-00013.safetensors",
369
  "model.layers.19.self_attn.o_proj.weight.quant_state.bitsandbytes__fp4": "model-00005-of-00013.safetensors",
370
+ "model.layers.19.self_attn.q_norm.weight": "model-00005-of-00013.safetensors",
371
  "model.layers.19.self_attn.q_proj.weight": "model-00005-of-00013.safetensors",
372
  "model.layers.19.self_attn.q_proj.weight.absmax": "model-00005-of-00013.safetensors",
373
  "model.layers.19.self_attn.q_proj.weight.quant_map": "model-00005-of-00013.safetensors",
 
389
  "model.layers.2.mlp.up_proj.weight.absmax": "model-00002-of-00013.safetensors",
390
  "model.layers.2.mlp.up_proj.weight.quant_map": "model-00002-of-00013.safetensors",
391
  "model.layers.2.mlp.up_proj.weight.quant_state.bitsandbytes__fp4": "model-00002-of-00013.safetensors",
392
+ "model.layers.2.self_attn.k_norm.weight": "model-00002-of-00013.safetensors",
393
  "model.layers.2.self_attn.k_proj.weight": "model-00002-of-00013.safetensors",
394
  "model.layers.2.self_attn.k_proj.weight.absmax": "model-00002-of-00013.safetensors",
395
  "model.layers.2.self_attn.k_proj.weight.quant_map": "model-00002-of-00013.safetensors",
 
398
  "model.layers.2.self_attn.o_proj.weight.absmax": "model-00002-of-00013.safetensors",
399
  "model.layers.2.self_attn.o_proj.weight.quant_map": "model-00002-of-00013.safetensors",
400
  "model.layers.2.self_attn.o_proj.weight.quant_state.bitsandbytes__fp4": "model-00002-of-00013.safetensors",
401
+ "model.layers.2.self_attn.q_norm.weight": "model-00002-of-00013.safetensors",
402
  "model.layers.2.self_attn.q_proj.weight": "model-00002-of-00013.safetensors",
403
  "model.layers.2.self_attn.q_proj.weight.absmax": "model-00002-of-00013.safetensors",
404
  "model.layers.2.self_attn.q_proj.weight.quant_map": "model-00002-of-00013.safetensors",
 
420
  "model.layers.20.mlp.up_proj.weight.absmax": "model-00005-of-00013.safetensors",
421
  "model.layers.20.mlp.up_proj.weight.quant_map": "model-00005-of-00013.safetensors",
422
  "model.layers.20.mlp.up_proj.weight.quant_state.bitsandbytes__fp4": "model-00005-of-00013.safetensors",
423
+ "model.layers.20.self_attn.k_norm.weight": "model-00005-of-00013.safetensors",
424
  "model.layers.20.self_attn.k_proj.weight": "model-00005-of-00013.safetensors",
425
  "model.layers.20.self_attn.k_proj.weight.absmax": "model-00005-of-00013.safetensors",
426
  "model.layers.20.self_attn.k_proj.weight.quant_map": "model-00005-of-00013.safetensors",
 
429
  "model.layers.20.self_attn.o_proj.weight.absmax": "model-00005-of-00013.safetensors",
430
  "model.layers.20.self_attn.o_proj.weight.quant_map": "model-00005-of-00013.safetensors",
431
  "model.layers.20.self_attn.o_proj.weight.quant_state.bitsandbytes__fp4": "model-00005-of-00013.safetensors",
432
+ "model.layers.20.self_attn.q_norm.weight": "model-00005-of-00013.safetensors",
433
  "model.layers.20.self_attn.q_proj.weight": "model-00005-of-00013.safetensors",
434
  "model.layers.20.self_attn.q_proj.weight.absmax": "model-00005-of-00013.safetensors",
435
  "model.layers.20.self_attn.q_proj.weight.quant_map": "model-00005-of-00013.safetensors",
 
451
  "model.layers.21.mlp.up_proj.weight.absmax": "model-00005-of-00013.safetensors",
452
  "model.layers.21.mlp.up_proj.weight.quant_map": "model-00005-of-00013.safetensors",
453
  "model.layers.21.mlp.up_proj.weight.quant_state.bitsandbytes__fp4": "model-00005-of-00013.safetensors",
454
+ "model.layers.21.self_attn.k_norm.weight": "model-00005-of-00013.safetensors",
455
  "model.layers.21.self_attn.k_proj.weight": "model-00005-of-00013.safetensors",
456
  "model.layers.21.self_attn.k_proj.weight.absmax": "model-00005-of-00013.safetensors",
457
  "model.layers.21.self_attn.k_proj.weight.quant_map": "model-00005-of-00013.safetensors",
 
460
  "model.layers.21.self_attn.o_proj.weight.absmax": "model-00005-of-00013.safetensors",
461
  "model.layers.21.self_attn.o_proj.weight.quant_map": "model-00005-of-00013.safetensors",
462
  "model.layers.21.self_attn.o_proj.weight.quant_state.bitsandbytes__fp4": "model-00005-of-00013.safetensors",
463
+ "model.layers.21.self_attn.q_norm.weight": "model-00005-of-00013.safetensors",
464
  "model.layers.21.self_attn.q_proj.weight": "model-00005-of-00013.safetensors",
465
  "model.layers.21.self_attn.q_proj.weight.absmax": "model-00005-of-00013.safetensors",
466
  "model.layers.21.self_attn.q_proj.weight.quant_map": "model-00005-of-00013.safetensors",
 
482
  "model.layers.22.mlp.up_proj.weight.absmax": "model-00006-of-00013.safetensors",
483
  "model.layers.22.mlp.up_proj.weight.quant_map": "model-00006-of-00013.safetensors",
484
  "model.layers.22.mlp.up_proj.weight.quant_state.bitsandbytes__fp4": "model-00006-of-00013.safetensors",
485
+ "model.layers.22.self_attn.k_norm.weight": "model-00005-of-00013.safetensors",
486
  "model.layers.22.self_attn.k_proj.weight": "model-00005-of-00013.safetensors",
487
  "model.layers.22.self_attn.k_proj.weight.absmax": "model-00005-of-00013.safetensors",
488
  "model.layers.22.self_attn.k_proj.weight.quant_map": "model-00005-of-00013.safetensors",
 
491
  "model.layers.22.self_attn.o_proj.weight.absmax": "model-00005-of-00013.safetensors",
492
  "model.layers.22.self_attn.o_proj.weight.quant_map": "model-00005-of-00013.safetensors",
493
  "model.layers.22.self_attn.o_proj.weight.quant_state.bitsandbytes__fp4": "model-00005-of-00013.safetensors",
494
+ "model.layers.22.self_attn.q_norm.weight": "model-00005-of-00013.safetensors",
495
  "model.layers.22.self_attn.q_proj.weight": "model-00005-of-00013.safetensors",
496
  "model.layers.22.self_attn.q_proj.weight.absmax": "model-00005-of-00013.safetensors",
497
  "model.layers.22.self_attn.q_proj.weight.quant_map": "model-00005-of-00013.safetensors",
 
513
  "model.layers.23.mlp.up_proj.weight.absmax": "model-00006-of-00013.safetensors",
514
  "model.layers.23.mlp.up_proj.weight.quant_map": "model-00006-of-00013.safetensors",
515
  "model.layers.23.mlp.up_proj.weight.quant_state.bitsandbytes__fp4": "model-00006-of-00013.safetensors",
516
+ "model.layers.23.self_attn.k_norm.weight": "model-00006-of-00013.safetensors",
517
  "model.layers.23.self_attn.k_proj.weight": "model-00006-of-00013.safetensors",
518
  "model.layers.23.self_attn.k_proj.weight.absmax": "model-00006-of-00013.safetensors",
519
  "model.layers.23.self_attn.k_proj.weight.quant_map": "model-00006-of-00013.safetensors",
 
522
  "model.layers.23.self_attn.o_proj.weight.absmax": "model-00006-of-00013.safetensors",
523
  "model.layers.23.self_attn.o_proj.weight.quant_map": "model-00006-of-00013.safetensors",
524
  "model.layers.23.self_attn.o_proj.weight.quant_state.bitsandbytes__fp4": "model-00006-of-00013.safetensors",
525
+ "model.layers.23.self_attn.q_norm.weight": "model-00006-of-00013.safetensors",
526
  "model.layers.23.self_attn.q_proj.weight": "model-00006-of-00013.safetensors",
527
  "model.layers.23.self_attn.q_proj.weight.absmax": "model-00006-of-00013.safetensors",
528
  "model.layers.23.self_attn.q_proj.weight.quant_map": "model-00006-of-00013.safetensors",
 
544
  "model.layers.24.mlp.up_proj.weight.absmax": "model-00006-of-00013.safetensors",
545
  "model.layers.24.mlp.up_proj.weight.quant_map": "model-00006-of-00013.safetensors",
546
  "model.layers.24.mlp.up_proj.weight.quant_state.bitsandbytes__fp4": "model-00006-of-00013.safetensors",
547
+ "model.layers.24.self_attn.k_norm.weight": "model-00006-of-00013.safetensors",
548
  "model.layers.24.self_attn.k_proj.weight": "model-00006-of-00013.safetensors",
549
  "model.layers.24.self_attn.k_proj.weight.absmax": "model-00006-of-00013.safetensors",
550
  "model.layers.24.self_attn.k_proj.weight.quant_map": "model-00006-of-00013.safetensors",
 
553
  "model.layers.24.self_attn.o_proj.weight.absmax": "model-00006-of-00013.safetensors",
554
  "model.layers.24.self_attn.o_proj.weight.quant_map": "model-00006-of-00013.safetensors",
555
  "model.layers.24.self_attn.o_proj.weight.quant_state.bitsandbytes__fp4": "model-00006-of-00013.safetensors",
556
+ "model.layers.24.self_attn.q_norm.weight": "model-00006-of-00013.safetensors",
557
  "model.layers.24.self_attn.q_proj.weight": "model-00006-of-00013.safetensors",
558
  "model.layers.24.self_attn.q_proj.weight.absmax": "model-00006-of-00013.safetensors",
559
  "model.layers.24.self_attn.q_proj.weight.quant_map": "model-00006-of-00013.safetensors",
 
575
  "model.layers.25.mlp.up_proj.weight.absmax": "model-00006-of-00013.safetensors",
576
  "model.layers.25.mlp.up_proj.weight.quant_map": "model-00006-of-00013.safetensors",
577
  "model.layers.25.mlp.up_proj.weight.quant_state.bitsandbytes__fp4": "model-00006-of-00013.safetensors",
578
+ "model.layers.25.self_attn.k_norm.weight": "model-00006-of-00013.safetensors",
579
  "model.layers.25.self_attn.k_proj.weight": "model-00006-of-00013.safetensors",
580
  "model.layers.25.self_attn.k_proj.weight.absmax": "model-00006-of-00013.safetensors",
581
  "model.layers.25.self_attn.k_proj.weight.quant_map": "model-00006-of-00013.safetensors",
 
584
  "model.layers.25.self_attn.o_proj.weight.absmax": "model-00006-of-00013.safetensors",
585
  "model.layers.25.self_attn.o_proj.weight.quant_map": "model-00006-of-00013.safetensors",
586
  "model.layers.25.self_attn.o_proj.weight.quant_state.bitsandbytes__fp4": "model-00006-of-00013.safetensors",
587
+ "model.layers.25.self_attn.q_norm.weight": "model-00006-of-00013.safetensors",
588
  "model.layers.25.self_attn.q_proj.weight": "model-00006-of-00013.safetensors",
589
  "model.layers.25.self_attn.q_proj.weight.absmax": "model-00006-of-00013.safetensors",
590
  "model.layers.25.self_attn.q_proj.weight.quant_map": "model-00006-of-00013.safetensors",
 
606
  "model.layers.26.mlp.up_proj.weight.absmax": "model-00006-of-00013.safetensors",
607
  "model.layers.26.mlp.up_proj.weight.quant_map": "model-00006-of-00013.safetensors",
608
  "model.layers.26.mlp.up_proj.weight.quant_state.bitsandbytes__fp4": "model-00006-of-00013.safetensors",
609
+ "model.layers.26.self_attn.k_norm.weight": "model-00006-of-00013.safetensors",
610
  "model.layers.26.self_attn.k_proj.weight": "model-00006-of-00013.safetensors",
611
  "model.layers.26.self_attn.k_proj.weight.absmax": "model-00006-of-00013.safetensors",
612
  "model.layers.26.self_attn.k_proj.weight.quant_map": "model-00006-of-00013.safetensors",
 
615
  "model.layers.26.self_attn.o_proj.weight.absmax": "model-00006-of-00013.safetensors",
616
  "model.layers.26.self_attn.o_proj.weight.quant_map": "model-00006-of-00013.safetensors",
617
  "model.layers.26.self_attn.o_proj.weight.quant_state.bitsandbytes__fp4": "model-00006-of-00013.safetensors",
618
+ "model.layers.26.self_attn.q_norm.weight": "model-00006-of-00013.safetensors",
619
  "model.layers.26.self_attn.q_proj.weight": "model-00006-of-00013.safetensors",
620
  "model.layers.26.self_attn.q_proj.weight.absmax": "model-00006-of-00013.safetensors",
621
  "model.layers.26.self_attn.q_proj.weight.quant_map": "model-00006-of-00013.safetensors",
 
637
  "model.layers.27.mlp.up_proj.weight.absmax": "model-00006-of-00013.safetensors",
638
  "model.layers.27.mlp.up_proj.weight.quant_map": "model-00006-of-00013.safetensors",
639
  "model.layers.27.mlp.up_proj.weight.quant_state.bitsandbytes__fp4": "model-00006-of-00013.safetensors",
640
+ "model.layers.27.self_attn.k_norm.weight": "model-00006-of-00013.safetensors",
641
  "model.layers.27.self_attn.k_proj.weight": "model-00006-of-00013.safetensors",
642
  "model.layers.27.self_attn.k_proj.weight.absmax": "model-00006-of-00013.safetensors",
643
  "model.layers.27.self_attn.k_proj.weight.quant_map": "model-00006-of-00013.safetensors",
 
646
  "model.layers.27.self_attn.o_proj.weight.absmax": "model-00006-of-00013.safetensors",
647
  "model.layers.27.self_attn.o_proj.weight.quant_map": "model-00006-of-00013.safetensors",
648
  "model.layers.27.self_attn.o_proj.weight.quant_state.bitsandbytes__fp4": "model-00006-of-00013.safetensors",
649
+ "model.layers.27.self_attn.q_norm.weight": "model-00006-of-00013.safetensors",
650
  "model.layers.27.self_attn.q_proj.weight": "model-00006-of-00013.safetensors",
651
  "model.layers.27.self_attn.q_proj.weight.absmax": "model-00006-of-00013.safetensors",
652
  "model.layers.27.self_attn.q_proj.weight.quant_map": "model-00006-of-00013.safetensors",
 
668
  "model.layers.28.mlp.up_proj.weight.absmax": "model-00007-of-00013.safetensors",
669
  "model.layers.28.mlp.up_proj.weight.quant_map": "model-00007-of-00013.safetensors",
670
  "model.layers.28.mlp.up_proj.weight.quant_state.bitsandbytes__fp4": "model-00007-of-00013.safetensors",
671
+ "model.layers.28.self_attn.k_norm.weight": "model-00007-of-00013.safetensors",
672
  "model.layers.28.self_attn.k_proj.weight": "model-00007-of-00013.safetensors",
673
  "model.layers.28.self_attn.k_proj.weight.absmax": "model-00007-of-00013.safetensors",
674
  "model.layers.28.self_attn.k_proj.weight.quant_map": "model-00007-of-00013.safetensors",
 
677
  "model.layers.28.self_attn.o_proj.weight.absmax": "model-00007-of-00013.safetensors",
678
  "model.layers.28.self_attn.o_proj.weight.quant_map": "model-00007-of-00013.safetensors",
679
  "model.layers.28.self_attn.o_proj.weight.quant_state.bitsandbytes__fp4": "model-00007-of-00013.safetensors",
680
+ "model.layers.28.self_attn.q_norm.weight": "model-00007-of-00013.safetensors",
681
  "model.layers.28.self_attn.q_proj.weight": "model-00007-of-00013.safetensors",
682
  "model.layers.28.self_attn.q_proj.weight.absmax": "model-00007-of-00013.safetensors",
683
  "model.layers.28.self_attn.q_proj.weight.quant_map": "model-00007-of-00013.safetensors",
 
699
  "model.layers.29.mlp.up_proj.weight.absmax": "model-00007-of-00013.safetensors",
700
  "model.layers.29.mlp.up_proj.weight.quant_map": "model-00007-of-00013.safetensors",
701
  "model.layers.29.mlp.up_proj.weight.quant_state.bitsandbytes__fp4": "model-00007-of-00013.safetensors",
702
+ "model.layers.29.self_attn.k_norm.weight": "model-00007-of-00013.safetensors",
703
  "model.layers.29.self_attn.k_proj.weight": "model-00007-of-00013.safetensors",
704
  "model.layers.29.self_attn.k_proj.weight.absmax": "model-00007-of-00013.safetensors",
705
  "model.layers.29.self_attn.k_proj.weight.quant_map": "model-00007-of-00013.safetensors",
 
708
  "model.layers.29.self_attn.o_proj.weight.absmax": "model-00007-of-00013.safetensors",
709
  "model.layers.29.self_attn.o_proj.weight.quant_map": "model-00007-of-00013.safetensors",
710
  "model.layers.29.self_attn.o_proj.weight.quant_state.bitsandbytes__fp4": "model-00007-of-00013.safetensors",
711
+ "model.layers.29.self_attn.q_norm.weight": "model-00007-of-00013.safetensors",
712
  "model.layers.29.self_attn.q_proj.weight": "model-00007-of-00013.safetensors",
713
  "model.layers.29.self_attn.q_proj.weight.absmax": "model-00007-of-00013.safetensors",
714
  "model.layers.29.self_attn.q_proj.weight.quant_map": "model-00007-of-00013.safetensors",
 
730
  "model.layers.3.mlp.up_proj.weight.absmax": "model-00002-of-00013.safetensors",
731
  "model.layers.3.mlp.up_proj.weight.quant_map": "model-00002-of-00013.safetensors",
732
  "model.layers.3.mlp.up_proj.weight.quant_state.bitsandbytes__fp4": "model-00002-of-00013.safetensors",
733
+ "model.layers.3.self_attn.k_norm.weight": "model-00002-of-00013.safetensors",
734
  "model.layers.3.self_attn.k_proj.weight": "model-00002-of-00013.safetensors",
735
  "model.layers.3.self_attn.k_proj.weight.absmax": "model-00002-of-00013.safetensors",
736
  "model.layers.3.self_attn.k_proj.weight.quant_map": "model-00002-of-00013.safetensors",
 
739
  "model.layers.3.self_attn.o_proj.weight.absmax": "model-00002-of-00013.safetensors",
740
  "model.layers.3.self_attn.o_proj.weight.quant_map": "model-00002-of-00013.safetensors",
741
  "model.layers.3.self_attn.o_proj.weight.quant_state.bitsandbytes__fp4": "model-00002-of-00013.safetensors",
742
+ "model.layers.3.self_attn.q_norm.weight": "model-00002-of-00013.safetensors",
743
  "model.layers.3.self_attn.q_proj.weight": "model-00002-of-00013.safetensors",
744
  "model.layers.3.self_attn.q_proj.weight.absmax": "model-00002-of-00013.safetensors",
745
  "model.layers.3.self_attn.q_proj.weight.quant_map": "model-00002-of-00013.safetensors",
 
761
  "model.layers.30.mlp.up_proj.weight.absmax": "model-00007-of-00013.safetensors",
762
  "model.layers.30.mlp.up_proj.weight.quant_map": "model-00007-of-00013.safetensors",
763
  "model.layers.30.mlp.up_proj.weight.quant_state.bitsandbytes__fp4": "model-00007-of-00013.safetensors",
764
+ "model.layers.30.self_attn.k_norm.weight": "model-00007-of-00013.safetensors",
765
  "model.layers.30.self_attn.k_proj.weight": "model-00007-of-00013.safetensors",
766
  "model.layers.30.self_attn.k_proj.weight.absmax": "model-00007-of-00013.safetensors",
767
  "model.layers.30.self_attn.k_proj.weight.quant_map": "model-00007-of-00013.safetensors",
 
770
  "model.layers.30.self_attn.o_proj.weight.absmax": "model-00007-of-00013.safetensors",
771
  "model.layers.30.self_attn.o_proj.weight.quant_map": "model-00007-of-00013.safetensors",
772
  "model.layers.30.self_attn.o_proj.weight.quant_state.bitsandbytes__fp4": "model-00007-of-00013.safetensors",
773
+ "model.layers.30.self_attn.q_norm.weight": "model-00007-of-00013.safetensors",
774
  "model.layers.30.self_attn.q_proj.weight": "model-00007-of-00013.safetensors",
775
  "model.layers.30.self_attn.q_proj.weight.absmax": "model-00007-of-00013.safetensors",
776
  "model.layers.30.self_attn.q_proj.weight.quant_map": "model-00007-of-00013.safetensors",
 
792
  "model.layers.31.mlp.up_proj.weight.absmax": "model-00007-of-00013.safetensors",
793
  "model.layers.31.mlp.up_proj.weight.quant_map": "model-00007-of-00013.safetensors",
794
  "model.layers.31.mlp.up_proj.weight.quant_state.bitsandbytes__fp4": "model-00007-of-00013.safetensors",
795
+ "model.layers.31.self_attn.k_norm.weight": "model-00007-of-00013.safetensors",
796
  "model.layers.31.self_attn.k_proj.weight": "model-00007-of-00013.safetensors",
797
  "model.layers.31.self_attn.k_proj.weight.absmax": "model-00007-of-00013.safetensors",
798
  "model.layers.31.self_attn.k_proj.weight.quant_map": "model-00007-of-00013.safetensors",
 
801
  "model.layers.31.self_attn.o_proj.weight.absmax": "model-00007-of-00013.safetensors",
802
  "model.layers.31.self_attn.o_proj.weight.quant_map": "model-00007-of-00013.safetensors",
803
  "model.layers.31.self_attn.o_proj.weight.quant_state.bitsandbytes__fp4": "model-00007-of-00013.safetensors",
804
+ "model.layers.31.self_attn.q_norm.weight": "model-00007-of-00013.safetensors",
805
  "model.layers.31.self_attn.q_proj.weight": "model-00007-of-00013.safetensors",
806
  "model.layers.31.self_attn.q_proj.weight.absmax": "model-00007-of-00013.safetensors",
807
  "model.layers.31.self_attn.q_proj.weight.quant_map": "model-00007-of-00013.safetensors",
 
823
  "model.layers.32.mlp.up_proj.weight.absmax": "model-00007-of-00013.safetensors",
824
  "model.layers.32.mlp.up_proj.weight.quant_map": "model-00007-of-00013.safetensors",
825
  "model.layers.32.mlp.up_proj.weight.quant_state.bitsandbytes__fp4": "model-00007-of-00013.safetensors",
826
+ "model.layers.32.self_attn.k_norm.weight": "model-00007-of-00013.safetensors",
827
  "model.layers.32.self_attn.k_proj.weight": "model-00007-of-00013.safetensors",
828
  "model.layers.32.self_attn.k_proj.weight.absmax": "model-00007-of-00013.safetensors",
829
  "model.layers.32.self_attn.k_proj.weight.quant_map": "model-00007-of-00013.safetensors",
 
832
  "model.layers.32.self_attn.o_proj.weight.absmax": "model-00007-of-00013.safetensors",
833
  "model.layers.32.self_attn.o_proj.weight.quant_map": "model-00007-of-00013.safetensors",
834
  "model.layers.32.self_attn.o_proj.weight.quant_state.bitsandbytes__fp4": "model-00007-of-00013.safetensors",
835
+ "model.layers.32.self_attn.q_norm.weight": "model-00007-of-00013.safetensors",
836
  "model.layers.32.self_attn.q_proj.weight": "model-00007-of-00013.safetensors",
837
  "model.layers.32.self_attn.q_proj.weight.absmax": "model-00007-of-00013.safetensors",
838
  "model.layers.32.self_attn.q_proj.weight.quant_map": "model-00007-of-00013.safetensors",
 
854
  "model.layers.33.mlp.up_proj.weight.absmax": "model-00008-of-00013.safetensors",
855
  "model.layers.33.mlp.up_proj.weight.quant_map": "model-00008-of-00013.safetensors",
856
  "model.layers.33.mlp.up_proj.weight.quant_state.bitsandbytes__fp4": "model-00008-of-00013.safetensors",
857
+ "model.layers.33.self_attn.k_norm.weight": "model-00007-of-00013.safetensors",
858
  "model.layers.33.self_attn.k_proj.weight": "model-00007-of-00013.safetensors",
859
  "model.layers.33.self_attn.k_proj.weight.absmax": "model-00007-of-00013.safetensors",
860
  "model.layers.33.self_attn.k_proj.weight.quant_map": "model-00007-of-00013.safetensors",
 
863
  "model.layers.33.self_attn.o_proj.weight.absmax": "model-00007-of-00013.safetensors",
864
  "model.layers.33.self_attn.o_proj.weight.quant_map": "model-00007-of-00013.safetensors",
865
  "model.layers.33.self_attn.o_proj.weight.quant_state.bitsandbytes__fp4": "model-00007-of-00013.safetensors",
866
+ "model.layers.33.self_attn.q_norm.weight": "model-00007-of-00013.safetensors",
867
  "model.layers.33.self_attn.q_proj.weight": "model-00007-of-00013.safetensors",
868
  "model.layers.33.self_attn.q_proj.weight.absmax": "model-00007-of-00013.safetensors",
869
  "model.layers.33.self_attn.q_proj.weight.quant_map": "model-00007-of-00013.safetensors",
 
885
  "model.layers.34.mlp.up_proj.weight.absmax": "model-00008-of-00013.safetensors",
886
  "model.layers.34.mlp.up_proj.weight.quant_map": "model-00008-of-00013.safetensors",
887
  "model.layers.34.mlp.up_proj.weight.quant_state.bitsandbytes__fp4": "model-00008-of-00013.safetensors",
888
+ "model.layers.34.self_attn.k_norm.weight": "model-00008-of-00013.safetensors",
889
  "model.layers.34.self_attn.k_proj.weight": "model-00008-of-00013.safetensors",
890
  "model.layers.34.self_attn.k_proj.weight.absmax": "model-00008-of-00013.safetensors",
891
  "model.layers.34.self_attn.k_proj.weight.quant_map": "model-00008-of-00013.safetensors",
 
894
  "model.layers.34.self_attn.o_proj.weight.absmax": "model-00008-of-00013.safetensors",
895
  "model.layers.34.self_attn.o_proj.weight.quant_map": "model-00008-of-00013.safetensors",
896
  "model.layers.34.self_attn.o_proj.weight.quant_state.bitsandbytes__fp4": "model-00008-of-00013.safetensors",
897
+ "model.layers.34.self_attn.q_norm.weight": "model-00008-of-00013.safetensors",
898
  "model.layers.34.self_attn.q_proj.weight": "model-00008-of-00013.safetensors",
899
  "model.layers.34.self_attn.q_proj.weight.absmax": "model-00008-of-00013.safetensors",
900
  "model.layers.34.self_attn.q_proj.weight.quant_map": "model-00008-of-00013.safetensors",
 
916
  "model.layers.35.mlp.up_proj.weight.absmax": "model-00008-of-00013.safetensors",
917
  "model.layers.35.mlp.up_proj.weight.quant_map": "model-00008-of-00013.safetensors",
918
  "model.layers.35.mlp.up_proj.weight.quant_state.bitsandbytes__fp4": "model-00008-of-00013.safetensors",
919
+ "model.layers.35.self_attn.k_norm.weight": "model-00008-of-00013.safetensors",
920
  "model.layers.35.self_attn.k_proj.weight": "model-00008-of-00013.safetensors",
921
  "model.layers.35.self_attn.k_proj.weight.absmax": "model-00008-of-00013.safetensors",
922
  "model.layers.35.self_attn.k_proj.weight.quant_map": "model-00008-of-00013.safetensors",
 
925
  "model.layers.35.self_attn.o_proj.weight.absmax": "model-00008-of-00013.safetensors",
926
  "model.layers.35.self_attn.o_proj.weight.quant_map": "model-00008-of-00013.safetensors",
927
  "model.layers.35.self_attn.o_proj.weight.quant_state.bitsandbytes__fp4": "model-00008-of-00013.safetensors",
928
+ "model.layers.35.self_attn.q_norm.weight": "model-00008-of-00013.safetensors",
929
  "model.layers.35.self_attn.q_proj.weight": "model-00008-of-00013.safetensors",
930
  "model.layers.35.self_attn.q_proj.weight.absmax": "model-00008-of-00013.safetensors",
931
  "model.layers.35.self_attn.q_proj.weight.quant_map": "model-00008-of-00013.safetensors",
 
947
  "model.layers.36.mlp.up_proj.weight.absmax": "model-00008-of-00013.safetensors",
948
  "model.layers.36.mlp.up_proj.weight.quant_map": "model-00008-of-00013.safetensors",
949
  "model.layers.36.mlp.up_proj.weight.quant_state.bitsandbytes__fp4": "model-00008-of-00013.safetensors",
950
+ "model.layers.36.self_attn.k_norm.weight": "model-00008-of-00013.safetensors",
951
  "model.layers.36.self_attn.k_proj.weight": "model-00008-of-00013.safetensors",
952
  "model.layers.36.self_attn.k_proj.weight.absmax": "model-00008-of-00013.safetensors",
953
  "model.layers.36.self_attn.k_proj.weight.quant_map": "model-00008-of-00013.safetensors",
 
956
  "model.layers.36.self_attn.o_proj.weight.absmax": "model-00008-of-00013.safetensors",
957
  "model.layers.36.self_attn.o_proj.weight.quant_map": "model-00008-of-00013.safetensors",
958
  "model.layers.36.self_attn.o_proj.weight.quant_state.bitsandbytes__fp4": "model-00008-of-00013.safetensors",
959
+ "model.layers.36.self_attn.q_norm.weight": "model-00008-of-00013.safetensors",
960
  "model.layers.36.self_attn.q_proj.weight": "model-00008-of-00013.safetensors",
961
  "model.layers.36.self_attn.q_proj.weight.absmax": "model-00008-of-00013.safetensors",
962
  "model.layers.36.self_attn.q_proj.weight.quant_map": "model-00008-of-00013.safetensors",
 
978
  "model.layers.37.mlp.up_proj.weight.absmax": "model-00008-of-00013.safetensors",
979
  "model.layers.37.mlp.up_proj.weight.quant_map": "model-00008-of-00013.safetensors",
980
  "model.layers.37.mlp.up_proj.weight.quant_state.bitsandbytes__fp4": "model-00008-of-00013.safetensors",
981
+ "model.layers.37.self_attn.k_norm.weight": "model-00008-of-00013.safetensors",
982
  "model.layers.37.self_attn.k_proj.weight": "model-00008-of-00013.safetensors",
983
  "model.layers.37.self_attn.k_proj.weight.absmax": "model-00008-of-00013.safetensors",
984
  "model.layers.37.self_attn.k_proj.weight.quant_map": "model-00008-of-00013.safetensors",
 
987
  "model.layers.37.self_attn.o_proj.weight.absmax": "model-00008-of-00013.safetensors",
988
  "model.layers.37.self_attn.o_proj.weight.quant_map": "model-00008-of-00013.safetensors",
989
  "model.layers.37.self_attn.o_proj.weight.quant_state.bitsandbytes__fp4": "model-00008-of-00013.safetensors",
990
+ "model.layers.37.self_attn.q_norm.weight": "model-00008-of-00013.safetensors",
991
  "model.layers.37.self_attn.q_proj.weight": "model-00008-of-00013.safetensors",
992
  "model.layers.37.self_attn.q_proj.weight.absmax": "model-00008-of-00013.safetensors",
993
  "model.layers.37.self_attn.q_proj.weight.quant_map": "model-00008-of-00013.safetensors",
 
1009
  "model.layers.38.mlp.up_proj.weight.absmax": "model-00008-of-00013.safetensors",
1010
  "model.layers.38.mlp.up_proj.weight.quant_map": "model-00008-of-00013.safetensors",
1011
  "model.layers.38.mlp.up_proj.weight.quant_state.bitsandbytes__fp4": "model-00008-of-00013.safetensors",
1012
+ "model.layers.38.self_attn.k_norm.weight": "model-00008-of-00013.safetensors",
1013
  "model.layers.38.self_attn.k_proj.weight": "model-00008-of-00013.safetensors",
1014
  "model.layers.38.self_attn.k_proj.weight.absmax": "model-00008-of-00013.safetensors",
1015
  "model.layers.38.self_attn.k_proj.weight.quant_map": "model-00008-of-00013.safetensors",
 
1018
  "model.layers.38.self_attn.o_proj.weight.absmax": "model-00008-of-00013.safetensors",
1019
  "model.layers.38.self_attn.o_proj.weight.quant_map": "model-00008-of-00013.safetensors",
1020
  "model.layers.38.self_attn.o_proj.weight.quant_state.bitsandbytes__fp4": "model-00008-of-00013.safetensors",
1021
+ "model.layers.38.self_attn.q_norm.weight": "model-00008-of-00013.safetensors",
1022
  "model.layers.38.self_attn.q_proj.weight": "model-00008-of-00013.safetensors",
1023
  "model.layers.38.self_attn.q_proj.weight.absmax": "model-00008-of-00013.safetensors",
1024
  "model.layers.38.self_attn.q_proj.weight.quant_map": "model-00008-of-00013.safetensors",
 
1040
  "model.layers.39.mlp.up_proj.weight.absmax": "model-00009-of-00013.safetensors",
1041
  "model.layers.39.mlp.up_proj.weight.quant_map": "model-00009-of-00013.safetensors",
1042
  "model.layers.39.mlp.up_proj.weight.quant_state.bitsandbytes__fp4": "model-00009-of-00013.safetensors",
1043
+ "model.layers.39.self_attn.k_norm.weight": "model-00009-of-00013.safetensors",
1044
  "model.layers.39.self_attn.k_proj.weight": "model-00009-of-00013.safetensors",
1045
  "model.layers.39.self_attn.k_proj.weight.absmax": "model-00009-of-00013.safetensors",
1046
  "model.layers.39.self_attn.k_proj.weight.quant_map": "model-00009-of-00013.safetensors",
 
1049
  "model.layers.39.self_attn.o_proj.weight.absmax": "model-00009-of-00013.safetensors",
1050
  "model.layers.39.self_attn.o_proj.weight.quant_map": "model-00009-of-00013.safetensors",
1051
  "model.layers.39.self_attn.o_proj.weight.quant_state.bitsandbytes__fp4": "model-00009-of-00013.safetensors",
1052
+ "model.layers.39.self_attn.q_norm.weight": "model-00009-of-00013.safetensors",
1053
  "model.layers.39.self_attn.q_proj.weight": "model-00009-of-00013.safetensors",
1054
  "model.layers.39.self_attn.q_proj.weight.absmax": "model-00009-of-00013.safetensors",
1055
  "model.layers.39.self_attn.q_proj.weight.quant_map": "model-00009-of-00013.safetensors",
 
1071
  "model.layers.4.mlp.up_proj.weight.absmax": "model-00002-of-00013.safetensors",
1072
  "model.layers.4.mlp.up_proj.weight.quant_map": "model-00002-of-00013.safetensors",
1073
  "model.layers.4.mlp.up_proj.weight.quant_state.bitsandbytes__fp4": "model-00002-of-00013.safetensors",
1074
+ "model.layers.4.self_attn.k_norm.weight": "model-00002-of-00013.safetensors",
1075
  "model.layers.4.self_attn.k_proj.weight": "model-00002-of-00013.safetensors",
1076
  "model.layers.4.self_attn.k_proj.weight.absmax": "model-00002-of-00013.safetensors",
1077
  "model.layers.4.self_attn.k_proj.weight.quant_map": "model-00002-of-00013.safetensors",
 
1080
  "model.layers.4.self_attn.o_proj.weight.absmax": "model-00002-of-00013.safetensors",
1081
  "model.layers.4.self_attn.o_proj.weight.quant_map": "model-00002-of-00013.safetensors",
1082
  "model.layers.4.self_attn.o_proj.weight.quant_state.bitsandbytes__fp4": "model-00002-of-00013.safetensors",
1083
+ "model.layers.4.self_attn.q_norm.weight": "model-00002-of-00013.safetensors",
1084
  "model.layers.4.self_attn.q_proj.weight": "model-00002-of-00013.safetensors",
1085
  "model.layers.4.self_attn.q_proj.weight.absmax": "model-00002-of-00013.safetensors",
1086
  "model.layers.4.self_attn.q_proj.weight.quant_map": "model-00002-of-00013.safetensors",
 
1102
  "model.layers.40.mlp.up_proj.weight.absmax": "model-00009-of-00013.safetensors",
1103
  "model.layers.40.mlp.up_proj.weight.quant_map": "model-00009-of-00013.safetensors",
1104
  "model.layers.40.mlp.up_proj.weight.quant_state.bitsandbytes__fp4": "model-00009-of-00013.safetensors",
1105
+ "model.layers.40.self_attn.k_norm.weight": "model-00009-of-00013.safetensors",
1106
  "model.layers.40.self_attn.k_proj.weight": "model-00009-of-00013.safetensors",
1107
  "model.layers.40.self_attn.k_proj.weight.absmax": "model-00009-of-00013.safetensors",
1108
  "model.layers.40.self_attn.k_proj.weight.quant_map": "model-00009-of-00013.safetensors",
 
1111
  "model.layers.40.self_attn.o_proj.weight.absmax": "model-00009-of-00013.safetensors",
1112
  "model.layers.40.self_attn.o_proj.weight.quant_map": "model-00009-of-00013.safetensors",
1113
  "model.layers.40.self_attn.o_proj.weight.quant_state.bitsandbytes__fp4": "model-00009-of-00013.safetensors",
1114
+ "model.layers.40.self_attn.q_norm.weight": "model-00009-of-00013.safetensors",
1115
  "model.layers.40.self_attn.q_proj.weight": "model-00009-of-00013.safetensors",
1116
  "model.layers.40.self_attn.q_proj.weight.absmax": "model-00009-of-00013.safetensors",
1117
  "model.layers.40.self_attn.q_proj.weight.quant_map": "model-00009-of-00013.safetensors",
 
1133
  "model.layers.41.mlp.up_proj.weight.absmax": "model-00009-of-00013.safetensors",
1134
  "model.layers.41.mlp.up_proj.weight.quant_map": "model-00009-of-00013.safetensors",
1135
  "model.layers.41.mlp.up_proj.weight.quant_state.bitsandbytes__fp4": "model-00009-of-00013.safetensors",
1136
+ "model.layers.41.self_attn.k_norm.weight": "model-00009-of-00013.safetensors",
1137
  "model.layers.41.self_attn.k_proj.weight": "model-00009-of-00013.safetensors",
1138
  "model.layers.41.self_attn.k_proj.weight.absmax": "model-00009-of-00013.safetensors",
1139
  "model.layers.41.self_attn.k_proj.weight.quant_map": "model-00009-of-00013.safetensors",
 
1142
  "model.layers.41.self_attn.o_proj.weight.absmax": "model-00009-of-00013.safetensors",
1143
  "model.layers.41.self_attn.o_proj.weight.quant_map": "model-00009-of-00013.safetensors",
1144
  "model.layers.41.self_attn.o_proj.weight.quant_state.bitsandbytes__fp4": "model-00009-of-00013.safetensors",
1145
+ "model.layers.41.self_attn.q_norm.weight": "model-00009-of-00013.safetensors",
1146
  "model.layers.41.self_attn.q_proj.weight": "model-00009-of-00013.safetensors",
1147
  "model.layers.41.self_attn.q_proj.weight.absmax": "model-00009-of-00013.safetensors",
1148
  "model.layers.41.self_attn.q_proj.weight.quant_map": "model-00009-of-00013.safetensors",
 
1164
  "model.layers.42.mlp.up_proj.weight.absmax": "model-00009-of-00013.safetensors",
1165
  "model.layers.42.mlp.up_proj.weight.quant_map": "model-00009-of-00013.safetensors",
1166
  "model.layers.42.mlp.up_proj.weight.quant_state.bitsandbytes__fp4": "model-00009-of-00013.safetensors",
1167
+ "model.layers.42.self_attn.k_norm.weight": "model-00009-of-00013.safetensors",
1168
  "model.layers.42.self_attn.k_proj.weight": "model-00009-of-00013.safetensors",
1169
  "model.layers.42.self_attn.k_proj.weight.absmax": "model-00009-of-00013.safetensors",
1170
  "model.layers.42.self_attn.k_proj.weight.quant_map": "model-00009-of-00013.safetensors",
 
1173
  "model.layers.42.self_attn.o_proj.weight.absmax": "model-00009-of-00013.safetensors",
1174
  "model.layers.42.self_attn.o_proj.weight.quant_map": "model-00009-of-00013.safetensors",
1175
  "model.layers.42.self_attn.o_proj.weight.quant_state.bitsandbytes__fp4": "model-00009-of-00013.safetensors",
1176
+ "model.layers.42.self_attn.q_norm.weight": "model-00009-of-00013.safetensors",
1177
  "model.layers.42.self_attn.q_proj.weight": "model-00009-of-00013.safetensors",
1178
  "model.layers.42.self_attn.q_proj.weight.absmax": "model-00009-of-00013.safetensors",
1179
  "model.layers.42.self_attn.q_proj.weight.quant_map": "model-00009-of-00013.safetensors",
 
1195
  "model.layers.43.mlp.up_proj.weight.absmax": "model-00009-of-00013.safetensors",
1196
  "model.layers.43.mlp.up_proj.weight.quant_map": "model-00009-of-00013.safetensors",
1197
  "model.layers.43.mlp.up_proj.weight.quant_state.bitsandbytes__fp4": "model-00009-of-00013.safetensors",
1198
+ "model.layers.43.self_attn.k_norm.weight": "model-00009-of-00013.safetensors",
1199
  "model.layers.43.self_attn.k_proj.weight": "model-00009-of-00013.safetensors",
1200
  "model.layers.43.self_attn.k_proj.weight.absmax": "model-00009-of-00013.safetensors",
1201
  "model.layers.43.self_attn.k_proj.weight.quant_map": "model-00009-of-00013.safetensors",
 
1204
  "model.layers.43.self_attn.o_proj.weight.absmax": "model-00009-of-00013.safetensors",
1205
  "model.layers.43.self_attn.o_proj.weight.quant_map": "model-00009-of-00013.safetensors",
1206
  "model.layers.43.self_attn.o_proj.weight.quant_state.bitsandbytes__fp4": "model-00009-of-00013.safetensors",
1207
+ "model.layers.43.self_attn.q_norm.weight": "model-00009-of-00013.safetensors",
1208
  "model.layers.43.self_attn.q_proj.weight": "model-00009-of-00013.safetensors",
1209
  "model.layers.43.self_attn.q_proj.weight.absmax": "model-00009-of-00013.safetensors",
1210
  "model.layers.43.self_attn.q_proj.weight.quant_map": "model-00009-of-00013.safetensors",
 
1226
  "model.layers.44.mlp.up_proj.weight.absmax": "model-00010-of-00013.safetensors",
1227
  "model.layers.44.mlp.up_proj.weight.quant_map": "model-00010-of-00013.safetensors",
1228
  "model.layers.44.mlp.up_proj.weight.quant_state.bitsandbytes__fp4": "model-00010-of-00013.safetensors",
1229
+ "model.layers.44.self_attn.k_norm.weight": "model-00009-of-00013.safetensors",
1230
  "model.layers.44.self_attn.k_proj.weight": "model-00009-of-00013.safetensors",
1231
  "model.layers.44.self_attn.k_proj.weight.absmax": "model-00009-of-00013.safetensors",
1232
  "model.layers.44.self_attn.k_proj.weight.quant_map": "model-00009-of-00013.safetensors",
 
1235
  "model.layers.44.self_attn.o_proj.weight.absmax": "model-00009-of-00013.safetensors",
1236
  "model.layers.44.self_attn.o_proj.weight.quant_map": "model-00009-of-00013.safetensors",
1237
  "model.layers.44.self_attn.o_proj.weight.quant_state.bitsandbytes__fp4": "model-00009-of-00013.safetensors",
1238
+ "model.layers.44.self_attn.q_norm.weight": "model-00009-of-00013.safetensors",
1239
  "model.layers.44.self_attn.q_proj.weight": "model-00009-of-00013.safetensors",
1240
  "model.layers.44.self_attn.q_proj.weight.absmax": "model-00009-of-00013.safetensors",
1241
  "model.layers.44.self_attn.q_proj.weight.quant_map": "model-00009-of-00013.safetensors",
 
1257
  "model.layers.45.mlp.up_proj.weight.absmax": "model-00010-of-00013.safetensors",
1258
  "model.layers.45.mlp.up_proj.weight.quant_map": "model-00010-of-00013.safetensors",
1259
  "model.layers.45.mlp.up_proj.weight.quant_state.bitsandbytes__fp4": "model-00010-of-00013.safetensors",
1260
+ "model.layers.45.self_attn.k_norm.weight": "model-00010-of-00013.safetensors",
1261
  "model.layers.45.self_attn.k_proj.weight": "model-00010-of-00013.safetensors",
1262
  "model.layers.45.self_attn.k_proj.weight.absmax": "model-00010-of-00013.safetensors",
1263
  "model.layers.45.self_attn.k_proj.weight.quant_map": "model-00010-of-00013.safetensors",
 
1266
  "model.layers.45.self_attn.o_proj.weight.absmax": "model-00010-of-00013.safetensors",
1267
  "model.layers.45.self_attn.o_proj.weight.quant_map": "model-00010-of-00013.safetensors",
1268
  "model.layers.45.self_attn.o_proj.weight.quant_state.bitsandbytes__fp4": "model-00010-of-00013.safetensors",
1269
+ "model.layers.45.self_attn.q_norm.weight": "model-00010-of-00013.safetensors",
1270
  "model.layers.45.self_attn.q_proj.weight": "model-00010-of-00013.safetensors",
1271
  "model.layers.45.self_attn.q_proj.weight.absmax": "model-00010-of-00013.safetensors",
1272
  "model.layers.45.self_attn.q_proj.weight.quant_map": "model-00010-of-00013.safetensors",
 
1288
  "model.layers.46.mlp.up_proj.weight.absmax": "model-00010-of-00013.safetensors",
1289
  "model.layers.46.mlp.up_proj.weight.quant_map": "model-00010-of-00013.safetensors",
1290
  "model.layers.46.mlp.up_proj.weight.quant_state.bitsandbytes__fp4": "model-00010-of-00013.safetensors",
1291
+ "model.layers.46.self_attn.k_norm.weight": "model-00010-of-00013.safetensors",
1292
  "model.layers.46.self_attn.k_proj.weight": "model-00010-of-00013.safetensors",
1293
  "model.layers.46.self_attn.k_proj.weight.absmax": "model-00010-of-00013.safetensors",
1294
  "model.layers.46.self_attn.k_proj.weight.quant_map": "model-00010-of-00013.safetensors",
 
1297
  "model.layers.46.self_attn.o_proj.weight.absmax": "model-00010-of-00013.safetensors",
1298
  "model.layers.46.self_attn.o_proj.weight.quant_map": "model-00010-of-00013.safetensors",
1299
  "model.layers.46.self_attn.o_proj.weight.quant_state.bitsandbytes__fp4": "model-00010-of-00013.safetensors",
1300
+ "model.layers.46.self_attn.q_norm.weight": "model-00010-of-00013.safetensors",
1301
  "model.layers.46.self_attn.q_proj.weight": "model-00010-of-00013.safetensors",
1302
  "model.layers.46.self_attn.q_proj.weight.absmax": "model-00010-of-00013.safetensors",
1303
  "model.layers.46.self_attn.q_proj.weight.quant_map": "model-00010-of-00013.safetensors",
 
1319
  "model.layers.47.mlp.up_proj.weight.absmax": "model-00010-of-00013.safetensors",
1320
  "model.layers.47.mlp.up_proj.weight.quant_map": "model-00010-of-00013.safetensors",
1321
  "model.layers.47.mlp.up_proj.weight.quant_state.bitsandbytes__fp4": "model-00010-of-00013.safetensors",
1322
+ "model.layers.47.self_attn.k_norm.weight": "model-00010-of-00013.safetensors",
1323
  "model.layers.47.self_attn.k_proj.weight": "model-00010-of-00013.safetensors",
1324
  "model.layers.47.self_attn.k_proj.weight.absmax": "model-00010-of-00013.safetensors",
1325
  "model.layers.47.self_attn.k_proj.weight.quant_map": "model-00010-of-00013.safetensors",
 
1328
  "model.layers.47.self_attn.o_proj.weight.absmax": "model-00010-of-00013.safetensors",
1329
  "model.layers.47.self_attn.o_proj.weight.quant_map": "model-00010-of-00013.safetensors",
1330
  "model.layers.47.self_attn.o_proj.weight.quant_state.bitsandbytes__fp4": "model-00010-of-00013.safetensors",
1331
+ "model.layers.47.self_attn.q_norm.weight": "model-00010-of-00013.safetensors",
1332
  "model.layers.47.self_attn.q_proj.weight": "model-00010-of-00013.safetensors",
1333
  "model.layers.47.self_attn.q_proj.weight.absmax": "model-00010-of-00013.safetensors",
1334
  "model.layers.47.self_attn.q_proj.weight.quant_map": "model-00010-of-00013.safetensors",
 
1350
  "model.layers.48.mlp.up_proj.weight.absmax": "model-00010-of-00013.safetensors",
1351
  "model.layers.48.mlp.up_proj.weight.quant_map": "model-00010-of-00013.safetensors",
1352
  "model.layers.48.mlp.up_proj.weight.quant_state.bitsandbytes__fp4": "model-00010-of-00013.safetensors",
1353
+ "model.layers.48.self_attn.k_norm.weight": "model-00010-of-00013.safetensors",
1354
  "model.layers.48.self_attn.k_proj.weight": "model-00010-of-00013.safetensors",
1355
  "model.layers.48.self_attn.k_proj.weight.absmax": "model-00010-of-00013.safetensors",
1356
  "model.layers.48.self_attn.k_proj.weight.quant_map": "model-00010-of-00013.safetensors",
 
1359
  "model.layers.48.self_attn.o_proj.weight.absmax": "model-00010-of-00013.safetensors",
1360
  "model.layers.48.self_attn.o_proj.weight.quant_map": "model-00010-of-00013.safetensors",
1361
  "model.layers.48.self_attn.o_proj.weight.quant_state.bitsandbytes__fp4": "model-00010-of-00013.safetensors",
1362
+ "model.layers.48.self_attn.q_norm.weight": "model-00010-of-00013.safetensors",
1363
  "model.layers.48.self_attn.q_proj.weight": "model-00010-of-00013.safetensors",
1364
  "model.layers.48.self_attn.q_proj.weight.absmax": "model-00010-of-00013.safetensors",
1365
  "model.layers.48.self_attn.q_proj.weight.quant_map": "model-00010-of-00013.safetensors",
 
1381
  "model.layers.49.mlp.up_proj.weight.absmax": "model-00010-of-00013.safetensors",
1382
  "model.layers.49.mlp.up_proj.weight.quant_map": "model-00010-of-00013.safetensors",
1383
  "model.layers.49.mlp.up_proj.weight.quant_state.bitsandbytes__fp4": "model-00010-of-00013.safetensors",
1384
+ "model.layers.49.self_attn.k_norm.weight": "model-00010-of-00013.safetensors",
1385
  "model.layers.49.self_attn.k_proj.weight": "model-00010-of-00013.safetensors",
1386
  "model.layers.49.self_attn.k_proj.weight.absmax": "model-00010-of-00013.safetensors",
1387
  "model.layers.49.self_attn.k_proj.weight.quant_map": "model-00010-of-00013.safetensors",
 
1390
  "model.layers.49.self_attn.o_proj.weight.absmax": "model-00010-of-00013.safetensors",
1391
  "model.layers.49.self_attn.o_proj.weight.quant_map": "model-00010-of-00013.safetensors",
1392
  "model.layers.49.self_attn.o_proj.weight.quant_state.bitsandbytes__fp4": "model-00010-of-00013.safetensors",
1393
+ "model.layers.49.self_attn.q_norm.weight": "model-00010-of-00013.safetensors",
1394
  "model.layers.49.self_attn.q_proj.weight": "model-00010-of-00013.safetensors",
1395
  "model.layers.49.self_attn.q_proj.weight.absmax": "model-00010-of-00013.safetensors",
1396
  "model.layers.49.self_attn.q_proj.weight.quant_map": "model-00010-of-00013.safetensors",
 
1412
  "model.layers.5.mlp.up_proj.weight.absmax": "model-00003-of-00013.safetensors",
1413
  "model.layers.5.mlp.up_proj.weight.quant_map": "model-00003-of-00013.safetensors",
1414
  "model.layers.5.mlp.up_proj.weight.quant_state.bitsandbytes__fp4": "model-00003-of-00013.safetensors",
1415
+ "model.layers.5.self_attn.k_norm.weight": "model-00002-of-00013.safetensors",
1416
  "model.layers.5.self_attn.k_proj.weight": "model-00002-of-00013.safetensors",
1417
  "model.layers.5.self_attn.k_proj.weight.absmax": "model-00002-of-00013.safetensors",
1418
  "model.layers.5.self_attn.k_proj.weight.quant_map": "model-00002-of-00013.safetensors",
 
1421
  "model.layers.5.self_attn.o_proj.weight.absmax": "model-00002-of-00013.safetensors",
1422
  "model.layers.5.self_attn.o_proj.weight.quant_map": "model-00002-of-00013.safetensors",
1423
  "model.layers.5.self_attn.o_proj.weight.quant_state.bitsandbytes__fp4": "model-00002-of-00013.safetensors",
1424
+ "model.layers.5.self_attn.q_norm.weight": "model-00002-of-00013.safetensors",
1425
  "model.layers.5.self_attn.q_proj.weight": "model-00002-of-00013.safetensors",
1426
  "model.layers.5.self_attn.q_proj.weight.absmax": "model-00002-of-00013.safetensors",
1427
  "model.layers.5.self_attn.q_proj.weight.quant_map": "model-00002-of-00013.safetensors",
 
1443
  "model.layers.50.mlp.up_proj.weight.absmax": "model-00011-of-00013.safetensors",
1444
  "model.layers.50.mlp.up_proj.weight.quant_map": "model-00011-of-00013.safetensors",
1445
  "model.layers.50.mlp.up_proj.weight.quant_state.bitsandbytes__fp4": "model-00011-of-00013.safetensors",
1446
+ "model.layers.50.self_attn.k_norm.weight": "model-00011-of-00013.safetensors",
1447
  "model.layers.50.self_attn.k_proj.weight": "model-00011-of-00013.safetensors",
1448
  "model.layers.50.self_attn.k_proj.weight.absmax": "model-00011-of-00013.safetensors",
1449
  "model.layers.50.self_attn.k_proj.weight.quant_map": "model-00011-of-00013.safetensors",
 
1452
  "model.layers.50.self_attn.o_proj.weight.absmax": "model-00011-of-00013.safetensors",
1453
  "model.layers.50.self_attn.o_proj.weight.quant_map": "model-00011-of-00013.safetensors",
1454
  "model.layers.50.self_attn.o_proj.weight.quant_state.bitsandbytes__fp4": "model-00011-of-00013.safetensors",
1455
+ "model.layers.50.self_attn.q_norm.weight": "model-00011-of-00013.safetensors",
1456
  "model.layers.50.self_attn.q_proj.weight": "model-00011-of-00013.safetensors",
1457
  "model.layers.50.self_attn.q_proj.weight.absmax": "model-00011-of-00013.safetensors",
1458
  "model.layers.50.self_attn.q_proj.weight.quant_map": "model-00011-of-00013.safetensors",
 
1474
  "model.layers.51.mlp.up_proj.weight.absmax": "model-00011-of-00013.safetensors",
1475
  "model.layers.51.mlp.up_proj.weight.quant_map": "model-00011-of-00013.safetensors",
1476
  "model.layers.51.mlp.up_proj.weight.quant_state.bitsandbytes__fp4": "model-00011-of-00013.safetensors",
1477
+ "model.layers.51.self_attn.k_norm.weight": "model-00011-of-00013.safetensors",
1478
  "model.layers.51.self_attn.k_proj.weight": "model-00011-of-00013.safetensors",
1479
  "model.layers.51.self_attn.k_proj.weight.absmax": "model-00011-of-00013.safetensors",
1480
  "model.layers.51.self_attn.k_proj.weight.quant_map": "model-00011-of-00013.safetensors",
 
1483
  "model.layers.51.self_attn.o_proj.weight.absmax": "model-00011-of-00013.safetensors",
1484
  "model.layers.51.self_attn.o_proj.weight.quant_map": "model-00011-of-00013.safetensors",
1485
  "model.layers.51.self_attn.o_proj.weight.quant_state.bitsandbytes__fp4": "model-00011-of-00013.safetensors",
1486
+ "model.layers.51.self_attn.q_norm.weight": "model-00011-of-00013.safetensors",
1487
  "model.layers.51.self_attn.q_proj.weight": "model-00011-of-00013.safetensors",
1488
  "model.layers.51.self_attn.q_proj.weight.absmax": "model-00011-of-00013.safetensors",
1489
  "model.layers.51.self_attn.q_proj.weight.quant_map": "model-00011-of-00013.safetensors",
 
1505
  "model.layers.52.mlp.up_proj.weight.absmax": "model-00011-of-00013.safetensors",
1506
  "model.layers.52.mlp.up_proj.weight.quant_map": "model-00011-of-00013.safetensors",
1507
  "model.layers.52.mlp.up_proj.weight.quant_state.bitsandbytes__fp4": "model-00011-of-00013.safetensors",
1508
+ "model.layers.52.self_attn.k_norm.weight": "model-00011-of-00013.safetensors",
1509
  "model.layers.52.self_attn.k_proj.weight": "model-00011-of-00013.safetensors",
1510
  "model.layers.52.self_attn.k_proj.weight.absmax": "model-00011-of-00013.safetensors",
1511
  "model.layers.52.self_attn.k_proj.weight.quant_map": "model-00011-of-00013.safetensors",
 
1514
  "model.layers.52.self_attn.o_proj.weight.absmax": "model-00011-of-00013.safetensors",
1515
  "model.layers.52.self_attn.o_proj.weight.quant_map": "model-00011-of-00013.safetensors",
1516
  "model.layers.52.self_attn.o_proj.weight.quant_state.bitsandbytes__fp4": "model-00011-of-00013.safetensors",
1517
+ "model.layers.52.self_attn.q_norm.weight": "model-00011-of-00013.safetensors",
1518
  "model.layers.52.self_attn.q_proj.weight": "model-00011-of-00013.safetensors",
1519
  "model.layers.52.self_attn.q_proj.weight.absmax": "model-00011-of-00013.safetensors",
1520
  "model.layers.52.self_attn.q_proj.weight.quant_map": "model-00011-of-00013.safetensors",
 
1536
  "model.layers.53.mlp.up_proj.weight.absmax": "model-00011-of-00013.safetensors",
1537
  "model.layers.53.mlp.up_proj.weight.quant_map": "model-00011-of-00013.safetensors",
1538
  "model.layers.53.mlp.up_proj.weight.quant_state.bitsandbytes__fp4": "model-00011-of-00013.safetensors",
1539
+ "model.layers.53.self_attn.k_norm.weight": "model-00011-of-00013.safetensors",
1540
  "model.layers.53.self_attn.k_proj.weight": "model-00011-of-00013.safetensors",
1541
  "model.layers.53.self_attn.k_proj.weight.absmax": "model-00011-of-00013.safetensors",
1542
  "model.layers.53.self_attn.k_proj.weight.quant_map": "model-00011-of-00013.safetensors",
 
1545
  "model.layers.53.self_attn.o_proj.weight.absmax": "model-00011-of-00013.safetensors",
1546
  "model.layers.53.self_attn.o_proj.weight.quant_map": "model-00011-of-00013.safetensors",
1547
  "model.layers.53.self_attn.o_proj.weight.quant_state.bitsandbytes__fp4": "model-00011-of-00013.safetensors",
1548
+ "model.layers.53.self_attn.q_norm.weight": "model-00011-of-00013.safetensors",
1549
  "model.layers.53.self_attn.q_proj.weight": "model-00011-of-00013.safetensors",
1550
  "model.layers.53.self_attn.q_proj.weight.absmax": "model-00011-of-00013.safetensors",
1551
  "model.layers.53.self_attn.q_proj.weight.quant_map": "model-00011-of-00013.safetensors",
 
1567
  "model.layers.54.mlp.up_proj.weight.absmax": "model-00011-of-00013.safetensors",
1568
  "model.layers.54.mlp.up_proj.weight.quant_map": "model-00011-of-00013.safetensors",
1569
  "model.layers.54.mlp.up_proj.weight.quant_state.bitsandbytes__fp4": "model-00011-of-00013.safetensors",
1570
+ "model.layers.54.self_attn.k_norm.weight": "model-00011-of-00013.safetensors",
1571
  "model.layers.54.self_attn.k_proj.weight": "model-00011-of-00013.safetensors",
1572
  "model.layers.54.self_attn.k_proj.weight.absmax": "model-00011-of-00013.safetensors",
1573
  "model.layers.54.self_attn.k_proj.weight.quant_map": "model-00011-of-00013.safetensors",
 
1576
  "model.layers.54.self_attn.o_proj.weight.absmax": "model-00011-of-00013.safetensors",
1577
  "model.layers.54.self_attn.o_proj.weight.quant_map": "model-00011-of-00013.safetensors",
1578
  "model.layers.54.self_attn.o_proj.weight.quant_state.bitsandbytes__fp4": "model-00011-of-00013.safetensors",
1579
+ "model.layers.54.self_attn.q_norm.weight": "model-00011-of-00013.safetensors",
1580
  "model.layers.54.self_attn.q_proj.weight": "model-00011-of-00013.safetensors",
1581
  "model.layers.54.self_attn.q_proj.weight.absmax": "model-00011-of-00013.safetensors",
1582
  "model.layers.54.self_attn.q_proj.weight.quant_map": "model-00011-of-00013.safetensors",
 
1598
  "model.layers.55.mlp.up_proj.weight.absmax": "model-00012-of-00013.safetensors",
1599
  "model.layers.55.mlp.up_proj.weight.quant_map": "model-00012-of-00013.safetensors",
1600
  "model.layers.55.mlp.up_proj.weight.quant_state.bitsandbytes__fp4": "model-00012-of-00013.safetensors",
1601
+ "model.layers.55.self_attn.k_norm.weight": "model-00011-of-00013.safetensors",
1602
  "model.layers.55.self_attn.k_proj.weight": "model-00011-of-00013.safetensors",
1603
  "model.layers.55.self_attn.k_proj.weight.absmax": "model-00011-of-00013.safetensors",
1604
  "model.layers.55.self_attn.k_proj.weight.quant_map": "model-00011-of-00013.safetensors",
 
1607
  "model.layers.55.self_attn.o_proj.weight.absmax": "model-00011-of-00013.safetensors",
1608
  "model.layers.55.self_attn.o_proj.weight.quant_map": "model-00011-of-00013.safetensors",
1609
  "model.layers.55.self_attn.o_proj.weight.quant_state.bitsandbytes__fp4": "model-00011-of-00013.safetensors",
1610
+ "model.layers.55.self_attn.q_norm.weight": "model-00011-of-00013.safetensors",
1611
  "model.layers.55.self_attn.q_proj.weight": "model-00011-of-00013.safetensors",
1612
  "model.layers.55.self_attn.q_proj.weight.absmax": "model-00011-of-00013.safetensors",
1613
  "model.layers.55.self_attn.q_proj.weight.quant_map": "model-00011-of-00013.safetensors",
 
1629
  "model.layers.56.mlp.up_proj.weight.absmax": "model-00012-of-00013.safetensors",
1630
  "model.layers.56.mlp.up_proj.weight.quant_map": "model-00012-of-00013.safetensors",
1631
  "model.layers.56.mlp.up_proj.weight.quant_state.bitsandbytes__fp4": "model-00012-of-00013.safetensors",
1632
+ "model.layers.56.self_attn.k_norm.weight": "model-00012-of-00013.safetensors",
1633
  "model.layers.56.self_attn.k_proj.weight": "model-00012-of-00013.safetensors",
1634
  "model.layers.56.self_attn.k_proj.weight.absmax": "model-00012-of-00013.safetensors",
1635
  "model.layers.56.self_attn.k_proj.weight.quant_map": "model-00012-of-00013.safetensors",
 
1638
  "model.layers.56.self_attn.o_proj.weight.absmax": "model-00012-of-00013.safetensors",
1639
  "model.layers.56.self_attn.o_proj.weight.quant_map": "model-00012-of-00013.safetensors",
1640
  "model.layers.56.self_attn.o_proj.weight.quant_state.bitsandbytes__fp4": "model-00012-of-00013.safetensors",
1641
+ "model.layers.56.self_attn.q_norm.weight": "model-00012-of-00013.safetensors",
1642
  "model.layers.56.self_attn.q_proj.weight": "model-00012-of-00013.safetensors",
1643
  "model.layers.56.self_attn.q_proj.weight.absmax": "model-00012-of-00013.safetensors",
1644
  "model.layers.56.self_attn.q_proj.weight.quant_map": "model-00012-of-00013.safetensors",
 
1660
  "model.layers.57.mlp.up_proj.weight.absmax": "model-00012-of-00013.safetensors",
1661
  "model.layers.57.mlp.up_proj.weight.quant_map": "model-00012-of-00013.safetensors",
1662
  "model.layers.57.mlp.up_proj.weight.quant_state.bitsandbytes__fp4": "model-00012-of-00013.safetensors",
1663
+ "model.layers.57.self_attn.k_norm.weight": "model-00012-of-00013.safetensors",
1664
  "model.layers.57.self_attn.k_proj.weight": "model-00012-of-00013.safetensors",
1665
  "model.layers.57.self_attn.k_proj.weight.absmax": "model-00012-of-00013.safetensors",
1666
  "model.layers.57.self_attn.k_proj.weight.quant_map": "model-00012-of-00013.safetensors",
 
1669
  "model.layers.57.self_attn.o_proj.weight.absmax": "model-00012-of-00013.safetensors",
1670
  "model.layers.57.self_attn.o_proj.weight.quant_map": "model-00012-of-00013.safetensors",
1671
  "model.layers.57.self_attn.o_proj.weight.quant_state.bitsandbytes__fp4": "model-00012-of-00013.safetensors",
1672
+ "model.layers.57.self_attn.q_norm.weight": "model-00012-of-00013.safetensors",
1673
  "model.layers.57.self_attn.q_proj.weight": "model-00012-of-00013.safetensors",
1674
  "model.layers.57.self_attn.q_proj.weight.absmax": "model-00012-of-00013.safetensors",
1675
  "model.layers.57.self_attn.q_proj.weight.quant_map": "model-00012-of-00013.safetensors",
 
1691
  "model.layers.58.mlp.up_proj.weight.absmax": "model-00012-of-00013.safetensors",
1692
  "model.layers.58.mlp.up_proj.weight.quant_map": "model-00012-of-00013.safetensors",
1693
  "model.layers.58.mlp.up_proj.weight.quant_state.bitsandbytes__fp4": "model-00012-of-00013.safetensors",
1694
+ "model.layers.58.self_attn.k_norm.weight": "model-00012-of-00013.safetensors",
1695
  "model.layers.58.self_attn.k_proj.weight": "model-00012-of-00013.safetensors",
1696
  "model.layers.58.self_attn.k_proj.weight.absmax": "model-00012-of-00013.safetensors",
1697
  "model.layers.58.self_attn.k_proj.weight.quant_map": "model-00012-of-00013.safetensors",
 
1700
  "model.layers.58.self_attn.o_proj.weight.absmax": "model-00012-of-00013.safetensors",
1701
  "model.layers.58.self_attn.o_proj.weight.quant_map": "model-00012-of-00013.safetensors",
1702
  "model.layers.58.self_attn.o_proj.weight.quant_state.bitsandbytes__fp4": "model-00012-of-00013.safetensors",
1703
+ "model.layers.58.self_attn.q_norm.weight": "model-00012-of-00013.safetensors",
1704
  "model.layers.58.self_attn.q_proj.weight": "model-00012-of-00013.safetensors",
1705
  "model.layers.58.self_attn.q_proj.weight.absmax": "model-00012-of-00013.safetensors",
1706
  "model.layers.58.self_attn.q_proj.weight.quant_map": "model-00012-of-00013.safetensors",
 
1722
  "model.layers.59.mlp.up_proj.weight.absmax": "model-00012-of-00013.safetensors",
1723
  "model.layers.59.mlp.up_proj.weight.quant_map": "model-00012-of-00013.safetensors",
1724
  "model.layers.59.mlp.up_proj.weight.quant_state.bitsandbytes__fp4": "model-00012-of-00013.safetensors",
1725
+ "model.layers.59.self_attn.k_norm.weight": "model-00012-of-00013.safetensors",
1726
  "model.layers.59.self_attn.k_proj.weight": "model-00012-of-00013.safetensors",
1727
  "model.layers.59.self_attn.k_proj.weight.absmax": "model-00012-of-00013.safetensors",
1728
  "model.layers.59.self_attn.k_proj.weight.quant_map": "model-00012-of-00013.safetensors",
 
1731
  "model.layers.59.self_attn.o_proj.weight.absmax": "model-00012-of-00013.safetensors",
1732
  "model.layers.59.self_attn.o_proj.weight.quant_map": "model-00012-of-00013.safetensors",
1733
  "model.layers.59.self_attn.o_proj.weight.quant_state.bitsandbytes__fp4": "model-00012-of-00013.safetensors",
1734
+ "model.layers.59.self_attn.q_norm.weight": "model-00012-of-00013.safetensors",
1735
  "model.layers.59.self_attn.q_proj.weight": "model-00012-of-00013.safetensors",
1736
  "model.layers.59.self_attn.q_proj.weight.absmax": "model-00012-of-00013.safetensors",
1737
  "model.layers.59.self_attn.q_proj.weight.quant_map": "model-00012-of-00013.safetensors",
 
1753
  "model.layers.6.mlp.up_proj.weight.absmax": "model-00003-of-00013.safetensors",
1754
  "model.layers.6.mlp.up_proj.weight.quant_map": "model-00003-of-00013.safetensors",
1755
  "model.layers.6.mlp.up_proj.weight.quant_state.bitsandbytes__fp4": "model-00003-of-00013.safetensors",
1756
+ "model.layers.6.self_attn.k_norm.weight": "model-00003-of-00013.safetensors",
1757
  "model.layers.6.self_attn.k_proj.weight": "model-00003-of-00013.safetensors",
1758
  "model.layers.6.self_attn.k_proj.weight.absmax": "model-00003-of-00013.safetensors",
1759
  "model.layers.6.self_attn.k_proj.weight.quant_map": "model-00003-of-00013.safetensors",
 
1762
  "model.layers.6.self_attn.o_proj.weight.absmax": "model-00003-of-00013.safetensors",
1763
  "model.layers.6.self_attn.o_proj.weight.quant_map": "model-00003-of-00013.safetensors",
1764
  "model.layers.6.self_attn.o_proj.weight.quant_state.bitsandbytes__fp4": "model-00003-of-00013.safetensors",
1765
+ "model.layers.6.self_attn.q_norm.weight": "model-00003-of-00013.safetensors",
1766
  "model.layers.6.self_attn.q_proj.weight": "model-00003-of-00013.safetensors",
1767
  "model.layers.6.self_attn.q_proj.weight.absmax": "model-00003-of-00013.safetensors",
1768
  "model.layers.6.self_attn.q_proj.weight.quant_map": "model-00003-of-00013.safetensors",
 
1784
  "model.layers.60.mlp.up_proj.weight.absmax": "model-00012-of-00013.safetensors",
1785
  "model.layers.60.mlp.up_proj.weight.quant_map": "model-00012-of-00013.safetensors",
1786
  "model.layers.60.mlp.up_proj.weight.quant_state.bitsandbytes__fp4": "model-00012-of-00013.safetensors",
1787
+ "model.layers.60.self_attn.k_norm.weight": "model-00012-of-00013.safetensors",
1788
  "model.layers.60.self_attn.k_proj.weight": "model-00012-of-00013.safetensors",
1789
  "model.layers.60.self_attn.k_proj.weight.absmax": "model-00012-of-00013.safetensors",
1790
  "model.layers.60.self_attn.k_proj.weight.quant_map": "model-00012-of-00013.safetensors",
 
1793
  "model.layers.60.self_attn.o_proj.weight.absmax": "model-00012-of-00013.safetensors",
1794
  "model.layers.60.self_attn.o_proj.weight.quant_map": "model-00012-of-00013.safetensors",
1795
  "model.layers.60.self_attn.o_proj.weight.quant_state.bitsandbytes__fp4": "model-00012-of-00013.safetensors",
1796
+ "model.layers.60.self_attn.q_norm.weight": "model-00012-of-00013.safetensors",
1797
  "model.layers.60.self_attn.q_proj.weight": "model-00012-of-00013.safetensors",
1798
  "model.layers.60.self_attn.q_proj.weight.absmax": "model-00012-of-00013.safetensors",
1799
  "model.layers.60.self_attn.q_proj.weight.quant_map": "model-00012-of-00013.safetensors",
 
1815
  "model.layers.61.mlp.up_proj.weight.absmax": "model-00013-of-00013.safetensors",
1816
  "model.layers.61.mlp.up_proj.weight.quant_map": "model-00013-of-00013.safetensors",
1817
  "model.layers.61.mlp.up_proj.weight.quant_state.bitsandbytes__fp4": "model-00013-of-00013.safetensors",
1818
+ "model.layers.61.self_attn.k_norm.weight": "model-00013-of-00013.safetensors",
1819
  "model.layers.61.self_attn.k_proj.weight": "model-00013-of-00013.safetensors",
1820
  "model.layers.61.self_attn.k_proj.weight.absmax": "model-00013-of-00013.safetensors",
1821
  "model.layers.61.self_attn.k_proj.weight.quant_map": "model-00013-of-00013.safetensors",
 
1824
  "model.layers.61.self_attn.o_proj.weight.absmax": "model-00013-of-00013.safetensors",
1825
  "model.layers.61.self_attn.o_proj.weight.quant_map": "model-00013-of-00013.safetensors",
1826
  "model.layers.61.self_attn.o_proj.weight.quant_state.bitsandbytes__fp4": "model-00013-of-00013.safetensors",
1827
+ "model.layers.61.self_attn.q_norm.weight": "model-00013-of-00013.safetensors",
1828
  "model.layers.61.self_attn.q_proj.weight": "model-00013-of-00013.safetensors",
1829
  "model.layers.61.self_attn.q_proj.weight.absmax": "model-00013-of-00013.safetensors",
1830
  "model.layers.61.self_attn.q_proj.weight.quant_map": "model-00013-of-00013.safetensors",
 
1846
  "model.layers.62.mlp.up_proj.weight.absmax": "model-00013-of-00013.safetensors",
1847
  "model.layers.62.mlp.up_proj.weight.quant_map": "model-00013-of-00013.safetensors",
1848
  "model.layers.62.mlp.up_proj.weight.quant_state.bitsandbytes__fp4": "model-00013-of-00013.safetensors",
1849
+ "model.layers.62.self_attn.k_norm.weight": "model-00013-of-00013.safetensors",
1850
  "model.layers.62.self_attn.k_proj.weight": "model-00013-of-00013.safetensors",
1851
  "model.layers.62.self_attn.k_proj.weight.absmax": "model-00013-of-00013.safetensors",
1852
  "model.layers.62.self_attn.k_proj.weight.quant_map": "model-00013-of-00013.safetensors",
 
1855
  "model.layers.62.self_attn.o_proj.weight.absmax": "model-00013-of-00013.safetensors",
1856
  "model.layers.62.self_attn.o_proj.weight.quant_map": "model-00013-of-00013.safetensors",
1857
  "model.layers.62.self_attn.o_proj.weight.quant_state.bitsandbytes__fp4": "model-00013-of-00013.safetensors",
1858
+ "model.layers.62.self_attn.q_norm.weight": "model-00013-of-00013.safetensors",
1859
  "model.layers.62.self_attn.q_proj.weight": "model-00013-of-00013.safetensors",
1860
  "model.layers.62.self_attn.q_proj.weight.absmax": "model-00013-of-00013.safetensors",
1861
  "model.layers.62.self_attn.q_proj.weight.quant_map": "model-00013-of-00013.safetensors",
 
1877
  "model.layers.63.mlp.up_proj.weight.absmax": "model-00013-of-00013.safetensors",
1878
  "model.layers.63.mlp.up_proj.weight.quant_map": "model-00013-of-00013.safetensors",
1879
  "model.layers.63.mlp.up_proj.weight.quant_state.bitsandbytes__fp4": "model-00013-of-00013.safetensors",
1880
+ "model.layers.63.self_attn.k_norm.weight": "model-00013-of-00013.safetensors",
1881
  "model.layers.63.self_attn.k_proj.weight": "model-00013-of-00013.safetensors",
1882
  "model.layers.63.self_attn.k_proj.weight.absmax": "model-00013-of-00013.safetensors",
1883
  "model.layers.63.self_attn.k_proj.weight.quant_map": "model-00013-of-00013.safetensors",
 
1886
  "model.layers.63.self_attn.o_proj.weight.absmax": "model-00013-of-00013.safetensors",
1887
  "model.layers.63.self_attn.o_proj.weight.quant_map": "model-00013-of-00013.safetensors",
1888
  "model.layers.63.self_attn.o_proj.weight.quant_state.bitsandbytes__fp4": "model-00013-of-00013.safetensors",
1889
+ "model.layers.63.self_attn.q_norm.weight": "model-00013-of-00013.safetensors",
1890
  "model.layers.63.self_attn.q_proj.weight": "model-00013-of-00013.safetensors",
1891
  "model.layers.63.self_attn.q_proj.weight.absmax": "model-00013-of-00013.safetensors",
1892
  "model.layers.63.self_attn.q_proj.weight.quant_map": "model-00013-of-00013.safetensors",
 
1908
  "model.layers.7.mlp.up_proj.weight.absmax": "model-00003-of-00013.safetensors",
1909
  "model.layers.7.mlp.up_proj.weight.quant_map": "model-00003-of-00013.safetensors",
1910
  "model.layers.7.mlp.up_proj.weight.quant_state.bitsandbytes__fp4": "model-00003-of-00013.safetensors",
1911
+ "model.layers.7.self_attn.k_norm.weight": "model-00003-of-00013.safetensors",
1912
  "model.layers.7.self_attn.k_proj.weight": "model-00003-of-00013.safetensors",
1913
  "model.layers.7.self_attn.k_proj.weight.absmax": "model-00003-of-00013.safetensors",
1914
  "model.layers.7.self_attn.k_proj.weight.quant_map": "model-00003-of-00013.safetensors",
 
1917
  "model.layers.7.self_attn.o_proj.weight.absmax": "model-00003-of-00013.safetensors",
1918
  "model.layers.7.self_attn.o_proj.weight.quant_map": "model-00003-of-00013.safetensors",
1919
  "model.layers.7.self_attn.o_proj.weight.quant_state.bitsandbytes__fp4": "model-00003-of-00013.safetensors",
1920
+ "model.layers.7.self_attn.q_norm.weight": "model-00003-of-00013.safetensors",
1921
  "model.layers.7.self_attn.q_proj.weight": "model-00003-of-00013.safetensors",
1922
  "model.layers.7.self_attn.q_proj.weight.absmax": "model-00003-of-00013.safetensors",
1923
  "model.layers.7.self_attn.q_proj.weight.quant_map": "model-00003-of-00013.safetensors",
 
1939
  "model.layers.8.mlp.up_proj.weight.absmax": "model-00003-of-00013.safetensors",
1940
  "model.layers.8.mlp.up_proj.weight.quant_map": "model-00003-of-00013.safetensors",
1941
  "model.layers.8.mlp.up_proj.weight.quant_state.bitsandbytes__fp4": "model-00003-of-00013.safetensors",
1942
+ "model.layers.8.self_attn.k_norm.weight": "model-00003-of-00013.safetensors",
1943
  "model.layers.8.self_attn.k_proj.weight": "model-00003-of-00013.safetensors",
1944
  "model.layers.8.self_attn.k_proj.weight.absmax": "model-00003-of-00013.safetensors",
1945
  "model.layers.8.self_attn.k_proj.weight.quant_map": "model-00003-of-00013.safetensors",
 
1948
  "model.layers.8.self_attn.o_proj.weight.absmax": "model-00003-of-00013.safetensors",
1949
  "model.layers.8.self_attn.o_proj.weight.quant_map": "model-00003-of-00013.safetensors",
1950
  "model.layers.8.self_attn.o_proj.weight.quant_state.bitsandbytes__fp4": "model-00003-of-00013.safetensors",
1951
+ "model.layers.8.self_attn.q_norm.weight": "model-00003-of-00013.safetensors",
1952
  "model.layers.8.self_attn.q_proj.weight": "model-00003-of-00013.safetensors",
1953
  "model.layers.8.self_attn.q_proj.weight.absmax": "model-00003-of-00013.safetensors",
1954
  "model.layers.8.self_attn.q_proj.weight.quant_map": "model-00003-of-00013.safetensors",
 
1970
  "model.layers.9.mlp.up_proj.weight.absmax": "model-00003-of-00013.safetensors",
1971
  "model.layers.9.mlp.up_proj.weight.quant_map": "model-00003-of-00013.safetensors",
1972
  "model.layers.9.mlp.up_proj.weight.quant_state.bitsandbytes__fp4": "model-00003-of-00013.safetensors",
1973
+ "model.layers.9.self_attn.k_norm.weight": "model-00003-of-00013.safetensors",
1974
  "model.layers.9.self_attn.k_proj.weight": "model-00003-of-00013.safetensors",
1975
  "model.layers.9.self_attn.k_proj.weight.absmax": "model-00003-of-00013.safetensors",
1976
  "model.layers.9.self_attn.k_proj.weight.quant_map": "model-00003-of-00013.safetensors",
 
1979
  "model.layers.9.self_attn.o_proj.weight.absmax": "model-00003-of-00013.safetensors",
1980
  "model.layers.9.self_attn.o_proj.weight.quant_map": "model-00003-of-00013.safetensors",
1981
  "model.layers.9.self_attn.o_proj.weight.quant_state.bitsandbytes__fp4": "model-00003-of-00013.safetensors",
1982
+ "model.layers.9.self_attn.q_norm.weight": "model-00003-of-00013.safetensors",
1983
  "model.layers.9.self_attn.q_proj.weight": "model-00003-of-00013.safetensors",
1984
  "model.layers.9.self_attn.q_proj.weight.absmax": "model-00003-of-00013.safetensors",
1985
  "model.layers.9.self_attn.q_proj.weight.quant_map": "model-00003-of-00013.safetensors",