Adding `safetensors` variant of this model
#3
by
SFconvertbot
- opened
- model.safetensors.index.json +309 -0
- model_00001-of-00032.safetensors +3 -0
- model_00002-of-00032.safetensors +3 -0
- model_00003-of-00032.safetensors +3 -0
- model_00004-of-00032.safetensors +3 -0
- model_00005-of-00032.safetensors +3 -0
- model_00006-of-00032.safetensors +3 -0
- model_00007-of-00032.safetensors +3 -0
- model_00008-of-00032.safetensors +3 -0
- model_00009-of-00032.safetensors +3 -0
- model_00010-of-00032.safetensors +3 -0
- model_00011-of-00032.safetensors +3 -0
- model_00012-of-00032.safetensors +3 -0
- model_00013-of-00032.safetensors +3 -0
- model_00014-of-00032.safetensors +3 -0
- model_00015-of-00032.safetensors +3 -0
- model_00016-of-00032.safetensors +3 -0
- model_00017-of-00032.safetensors +3 -0
- model_00018-of-00032.safetensors +3 -0
- model_00019-of-00032.safetensors +3 -0
- model_00020-of-00032.safetensors +3 -0
- model_00021-of-00032.safetensors +3 -0
- model_00022-of-00032.safetensors +3 -0
- model_00023-of-00032.safetensors +3 -0
- model_00024-of-00032.safetensors +3 -0
- model_00025-of-00032.safetensors +3 -0
- model_00026-of-00032.safetensors +3 -0
- model_00027-of-00032.safetensors +3 -0
- model_00028-of-00032.safetensors +3 -0
- model_00029-of-00032.safetensors +3 -0
- model_00030-of-00032.safetensors +3 -0
- model_00031-of-00032.safetensors +3 -0
- model_00032-of-00032.safetensors +3 -0
model.safetensors.index.json
ADDED
@@ -0,0 +1,309 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"metadata": {
|
3 |
+
"total_size": 14400036864
|
4 |
+
},
|
5 |
+
"weight_map": {
|
6 |
+
"h.0.input_layernorm.weight": "model_00002-of-00032.safetensors",
|
7 |
+
"h.0.mlp.down_proj.bias": "model_00002-of-00032.safetensors",
|
8 |
+
"h.0.mlp.down_proj.weight": "model_00002-of-00032.safetensors",
|
9 |
+
"h.0.mlp.gate_proj.weight": "model_00002-of-00032.safetensors",
|
10 |
+
"h.0.mlp.up_proj.weight": "model_00002-of-00032.safetensors",
|
11 |
+
"h.0.post_attention_layernorm.weight": "model_00002-of-00032.safetensors",
|
12 |
+
"h.0.self_attention.dense.bias": "model_00002-of-00032.safetensors",
|
13 |
+
"h.0.self_attention.dense.weight": "model_00002-of-00032.safetensors",
|
14 |
+
"h.0.self_attention.key_value.weight": "model_00002-of-00032.safetensors",
|
15 |
+
"h.0.self_attention.query.weight": "model_00002-of-00032.safetensors",
|
16 |
+
"h.1.input_layernorm.weight": "model_00003-of-00032.safetensors",
|
17 |
+
"h.1.mlp.down_proj.bias": "model_00003-of-00032.safetensors",
|
18 |
+
"h.1.mlp.down_proj.weight": "model_00003-of-00032.safetensors",
|
19 |
+
"h.1.mlp.gate_proj.weight": "model_00003-of-00032.safetensors",
|
20 |
+
"h.1.mlp.up_proj.weight": "model_00003-of-00032.safetensors",
|
21 |
+
"h.1.post_attention_layernorm.weight": "model_00003-of-00032.safetensors",
|
22 |
+
"h.1.self_attention.dense.bias": "model_00003-of-00032.safetensors",
|
23 |
+
"h.1.self_attention.dense.weight": "model_00003-of-00032.safetensors",
|
24 |
+
"h.1.self_attention.key_value.weight": "model_00003-of-00032.safetensors",
|
25 |
+
"h.1.self_attention.query.weight": "model_00003-of-00032.safetensors",
|
26 |
+
"h.10.input_layernorm.weight": "model_00012-of-00032.safetensors",
|
27 |
+
"h.10.mlp.down_proj.bias": "model_00012-of-00032.safetensors",
|
28 |
+
"h.10.mlp.down_proj.weight": "model_00012-of-00032.safetensors",
|
29 |
+
"h.10.mlp.gate_proj.weight": "model_00012-of-00032.safetensors",
|
30 |
+
"h.10.mlp.up_proj.weight": "model_00012-of-00032.safetensors",
|
31 |
+
"h.10.post_attention_layernorm.weight": "model_00012-of-00032.safetensors",
|
32 |
+
"h.10.self_attention.dense.bias": "model_00012-of-00032.safetensors",
|
33 |
+
"h.10.self_attention.dense.weight": "model_00012-of-00032.safetensors",
|
34 |
+
"h.10.self_attention.key_value.weight": "model_00012-of-00032.safetensors",
|
35 |
+
"h.10.self_attention.query.weight": "model_00012-of-00032.safetensors",
|
36 |
+
"h.11.input_layernorm.weight": "model_00013-of-00032.safetensors",
|
37 |
+
"h.11.mlp.down_proj.bias": "model_00013-of-00032.safetensors",
|
38 |
+
"h.11.mlp.down_proj.weight": "model_00013-of-00032.safetensors",
|
39 |
+
"h.11.mlp.gate_proj.weight": "model_00013-of-00032.safetensors",
|
40 |
+
"h.11.mlp.up_proj.weight": "model_00013-of-00032.safetensors",
|
41 |
+
"h.11.post_attention_layernorm.weight": "model_00013-of-00032.safetensors",
|
42 |
+
"h.11.self_attention.dense.bias": "model_00013-of-00032.safetensors",
|
43 |
+
"h.11.self_attention.dense.weight": "model_00013-of-00032.safetensors",
|
44 |
+
"h.11.self_attention.key_value.weight": "model_00013-of-00032.safetensors",
|
45 |
+
"h.11.self_attention.query.weight": "model_00013-of-00032.safetensors",
|
46 |
+
"h.12.input_layernorm.weight": "model_00014-of-00032.safetensors",
|
47 |
+
"h.12.mlp.down_proj.bias": "model_00014-of-00032.safetensors",
|
48 |
+
"h.12.mlp.down_proj.weight": "model_00014-of-00032.safetensors",
|
49 |
+
"h.12.mlp.gate_proj.weight": "model_00014-of-00032.safetensors",
|
50 |
+
"h.12.mlp.up_proj.weight": "model_00014-of-00032.safetensors",
|
51 |
+
"h.12.post_attention_layernorm.weight": "model_00014-of-00032.safetensors",
|
52 |
+
"h.12.self_attention.dense.bias": "model_00014-of-00032.safetensors",
|
53 |
+
"h.12.self_attention.dense.weight": "model_00014-of-00032.safetensors",
|
54 |
+
"h.12.self_attention.key_value.weight": "model_00014-of-00032.safetensors",
|
55 |
+
"h.12.self_attention.query.weight": "model_00014-of-00032.safetensors",
|
56 |
+
"h.13.input_layernorm.weight": "model_00015-of-00032.safetensors",
|
57 |
+
"h.13.mlp.down_proj.bias": "model_00015-of-00032.safetensors",
|
58 |
+
"h.13.mlp.down_proj.weight": "model_00015-of-00032.safetensors",
|
59 |
+
"h.13.mlp.gate_proj.weight": "model_00015-of-00032.safetensors",
|
60 |
+
"h.13.mlp.up_proj.weight": "model_00015-of-00032.safetensors",
|
61 |
+
"h.13.post_attention_layernorm.weight": "model_00015-of-00032.safetensors",
|
62 |
+
"h.13.self_attention.dense.bias": "model_00015-of-00032.safetensors",
|
63 |
+
"h.13.self_attention.dense.weight": "model_00015-of-00032.safetensors",
|
64 |
+
"h.13.self_attention.key_value.weight": "model_00015-of-00032.safetensors",
|
65 |
+
"h.13.self_attention.query.weight": "model_00015-of-00032.safetensors",
|
66 |
+
"h.14.input_layernorm.weight": "model_00016-of-00032.safetensors",
|
67 |
+
"h.14.mlp.down_proj.bias": "model_00016-of-00032.safetensors",
|
68 |
+
"h.14.mlp.down_proj.weight": "model_00016-of-00032.safetensors",
|
69 |
+
"h.14.mlp.gate_proj.weight": "model_00016-of-00032.safetensors",
|
70 |
+
"h.14.mlp.up_proj.weight": "model_00016-of-00032.safetensors",
|
71 |
+
"h.14.post_attention_layernorm.weight": "model_00016-of-00032.safetensors",
|
72 |
+
"h.14.self_attention.dense.bias": "model_00016-of-00032.safetensors",
|
73 |
+
"h.14.self_attention.dense.weight": "model_00016-of-00032.safetensors",
|
74 |
+
"h.14.self_attention.key_value.weight": "model_00016-of-00032.safetensors",
|
75 |
+
"h.14.self_attention.query.weight": "model_00016-of-00032.safetensors",
|
76 |
+
"h.15.input_layernorm.weight": "model_00017-of-00032.safetensors",
|
77 |
+
"h.15.mlp.down_proj.bias": "model_00017-of-00032.safetensors",
|
78 |
+
"h.15.mlp.down_proj.weight": "model_00017-of-00032.safetensors",
|
79 |
+
"h.15.mlp.gate_proj.weight": "model_00017-of-00032.safetensors",
|
80 |
+
"h.15.mlp.up_proj.weight": "model_00017-of-00032.safetensors",
|
81 |
+
"h.15.post_attention_layernorm.weight": "model_00017-of-00032.safetensors",
|
82 |
+
"h.15.self_attention.dense.bias": "model_00017-of-00032.safetensors",
|
83 |
+
"h.15.self_attention.dense.weight": "model_00017-of-00032.safetensors",
|
84 |
+
"h.15.self_attention.key_value.weight": "model_00017-of-00032.safetensors",
|
85 |
+
"h.15.self_attention.query.weight": "model_00017-of-00032.safetensors",
|
86 |
+
"h.16.input_layernorm.weight": "model_00018-of-00032.safetensors",
|
87 |
+
"h.16.mlp.down_proj.bias": "model_00018-of-00032.safetensors",
|
88 |
+
"h.16.mlp.down_proj.weight": "model_00018-of-00032.safetensors",
|
89 |
+
"h.16.mlp.gate_proj.weight": "model_00018-of-00032.safetensors",
|
90 |
+
"h.16.mlp.up_proj.weight": "model_00018-of-00032.safetensors",
|
91 |
+
"h.16.post_attention_layernorm.weight": "model_00018-of-00032.safetensors",
|
92 |
+
"h.16.self_attention.dense.bias": "model_00018-of-00032.safetensors",
|
93 |
+
"h.16.self_attention.dense.weight": "model_00018-of-00032.safetensors",
|
94 |
+
"h.16.self_attention.key_value.weight": "model_00018-of-00032.safetensors",
|
95 |
+
"h.16.self_attention.query.weight": "model_00018-of-00032.safetensors",
|
96 |
+
"h.17.input_layernorm.weight": "model_00019-of-00032.safetensors",
|
97 |
+
"h.17.mlp.down_proj.bias": "model_00019-of-00032.safetensors",
|
98 |
+
"h.17.mlp.down_proj.weight": "model_00019-of-00032.safetensors",
|
99 |
+
"h.17.mlp.gate_proj.weight": "model_00019-of-00032.safetensors",
|
100 |
+
"h.17.mlp.up_proj.weight": "model_00019-of-00032.safetensors",
|
101 |
+
"h.17.post_attention_layernorm.weight": "model_00019-of-00032.safetensors",
|
102 |
+
"h.17.self_attention.dense.bias": "model_00019-of-00032.safetensors",
|
103 |
+
"h.17.self_attention.dense.weight": "model_00019-of-00032.safetensors",
|
104 |
+
"h.17.self_attention.key_value.weight": "model_00019-of-00032.safetensors",
|
105 |
+
"h.17.self_attention.query.weight": "model_00019-of-00032.safetensors",
|
106 |
+
"h.18.input_layernorm.weight": "model_00020-of-00032.safetensors",
|
107 |
+
"h.18.mlp.down_proj.bias": "model_00020-of-00032.safetensors",
|
108 |
+
"h.18.mlp.down_proj.weight": "model_00020-of-00032.safetensors",
|
109 |
+
"h.18.mlp.gate_proj.weight": "model_00020-of-00032.safetensors",
|
110 |
+
"h.18.mlp.up_proj.weight": "model_00020-of-00032.safetensors",
|
111 |
+
"h.18.post_attention_layernorm.weight": "model_00020-of-00032.safetensors",
|
112 |
+
"h.18.self_attention.dense.bias": "model_00020-of-00032.safetensors",
|
113 |
+
"h.18.self_attention.dense.weight": "model_00020-of-00032.safetensors",
|
114 |
+
"h.18.self_attention.key_value.weight": "model_00020-of-00032.safetensors",
|
115 |
+
"h.18.self_attention.query.weight": "model_00020-of-00032.safetensors",
|
116 |
+
"h.19.input_layernorm.weight": "model_00021-of-00032.safetensors",
|
117 |
+
"h.19.mlp.down_proj.bias": "model_00021-of-00032.safetensors",
|
118 |
+
"h.19.mlp.down_proj.weight": "model_00021-of-00032.safetensors",
|
119 |
+
"h.19.mlp.gate_proj.weight": "model_00021-of-00032.safetensors",
|
120 |
+
"h.19.mlp.up_proj.weight": "model_00021-of-00032.safetensors",
|
121 |
+
"h.19.post_attention_layernorm.weight": "model_00021-of-00032.safetensors",
|
122 |
+
"h.19.self_attention.dense.bias": "model_00021-of-00032.safetensors",
|
123 |
+
"h.19.self_attention.dense.weight": "model_00021-of-00032.safetensors",
|
124 |
+
"h.19.self_attention.key_value.weight": "model_00021-of-00032.safetensors",
|
125 |
+
"h.19.self_attention.query.weight": "model_00021-of-00032.safetensors",
|
126 |
+
"h.2.input_layernorm.weight": "model_00004-of-00032.safetensors",
|
127 |
+
"h.2.mlp.down_proj.bias": "model_00004-of-00032.safetensors",
|
128 |
+
"h.2.mlp.down_proj.weight": "model_00004-of-00032.safetensors",
|
129 |
+
"h.2.mlp.gate_proj.weight": "model_00004-of-00032.safetensors",
|
130 |
+
"h.2.mlp.up_proj.weight": "model_00004-of-00032.safetensors",
|
131 |
+
"h.2.post_attention_layernorm.weight": "model_00004-of-00032.safetensors",
|
132 |
+
"h.2.self_attention.dense.bias": "model_00004-of-00032.safetensors",
|
133 |
+
"h.2.self_attention.dense.weight": "model_00004-of-00032.safetensors",
|
134 |
+
"h.2.self_attention.key_value.weight": "model_00004-of-00032.safetensors",
|
135 |
+
"h.2.self_attention.query.weight": "model_00004-of-00032.safetensors",
|
136 |
+
"h.20.input_layernorm.weight": "model_00022-of-00032.safetensors",
|
137 |
+
"h.20.mlp.down_proj.bias": "model_00022-of-00032.safetensors",
|
138 |
+
"h.20.mlp.down_proj.weight": "model_00022-of-00032.safetensors",
|
139 |
+
"h.20.mlp.gate_proj.weight": "model_00022-of-00032.safetensors",
|
140 |
+
"h.20.mlp.up_proj.weight": "model_00022-of-00032.safetensors",
|
141 |
+
"h.20.post_attention_layernorm.weight": "model_00022-of-00032.safetensors",
|
142 |
+
"h.20.self_attention.dense.bias": "model_00022-of-00032.safetensors",
|
143 |
+
"h.20.self_attention.dense.weight": "model_00022-of-00032.safetensors",
|
144 |
+
"h.20.self_attention.key_value.weight": "model_00022-of-00032.safetensors",
|
145 |
+
"h.20.self_attention.query.weight": "model_00022-of-00032.safetensors",
|
146 |
+
"h.21.input_layernorm.weight": "model_00023-of-00032.safetensors",
|
147 |
+
"h.21.mlp.down_proj.bias": "model_00023-of-00032.safetensors",
|
148 |
+
"h.21.mlp.down_proj.weight": "model_00023-of-00032.safetensors",
|
149 |
+
"h.21.mlp.gate_proj.weight": "model_00023-of-00032.safetensors",
|
150 |
+
"h.21.mlp.up_proj.weight": "model_00023-of-00032.safetensors",
|
151 |
+
"h.21.post_attention_layernorm.weight": "model_00023-of-00032.safetensors",
|
152 |
+
"h.21.self_attention.dense.bias": "model_00023-of-00032.safetensors",
|
153 |
+
"h.21.self_attention.dense.weight": "model_00023-of-00032.safetensors",
|
154 |
+
"h.21.self_attention.key_value.weight": "model_00023-of-00032.safetensors",
|
155 |
+
"h.21.self_attention.query.weight": "model_00023-of-00032.safetensors",
|
156 |
+
"h.22.input_layernorm.weight": "model_00024-of-00032.safetensors",
|
157 |
+
"h.22.mlp.down_proj.bias": "model_00024-of-00032.safetensors",
|
158 |
+
"h.22.mlp.down_proj.weight": "model_00024-of-00032.safetensors",
|
159 |
+
"h.22.mlp.gate_proj.weight": "model_00024-of-00032.safetensors",
|
160 |
+
"h.22.mlp.up_proj.weight": "model_00024-of-00032.safetensors",
|
161 |
+
"h.22.post_attention_layernorm.weight": "model_00024-of-00032.safetensors",
|
162 |
+
"h.22.self_attention.dense.bias": "model_00024-of-00032.safetensors",
|
163 |
+
"h.22.self_attention.dense.weight": "model_00024-of-00032.safetensors",
|
164 |
+
"h.22.self_attention.key_value.weight": "model_00024-of-00032.safetensors",
|
165 |
+
"h.22.self_attention.query.weight": "model_00024-of-00032.safetensors",
|
166 |
+
"h.23.input_layernorm.weight": "model_00025-of-00032.safetensors",
|
167 |
+
"h.23.mlp.down_proj.bias": "model_00025-of-00032.safetensors",
|
168 |
+
"h.23.mlp.down_proj.weight": "model_00025-of-00032.safetensors",
|
169 |
+
"h.23.mlp.gate_proj.weight": "model_00025-of-00032.safetensors",
|
170 |
+
"h.23.mlp.up_proj.weight": "model_00025-of-00032.safetensors",
|
171 |
+
"h.23.post_attention_layernorm.weight": "model_00025-of-00032.safetensors",
|
172 |
+
"h.23.self_attention.dense.bias": "model_00025-of-00032.safetensors",
|
173 |
+
"h.23.self_attention.dense.weight": "model_00025-of-00032.safetensors",
|
174 |
+
"h.23.self_attention.key_value.weight": "model_00025-of-00032.safetensors",
|
175 |
+
"h.23.self_attention.query.weight": "model_00025-of-00032.safetensors",
|
176 |
+
"h.24.input_layernorm.weight": "model_00026-of-00032.safetensors",
|
177 |
+
"h.24.mlp.down_proj.bias": "model_00026-of-00032.safetensors",
|
178 |
+
"h.24.mlp.down_proj.weight": "model_00026-of-00032.safetensors",
|
179 |
+
"h.24.mlp.gate_proj.weight": "model_00026-of-00032.safetensors",
|
180 |
+
"h.24.mlp.up_proj.weight": "model_00026-of-00032.safetensors",
|
181 |
+
"h.24.post_attention_layernorm.weight": "model_00026-of-00032.safetensors",
|
182 |
+
"h.24.self_attention.dense.bias": "model_00026-of-00032.safetensors",
|
183 |
+
"h.24.self_attention.dense.weight": "model_00026-of-00032.safetensors",
|
184 |
+
"h.24.self_attention.key_value.weight": "model_00026-of-00032.safetensors",
|
185 |
+
"h.24.self_attention.query.weight": "model_00026-of-00032.safetensors",
|
186 |
+
"h.25.input_layernorm.weight": "model_00027-of-00032.safetensors",
|
187 |
+
"h.25.mlp.down_proj.bias": "model_00027-of-00032.safetensors",
|
188 |
+
"h.25.mlp.down_proj.weight": "model_00027-of-00032.safetensors",
|
189 |
+
"h.25.mlp.gate_proj.weight": "model_00027-of-00032.safetensors",
|
190 |
+
"h.25.mlp.up_proj.weight": "model_00027-of-00032.safetensors",
|
191 |
+
"h.25.post_attention_layernorm.weight": "model_00027-of-00032.safetensors",
|
192 |
+
"h.25.self_attention.dense.bias": "model_00027-of-00032.safetensors",
|
193 |
+
"h.25.self_attention.dense.weight": "model_00027-of-00032.safetensors",
|
194 |
+
"h.25.self_attention.key_value.weight": "model_00027-of-00032.safetensors",
|
195 |
+
"h.25.self_attention.query.weight": "model_00027-of-00032.safetensors",
|
196 |
+
"h.26.input_layernorm.weight": "model_00028-of-00032.safetensors",
|
197 |
+
"h.26.mlp.down_proj.bias": "model_00028-of-00032.safetensors",
|
198 |
+
"h.26.mlp.down_proj.weight": "model_00028-of-00032.safetensors",
|
199 |
+
"h.26.mlp.gate_proj.weight": "model_00028-of-00032.safetensors",
|
200 |
+
"h.26.mlp.up_proj.weight": "model_00028-of-00032.safetensors",
|
201 |
+
"h.26.post_attention_layernorm.weight": "model_00028-of-00032.safetensors",
|
202 |
+
"h.26.self_attention.dense.bias": "model_00028-of-00032.safetensors",
|
203 |
+
"h.26.self_attention.dense.weight": "model_00028-of-00032.safetensors",
|
204 |
+
"h.26.self_attention.key_value.weight": "model_00028-of-00032.safetensors",
|
205 |
+
"h.26.self_attention.query.weight": "model_00028-of-00032.safetensors",
|
206 |
+
"h.27.input_layernorm.weight": "model_00029-of-00032.safetensors",
|
207 |
+
"h.27.mlp.down_proj.bias": "model_00029-of-00032.safetensors",
|
208 |
+
"h.27.mlp.down_proj.weight": "model_00029-of-00032.safetensors",
|
209 |
+
"h.27.mlp.gate_proj.weight": "model_00029-of-00032.safetensors",
|
210 |
+
"h.27.mlp.up_proj.weight": "model_00029-of-00032.safetensors",
|
211 |
+
"h.27.post_attention_layernorm.weight": "model_00029-of-00032.safetensors",
|
212 |
+
"h.27.self_attention.dense.bias": "model_00029-of-00032.safetensors",
|
213 |
+
"h.27.self_attention.dense.weight": "model_00029-of-00032.safetensors",
|
214 |
+
"h.27.self_attention.key_value.weight": "model_00029-of-00032.safetensors",
|
215 |
+
"h.27.self_attention.query.weight": "model_00029-of-00032.safetensors",
|
216 |
+
"h.28.input_layernorm.weight": "model_00030-of-00032.safetensors",
|
217 |
+
"h.28.mlp.down_proj.bias": "model_00030-of-00032.safetensors",
|
218 |
+
"h.28.mlp.down_proj.weight": "model_00030-of-00032.safetensors",
|
219 |
+
"h.28.mlp.gate_proj.weight": "model_00030-of-00032.safetensors",
|
220 |
+
"h.28.mlp.up_proj.weight": "model_00030-of-00032.safetensors",
|
221 |
+
"h.28.post_attention_layernorm.weight": "model_00030-of-00032.safetensors",
|
222 |
+
"h.28.self_attention.dense.bias": "model_00030-of-00032.safetensors",
|
223 |
+
"h.28.self_attention.dense.weight": "model_00030-of-00032.safetensors",
|
224 |
+
"h.28.self_attention.key_value.weight": "model_00030-of-00032.safetensors",
|
225 |
+
"h.28.self_attention.query.weight": "model_00030-of-00032.safetensors",
|
226 |
+
"h.29.input_layernorm.weight": "model_00031-of-00032.safetensors",
|
227 |
+
"h.29.mlp.down_proj.bias": "model_00031-of-00032.safetensors",
|
228 |
+
"h.29.mlp.down_proj.weight": "model_00031-of-00032.safetensors",
|
229 |
+
"h.29.mlp.gate_proj.weight": "model_00031-of-00032.safetensors",
|
230 |
+
"h.29.mlp.up_proj.weight": "model_00031-of-00032.safetensors",
|
231 |
+
"h.29.post_attention_layernorm.weight": "model_00031-of-00032.safetensors",
|
232 |
+
"h.29.self_attention.dense.bias": "model_00031-of-00032.safetensors",
|
233 |
+
"h.29.self_attention.dense.weight": "model_00031-of-00032.safetensors",
|
234 |
+
"h.29.self_attention.key_value.weight": "model_00031-of-00032.safetensors",
|
235 |
+
"h.29.self_attention.query.weight": "model_00031-of-00032.safetensors",
|
236 |
+
"h.3.input_layernorm.weight": "model_00005-of-00032.safetensors",
|
237 |
+
"h.3.mlp.down_proj.bias": "model_00005-of-00032.safetensors",
|
238 |
+
"h.3.mlp.down_proj.weight": "model_00005-of-00032.safetensors",
|
239 |
+
"h.3.mlp.gate_proj.weight": "model_00005-of-00032.safetensors",
|
240 |
+
"h.3.mlp.up_proj.weight": "model_00005-of-00032.safetensors",
|
241 |
+
"h.3.post_attention_layernorm.weight": "model_00005-of-00032.safetensors",
|
242 |
+
"h.3.self_attention.dense.bias": "model_00005-of-00032.safetensors",
|
243 |
+
"h.3.self_attention.dense.weight": "model_00005-of-00032.safetensors",
|
244 |
+
"h.3.self_attention.key_value.weight": "model_00005-of-00032.safetensors",
|
245 |
+
"h.3.self_attention.query.weight": "model_00005-of-00032.safetensors",
|
246 |
+
"h.4.input_layernorm.weight": "model_00006-of-00032.safetensors",
|
247 |
+
"h.4.mlp.down_proj.bias": "model_00006-of-00032.safetensors",
|
248 |
+
"h.4.mlp.down_proj.weight": "model_00006-of-00032.safetensors",
|
249 |
+
"h.4.mlp.gate_proj.weight": "model_00006-of-00032.safetensors",
|
250 |
+
"h.4.mlp.up_proj.weight": "model_00006-of-00032.safetensors",
|
251 |
+
"h.4.post_attention_layernorm.weight": "model_00006-of-00032.safetensors",
|
252 |
+
"h.4.self_attention.dense.bias": "model_00006-of-00032.safetensors",
|
253 |
+
"h.4.self_attention.dense.weight": "model_00006-of-00032.safetensors",
|
254 |
+
"h.4.self_attention.key_value.weight": "model_00006-of-00032.safetensors",
|
255 |
+
"h.4.self_attention.query.weight": "model_00006-of-00032.safetensors",
|
256 |
+
"h.5.input_layernorm.weight": "model_00007-of-00032.safetensors",
|
257 |
+
"h.5.mlp.down_proj.bias": "model_00007-of-00032.safetensors",
|
258 |
+
"h.5.mlp.down_proj.weight": "model_00007-of-00032.safetensors",
|
259 |
+
"h.5.mlp.gate_proj.weight": "model_00007-of-00032.safetensors",
|
260 |
+
"h.5.mlp.up_proj.weight": "model_00007-of-00032.safetensors",
|
261 |
+
"h.5.post_attention_layernorm.weight": "model_00007-of-00032.safetensors",
|
262 |
+
"h.5.self_attention.dense.bias": "model_00007-of-00032.safetensors",
|
263 |
+
"h.5.self_attention.dense.weight": "model_00007-of-00032.safetensors",
|
264 |
+
"h.5.self_attention.key_value.weight": "model_00007-of-00032.safetensors",
|
265 |
+
"h.5.self_attention.query.weight": "model_00007-of-00032.safetensors",
|
266 |
+
"h.6.input_layernorm.weight": "model_00008-of-00032.safetensors",
|
267 |
+
"h.6.mlp.down_proj.bias": "model_00008-of-00032.safetensors",
|
268 |
+
"h.6.mlp.down_proj.weight": "model_00008-of-00032.safetensors",
|
269 |
+
"h.6.mlp.gate_proj.weight": "model_00008-of-00032.safetensors",
|
270 |
+
"h.6.mlp.up_proj.weight": "model_00008-of-00032.safetensors",
|
271 |
+
"h.6.post_attention_layernorm.weight": "model_00008-of-00032.safetensors",
|
272 |
+
"h.6.self_attention.dense.bias": "model_00008-of-00032.safetensors",
|
273 |
+
"h.6.self_attention.dense.weight": "model_00008-of-00032.safetensors",
|
274 |
+
"h.6.self_attention.key_value.weight": "model_00008-of-00032.safetensors",
|
275 |
+
"h.6.self_attention.query.weight": "model_00008-of-00032.safetensors",
|
276 |
+
"h.7.input_layernorm.weight": "model_00009-of-00032.safetensors",
|
277 |
+
"h.7.mlp.down_proj.bias": "model_00009-of-00032.safetensors",
|
278 |
+
"h.7.mlp.down_proj.weight": "model_00009-of-00032.safetensors",
|
279 |
+
"h.7.mlp.gate_proj.weight": "model_00009-of-00032.safetensors",
|
280 |
+
"h.7.mlp.up_proj.weight": "model_00009-of-00032.safetensors",
|
281 |
+
"h.7.post_attention_layernorm.weight": "model_00009-of-00032.safetensors",
|
282 |
+
"h.7.self_attention.dense.bias": "model_00009-of-00032.safetensors",
|
283 |
+
"h.7.self_attention.dense.weight": "model_00009-of-00032.safetensors",
|
284 |
+
"h.7.self_attention.key_value.weight": "model_00009-of-00032.safetensors",
|
285 |
+
"h.7.self_attention.query.weight": "model_00009-of-00032.safetensors",
|
286 |
+
"h.8.input_layernorm.weight": "model_00010-of-00032.safetensors",
|
287 |
+
"h.8.mlp.down_proj.bias": "model_00010-of-00032.safetensors",
|
288 |
+
"h.8.mlp.down_proj.weight": "model_00010-of-00032.safetensors",
|
289 |
+
"h.8.mlp.gate_proj.weight": "model_00010-of-00032.safetensors",
|
290 |
+
"h.8.mlp.up_proj.weight": "model_00010-of-00032.safetensors",
|
291 |
+
"h.8.post_attention_layernorm.weight": "model_00010-of-00032.safetensors",
|
292 |
+
"h.8.self_attention.dense.bias": "model_00010-of-00032.safetensors",
|
293 |
+
"h.8.self_attention.dense.weight": "model_00010-of-00032.safetensors",
|
294 |
+
"h.8.self_attention.key_value.weight": "model_00010-of-00032.safetensors",
|
295 |
+
"h.8.self_attention.query.weight": "model_00010-of-00032.safetensors",
|
296 |
+
"h.9.input_layernorm.weight": "model_00011-of-00032.safetensors",
|
297 |
+
"h.9.mlp.down_proj.bias": "model_00011-of-00032.safetensors",
|
298 |
+
"h.9.mlp.down_proj.weight": "model_00011-of-00032.safetensors",
|
299 |
+
"h.9.mlp.gate_proj.weight": "model_00011-of-00032.safetensors",
|
300 |
+
"h.9.mlp.up_proj.weight": "model_00011-of-00032.safetensors",
|
301 |
+
"h.9.post_attention_layernorm.weight": "model_00011-of-00032.safetensors",
|
302 |
+
"h.9.self_attention.dense.bias": "model_00011-of-00032.safetensors",
|
303 |
+
"h.9.self_attention.dense.weight": "model_00011-of-00032.safetensors",
|
304 |
+
"h.9.self_attention.key_value.weight": "model_00011-of-00032.safetensors",
|
305 |
+
"h.9.self_attention.query.weight": "model_00011-of-00032.safetensors",
|
306 |
+
"ln_f.weight": "model_00032-of-00032.safetensors",
|
307 |
+
"word_embeddings.weight": "model_00001-of-00032.safetensors"
|
308 |
+
}
|
309 |
+
}
|
model_00001-of-00032.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:556b983983462cf0ced16ff9b10adc81f753ef96b998e417cdbc0b0eaf7c0adf
|
3 |
+
size 1312817288
|
model_00002-of-00032.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:5d07841a04997ef05cf31c99c3569808388b83ab4af50f58b58d950e26d5fe2a
|
3 |
+
size 436241424
|
model_00003-of-00032.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:dd0038354562dd55c122b08284138267b72f4545c48564223e38e081a3eb3a6f
|
3 |
+
size 436241424
|
model_00004-of-00032.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:fe807fc83148692b2ec70e918664c02d22ea6f89b365dde16329dc7999beb887
|
3 |
+
size 436241424
|
model_00005-of-00032.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:b41b236a81d35c2aef4f36428bb930cb23985d2ca972e6669aadc994a50dac58
|
3 |
+
size 436241424
|
model_00006-of-00032.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:6a6d02efa14f3803c99ff8431e32db10a324259c813c7bb359c7b4d5cd41a97f
|
3 |
+
size 436241424
|
model_00007-of-00032.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:36606ee9f6741f01bc33f5afe7bea73584ff420606ff4b40933bd9f112f3da6f
|
3 |
+
size 436241424
|
model_00008-of-00032.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:cfc5d37ea020e0ad99589a9d51b35b35ea0fe76ad1a634a36980debcb4d16400
|
3 |
+
size 436241424
|
model_00009-of-00032.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:b756a9f2b0e7a28e16702179e5433813b4bcb48a7ca5f2bd5772a1d85dd34321
|
3 |
+
size 436241424
|
model_00010-of-00032.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:86ed0c1042f3062985a5bf6bfb99c51fe4ff276afec16fd7e848621798920da7
|
3 |
+
size 436241424
|
model_00011-of-00032.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:655e1bbf708b67c33e250ef6671575654d60691271d7c6d54d88981c4da89bd7
|
3 |
+
size 436241424
|
model_00012-of-00032.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:98ba8f41864ffff13df9dba654006dd61bae37a1b2c14fc4296e9a544b06c7ae
|
3 |
+
size 436241432
|
model_00013-of-00032.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:7724008a24175457d81bdd9277db6994eb68592e4feadae779fd8cbfa69bdbd8
|
3 |
+
size 436241432
|
model_00014-of-00032.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:47410f36efe7c42d1c0e91afcf1f3aaecccabdf5ace91b32f2a13a7b71b9f111
|
3 |
+
size 436241432
|
model_00015-of-00032.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:32913deea454acbb070637aed2100e3bff2113acffd53eaf3f7f58862586a1df
|
3 |
+
size 436241432
|
model_00016-of-00032.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:e13141a727092e5e33a125be5fb46b0ba0a336ab7061dbb378ed5683e82b3546
|
3 |
+
size 436241432
|
model_00017-of-00032.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:57d82c20329c3043f9dda91862c741a1169d48e729d6eddfac414506f4e235d9
|
3 |
+
size 436241432
|
model_00018-of-00032.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:bf5ab7bc4ba0d031db392687e0a69e18a19d93e85245cd09ee24d4d6688a8b59
|
3 |
+
size 436241432
|
model_00019-of-00032.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:614365ced74c8e8fc627336a3c88ddfd4f0a1e4de3d4bdec1ea2028c9f4cdda7
|
3 |
+
size 436241432
|
model_00020-of-00032.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:183a12959f09bbf775c1dc0b8dd8b73d06540e5c3f33af424adaa00867273a71
|
3 |
+
size 436241432
|
model_00021-of-00032.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:65883d7503c3c25f8a42efba949a87c26120e80232228f96971511a77bed7ac6
|
3 |
+
size 436241432
|
model_00022-of-00032.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:74f31513119aed1bd6164e65f23279dc054bb31a9e934aab0792d54ef24786dc
|
3 |
+
size 436241432
|
model_00023-of-00032.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:2d15bb43c898d8a1826ac55bb51660b2ec219953f399d01fb75fd2867e9ef261
|
3 |
+
size 436241432
|
model_00024-of-00032.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:ae95648ac37ae15198e9c938b090085c4b555fbd8d6dbcb24fa2fa380047e153
|
3 |
+
size 436241432
|
model_00025-of-00032.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:5f6b4590d594fc3fe5be2c8a791d4153eaadb635009617e46deb8249305d2220
|
3 |
+
size 436241432
|
model_00026-of-00032.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:a8e6ac79bf617752319e67f22384f717c85d88a480f45aee81cb487ac95e387d
|
3 |
+
size 436241432
|
model_00027-of-00032.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:50a89bbd6758d20d76da31ccdc55afd007b377316c38e4070972a64e8b39c713
|
3 |
+
size 436241432
|
model_00028-of-00032.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:2ee5676f9aba0f1ab94511df1b1e0872843f9fd91a9126cf49d7e6d3895ecd47
|
3 |
+
size 436241432
|
model_00029-of-00032.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:315d3c41be7a8c5a2d3efd8b7af044d03c3fade6ceb24f986ed5bd2d96c2b824
|
3 |
+
size 436241432
|
model_00030-of-00032.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:4fafc37807f22d35cf9e8e0949daa4b98fca7dd3fc022ba5af6527aa97decf3f
|
3 |
+
size 436241432
|
model_00031-of-00032.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:e296f5c162857c10d763a1a81ed998c4a743c5c079fbe924b115bfc27fcd3c10
|
3 |
+
size 436241432
|
model_00032-of-00032.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:dcd05006bfcdb8199627fe470dedb7cb94705f1cf2f9ffd6d45048e5b5508186
|
3 |
+
size 8304
|