diff --git a/model.safetensors.index.json b/model.safetensors.index.json new file mode 100644 index 0000000000000000000000000000000000000000..6c15db7f6fed4cfcc1473fe0663eb4e5867e6675 --- /dev/null +++ b/model.safetensors.index.json @@ -0,0 +1,550 @@ +{ + "metadata": { + "total_size": 40182671360 + }, + "weight_map": { + "lm_head.weight": "model_00063-of-00063.safetensors", + "model.embed_tokens.weight": "model_00001-of-00063.safetensors", + "model.layers.0.input_layernorm.weight": "model_00002-of-00063.safetensors", + "model.layers.0.mlp.down_proj.weight": "model_00002-of-00063.safetensors", + "model.layers.0.mlp.gate_proj.weight": "model_00002-of-00063.safetensors", + "model.layers.0.mlp.up_proj.weight": "model_00002-of-00063.safetensors", + "model.layers.0.post_attention_layernorm.weight": "model_00002-of-00063.safetensors", + "model.layers.0.self_attn.k_proj.weight": "model_00002-of-00063.safetensors", + "model.layers.0.self_attn.o_proj.weight": "model_00002-of-00063.safetensors", + "model.layers.0.self_attn.q_proj.weight": "model_00002-of-00063.safetensors", + "model.layers.0.self_attn.v_proj.weight": "model_00002-of-00063.safetensors", + "model.layers.1.input_layernorm.weight": "model_00003-of-00063.safetensors", + "model.layers.1.mlp.down_proj.weight": "model_00003-of-00063.safetensors", + "model.layers.1.mlp.gate_proj.weight": "model_00003-of-00063.safetensors", + "model.layers.1.mlp.up_proj.weight": "model_00003-of-00063.safetensors", + "model.layers.1.post_attention_layernorm.weight": "model_00003-of-00063.safetensors", + "model.layers.1.self_attn.k_proj.weight": "model_00003-of-00063.safetensors", + "model.layers.1.self_attn.o_proj.weight": "model_00003-of-00063.safetensors", + "model.layers.1.self_attn.q_proj.weight": "model_00003-of-00063.safetensors", + "model.layers.1.self_attn.v_proj.weight": "model_00003-of-00063.safetensors", + "model.layers.10.input_layernorm.weight": "model_00012-of-00063.safetensors", + "model.layers.10.mlp.down_proj.weight": "model_00012-of-00063.safetensors", + "model.layers.10.mlp.gate_proj.weight": "model_00012-of-00063.safetensors", + "model.layers.10.mlp.up_proj.weight": "model_00012-of-00063.safetensors", + "model.layers.10.post_attention_layernorm.weight": "model_00012-of-00063.safetensors", + "model.layers.10.self_attn.k_proj.weight": "model_00012-of-00063.safetensors", + "model.layers.10.self_attn.o_proj.weight": "model_00012-of-00063.safetensors", + "model.layers.10.self_attn.q_proj.weight": "model_00012-of-00063.safetensors", + "model.layers.10.self_attn.v_proj.weight": "model_00012-of-00063.safetensors", + "model.layers.11.input_layernorm.weight": "model_00013-of-00063.safetensors", + "model.layers.11.mlp.down_proj.weight": "model_00013-of-00063.safetensors", + "model.layers.11.mlp.gate_proj.weight": "model_00013-of-00063.safetensors", + "model.layers.11.mlp.up_proj.weight": "model_00013-of-00063.safetensors", + "model.layers.11.post_attention_layernorm.weight": "model_00013-of-00063.safetensors", + "model.layers.11.self_attn.k_proj.weight": "model_00013-of-00063.safetensors", + "model.layers.11.self_attn.o_proj.weight": "model_00013-of-00063.safetensors", + "model.layers.11.self_attn.q_proj.weight": "model_00013-of-00063.safetensors", + "model.layers.11.self_attn.v_proj.weight": "model_00013-of-00063.safetensors", + "model.layers.12.input_layernorm.weight": "model_00014-of-00063.safetensors", + "model.layers.12.mlp.down_proj.weight": "model_00014-of-00063.safetensors", + "model.layers.12.mlp.gate_proj.weight": "model_00014-of-00063.safetensors", + "model.layers.12.mlp.up_proj.weight": "model_00014-of-00063.safetensors", + "model.layers.12.post_attention_layernorm.weight": "model_00014-of-00063.safetensors", + "model.layers.12.self_attn.k_proj.weight": "model_00014-of-00063.safetensors", + "model.layers.12.self_attn.o_proj.weight": "model_00014-of-00063.safetensors", + "model.layers.12.self_attn.q_proj.weight": "model_00014-of-00063.safetensors", + "model.layers.12.self_attn.v_proj.weight": "model_00014-of-00063.safetensors", + "model.layers.13.input_layernorm.weight": "model_00015-of-00063.safetensors", + "model.layers.13.mlp.down_proj.weight": "model_00015-of-00063.safetensors", + "model.layers.13.mlp.gate_proj.weight": "model_00015-of-00063.safetensors", + "model.layers.13.mlp.up_proj.weight": "model_00015-of-00063.safetensors", + "model.layers.13.post_attention_layernorm.weight": "model_00015-of-00063.safetensors", + "model.layers.13.self_attn.k_proj.weight": "model_00015-of-00063.safetensors", + "model.layers.13.self_attn.o_proj.weight": "model_00015-of-00063.safetensors", + "model.layers.13.self_attn.q_proj.weight": "model_00015-of-00063.safetensors", + "model.layers.13.self_attn.v_proj.weight": "model_00015-of-00063.safetensors", + "model.layers.14.input_layernorm.weight": "model_00016-of-00063.safetensors", + "model.layers.14.mlp.down_proj.weight": "model_00016-of-00063.safetensors", + "model.layers.14.mlp.gate_proj.weight": "model_00016-of-00063.safetensors", + "model.layers.14.mlp.up_proj.weight": "model_00016-of-00063.safetensors", + "model.layers.14.post_attention_layernorm.weight": "model_00016-of-00063.safetensors", + "model.layers.14.self_attn.k_proj.weight": "model_00016-of-00063.safetensors", + "model.layers.14.self_attn.o_proj.weight": "model_00016-of-00063.safetensors", + "model.layers.14.self_attn.q_proj.weight": "model_00016-of-00063.safetensors", + "model.layers.14.self_attn.v_proj.weight": "model_00016-of-00063.safetensors", + "model.layers.15.input_layernorm.weight": "model_00017-of-00063.safetensors", + "model.layers.15.mlp.down_proj.weight": "model_00017-of-00063.safetensors", + "model.layers.15.mlp.gate_proj.weight": "model_00017-of-00063.safetensors", + "model.layers.15.mlp.up_proj.weight": "model_00017-of-00063.safetensors", + "model.layers.15.post_attention_layernorm.weight": "model_00017-of-00063.safetensors", + "model.layers.15.self_attn.k_proj.weight": "model_00017-of-00063.safetensors", + "model.layers.15.self_attn.o_proj.weight": "model_00017-of-00063.safetensors", + "model.layers.15.self_attn.q_proj.weight": "model_00017-of-00063.safetensors", + "model.layers.15.self_attn.v_proj.weight": "model_00017-of-00063.safetensors", + "model.layers.16.input_layernorm.weight": "model_00018-of-00063.safetensors", + "model.layers.16.mlp.down_proj.weight": "model_00018-of-00063.safetensors", + "model.layers.16.mlp.gate_proj.weight": "model_00018-of-00063.safetensors", + "model.layers.16.mlp.up_proj.weight": "model_00018-of-00063.safetensors", + "model.layers.16.post_attention_layernorm.weight": "model_00018-of-00063.safetensors", + "model.layers.16.self_attn.k_proj.weight": "model_00018-of-00063.safetensors", + "model.layers.16.self_attn.o_proj.weight": "model_00018-of-00063.safetensors", + "model.layers.16.self_attn.q_proj.weight": "model_00018-of-00063.safetensors", + "model.layers.16.self_attn.v_proj.weight": "model_00018-of-00063.safetensors", + "model.layers.17.input_layernorm.weight": "model_00019-of-00063.safetensors", + "model.layers.17.mlp.down_proj.weight": "model_00019-of-00063.safetensors", + "model.layers.17.mlp.gate_proj.weight": "model_00019-of-00063.safetensors", + "model.layers.17.mlp.up_proj.weight": "model_00019-of-00063.safetensors", + "model.layers.17.post_attention_layernorm.weight": "model_00019-of-00063.safetensors", + "model.layers.17.self_attn.k_proj.weight": "model_00019-of-00063.safetensors", + "model.layers.17.self_attn.o_proj.weight": "model_00019-of-00063.safetensors", + "model.layers.17.self_attn.q_proj.weight": "model_00019-of-00063.safetensors", + "model.layers.17.self_attn.v_proj.weight": "model_00019-of-00063.safetensors", + "model.layers.18.input_layernorm.weight": "model_00020-of-00063.safetensors", + "model.layers.18.mlp.down_proj.weight": "model_00020-of-00063.safetensors", + "model.layers.18.mlp.gate_proj.weight": "model_00020-of-00063.safetensors", + "model.layers.18.mlp.up_proj.weight": "model_00020-of-00063.safetensors", + "model.layers.18.post_attention_layernorm.weight": "model_00020-of-00063.safetensors", + "model.layers.18.self_attn.k_proj.weight": "model_00020-of-00063.safetensors", + "model.layers.18.self_attn.o_proj.weight": "model_00020-of-00063.safetensors", + "model.layers.18.self_attn.q_proj.weight": "model_00020-of-00063.safetensors", + "model.layers.18.self_attn.v_proj.weight": "model_00020-of-00063.safetensors", + "model.layers.19.input_layernorm.weight": "model_00021-of-00063.safetensors", + "model.layers.19.mlp.down_proj.weight": "model_00021-of-00063.safetensors", + "model.layers.19.mlp.gate_proj.weight": "model_00021-of-00063.safetensors", + "model.layers.19.mlp.up_proj.weight": "model_00021-of-00063.safetensors", + "model.layers.19.post_attention_layernorm.weight": "model_00021-of-00063.safetensors", + "model.layers.19.self_attn.k_proj.weight": "model_00021-of-00063.safetensors", + "model.layers.19.self_attn.o_proj.weight": "model_00021-of-00063.safetensors", + "model.layers.19.self_attn.q_proj.weight": "model_00021-of-00063.safetensors", + "model.layers.19.self_attn.v_proj.weight": "model_00021-of-00063.safetensors", + "model.layers.2.input_layernorm.weight": "model_00004-of-00063.safetensors", + "model.layers.2.mlp.down_proj.weight": "model_00004-of-00063.safetensors", + "model.layers.2.mlp.gate_proj.weight": "model_00004-of-00063.safetensors", + "model.layers.2.mlp.up_proj.weight": "model_00004-of-00063.safetensors", + "model.layers.2.post_attention_layernorm.weight": "model_00004-of-00063.safetensors", + "model.layers.2.self_attn.k_proj.weight": "model_00004-of-00063.safetensors", + "model.layers.2.self_attn.o_proj.weight": "model_00004-of-00063.safetensors", + "model.layers.2.self_attn.q_proj.weight": "model_00004-of-00063.safetensors", + "model.layers.2.self_attn.v_proj.weight": "model_00004-of-00063.safetensors", + "model.layers.20.input_layernorm.weight": "model_00022-of-00063.safetensors", + "model.layers.20.mlp.down_proj.weight": "model_00022-of-00063.safetensors", + "model.layers.20.mlp.gate_proj.weight": "model_00022-of-00063.safetensors", + "model.layers.20.mlp.up_proj.weight": "model_00022-of-00063.safetensors", + "model.layers.20.post_attention_layernorm.weight": "model_00022-of-00063.safetensors", + "model.layers.20.self_attn.k_proj.weight": "model_00022-of-00063.safetensors", + "model.layers.20.self_attn.o_proj.weight": "model_00022-of-00063.safetensors", + "model.layers.20.self_attn.q_proj.weight": "model_00022-of-00063.safetensors", + "model.layers.20.self_attn.v_proj.weight": "model_00022-of-00063.safetensors", + "model.layers.21.input_layernorm.weight": "model_00023-of-00063.safetensors", + "model.layers.21.mlp.down_proj.weight": "model_00023-of-00063.safetensors", + "model.layers.21.mlp.gate_proj.weight": "model_00023-of-00063.safetensors", + "model.layers.21.mlp.up_proj.weight": "model_00023-of-00063.safetensors", + "model.layers.21.post_attention_layernorm.weight": "model_00023-of-00063.safetensors", + "model.layers.21.self_attn.k_proj.weight": "model_00023-of-00063.safetensors", + "model.layers.21.self_attn.o_proj.weight": "model_00023-of-00063.safetensors", + "model.layers.21.self_attn.q_proj.weight": "model_00023-of-00063.safetensors", + "model.layers.21.self_attn.v_proj.weight": "model_00023-of-00063.safetensors", + "model.layers.22.input_layernorm.weight": "model_00024-of-00063.safetensors", + "model.layers.22.mlp.down_proj.weight": "model_00024-of-00063.safetensors", + "model.layers.22.mlp.gate_proj.weight": "model_00024-of-00063.safetensors", + "model.layers.22.mlp.up_proj.weight": "model_00024-of-00063.safetensors", + "model.layers.22.post_attention_layernorm.weight": "model_00024-of-00063.safetensors", + "model.layers.22.self_attn.k_proj.weight": "model_00024-of-00063.safetensors", + "model.layers.22.self_attn.o_proj.weight": "model_00024-of-00063.safetensors", + "model.layers.22.self_attn.q_proj.weight": "model_00024-of-00063.safetensors", + "model.layers.22.self_attn.v_proj.weight": "model_00024-of-00063.safetensors", + "model.layers.23.input_layernorm.weight": "model_00025-of-00063.safetensors", + "model.layers.23.mlp.down_proj.weight": "model_00025-of-00063.safetensors", + "model.layers.23.mlp.gate_proj.weight": "model_00025-of-00063.safetensors", + "model.layers.23.mlp.up_proj.weight": "model_00025-of-00063.safetensors", + "model.layers.23.post_attention_layernorm.weight": "model_00025-of-00063.safetensors", + "model.layers.23.self_attn.k_proj.weight": "model_00025-of-00063.safetensors", + "model.layers.23.self_attn.o_proj.weight": "model_00025-of-00063.safetensors", + "model.layers.23.self_attn.q_proj.weight": "model_00025-of-00063.safetensors", + "model.layers.23.self_attn.v_proj.weight": "model_00025-of-00063.safetensors", + "model.layers.24.input_layernorm.weight": "model_00026-of-00063.safetensors", + "model.layers.24.mlp.down_proj.weight": "model_00026-of-00063.safetensors", + "model.layers.24.mlp.gate_proj.weight": "model_00026-of-00063.safetensors", + "model.layers.24.mlp.up_proj.weight": "model_00026-of-00063.safetensors", + "model.layers.24.post_attention_layernorm.weight": "model_00026-of-00063.safetensors", + "model.layers.24.self_attn.k_proj.weight": "model_00026-of-00063.safetensors", + "model.layers.24.self_attn.o_proj.weight": "model_00026-of-00063.safetensors", + "model.layers.24.self_attn.q_proj.weight": "model_00026-of-00063.safetensors", + "model.layers.24.self_attn.v_proj.weight": "model_00026-of-00063.safetensors", + "model.layers.25.input_layernorm.weight": "model_00027-of-00063.safetensors", + "model.layers.25.mlp.down_proj.weight": "model_00027-of-00063.safetensors", + "model.layers.25.mlp.gate_proj.weight": "model_00027-of-00063.safetensors", + "model.layers.25.mlp.up_proj.weight": "model_00027-of-00063.safetensors", + "model.layers.25.post_attention_layernorm.weight": "model_00027-of-00063.safetensors", + "model.layers.25.self_attn.k_proj.weight": "model_00027-of-00063.safetensors", + "model.layers.25.self_attn.o_proj.weight": "model_00027-of-00063.safetensors", + "model.layers.25.self_attn.q_proj.weight": "model_00027-of-00063.safetensors", + "model.layers.25.self_attn.v_proj.weight": "model_00027-of-00063.safetensors", + "model.layers.26.input_layernorm.weight": "model_00028-of-00063.safetensors", + "model.layers.26.mlp.down_proj.weight": "model_00028-of-00063.safetensors", + "model.layers.26.mlp.gate_proj.weight": "model_00028-of-00063.safetensors", + "model.layers.26.mlp.up_proj.weight": "model_00028-of-00063.safetensors", + "model.layers.26.post_attention_layernorm.weight": "model_00028-of-00063.safetensors", + "model.layers.26.self_attn.k_proj.weight": "model_00028-of-00063.safetensors", + "model.layers.26.self_attn.o_proj.weight": "model_00028-of-00063.safetensors", + "model.layers.26.self_attn.q_proj.weight": "model_00028-of-00063.safetensors", + "model.layers.26.self_attn.v_proj.weight": "model_00028-of-00063.safetensors", + "model.layers.27.input_layernorm.weight": "model_00029-of-00063.safetensors", + "model.layers.27.mlp.down_proj.weight": "model_00029-of-00063.safetensors", + "model.layers.27.mlp.gate_proj.weight": "model_00029-of-00063.safetensors", + "model.layers.27.mlp.up_proj.weight": "model_00029-of-00063.safetensors", + "model.layers.27.post_attention_layernorm.weight": "model_00029-of-00063.safetensors", + "model.layers.27.self_attn.k_proj.weight": "model_00029-of-00063.safetensors", + "model.layers.27.self_attn.o_proj.weight": "model_00029-of-00063.safetensors", + "model.layers.27.self_attn.q_proj.weight": "model_00029-of-00063.safetensors", + "model.layers.27.self_attn.v_proj.weight": "model_00029-of-00063.safetensors", + "model.layers.28.input_layernorm.weight": "model_00030-of-00063.safetensors", + "model.layers.28.mlp.down_proj.weight": "model_00030-of-00063.safetensors", + "model.layers.28.mlp.gate_proj.weight": "model_00030-of-00063.safetensors", + "model.layers.28.mlp.up_proj.weight": "model_00030-of-00063.safetensors", + "model.layers.28.post_attention_layernorm.weight": "model_00030-of-00063.safetensors", + "model.layers.28.self_attn.k_proj.weight": "model_00030-of-00063.safetensors", + "model.layers.28.self_attn.o_proj.weight": "model_00030-of-00063.safetensors", + "model.layers.28.self_attn.q_proj.weight": "model_00030-of-00063.safetensors", + "model.layers.28.self_attn.v_proj.weight": "model_00030-of-00063.safetensors", + "model.layers.29.input_layernorm.weight": "model_00031-of-00063.safetensors", + "model.layers.29.mlp.down_proj.weight": "model_00031-of-00063.safetensors", + "model.layers.29.mlp.gate_proj.weight": "model_00031-of-00063.safetensors", + "model.layers.29.mlp.up_proj.weight": "model_00031-of-00063.safetensors", + "model.layers.29.post_attention_layernorm.weight": "model_00031-of-00063.safetensors", + "model.layers.29.self_attn.k_proj.weight": "model_00031-of-00063.safetensors", + "model.layers.29.self_attn.o_proj.weight": "model_00031-of-00063.safetensors", + "model.layers.29.self_attn.q_proj.weight": "model_00031-of-00063.safetensors", + "model.layers.29.self_attn.v_proj.weight": "model_00031-of-00063.safetensors", + "model.layers.3.input_layernorm.weight": "model_00005-of-00063.safetensors", + "model.layers.3.mlp.down_proj.weight": "model_00005-of-00063.safetensors", + "model.layers.3.mlp.gate_proj.weight": "model_00005-of-00063.safetensors", + "model.layers.3.mlp.up_proj.weight": "model_00005-of-00063.safetensors", + "model.layers.3.post_attention_layernorm.weight": "model_00005-of-00063.safetensors", + "model.layers.3.self_attn.k_proj.weight": "model_00005-of-00063.safetensors", + "model.layers.3.self_attn.o_proj.weight": "model_00005-of-00063.safetensors", + "model.layers.3.self_attn.q_proj.weight": "model_00005-of-00063.safetensors", + "model.layers.3.self_attn.v_proj.weight": "model_00005-of-00063.safetensors", + "model.layers.30.input_layernorm.weight": "model_00032-of-00063.safetensors", + "model.layers.30.mlp.down_proj.weight": "model_00032-of-00063.safetensors", + "model.layers.30.mlp.gate_proj.weight": "model_00032-of-00063.safetensors", + "model.layers.30.mlp.up_proj.weight": "model_00032-of-00063.safetensors", + "model.layers.30.post_attention_layernorm.weight": "model_00032-of-00063.safetensors", + "model.layers.30.self_attn.k_proj.weight": "model_00032-of-00063.safetensors", + "model.layers.30.self_attn.o_proj.weight": "model_00032-of-00063.safetensors", + "model.layers.30.self_attn.q_proj.weight": "model_00032-of-00063.safetensors", + "model.layers.30.self_attn.v_proj.weight": "model_00032-of-00063.safetensors", + "model.layers.31.input_layernorm.weight": "model_00033-of-00063.safetensors", + "model.layers.31.mlp.down_proj.weight": "model_00033-of-00063.safetensors", + "model.layers.31.mlp.gate_proj.weight": "model_00033-of-00063.safetensors", + "model.layers.31.mlp.up_proj.weight": "model_00033-of-00063.safetensors", + "model.layers.31.post_attention_layernorm.weight": "model_00033-of-00063.safetensors", + "model.layers.31.self_attn.k_proj.weight": "model_00033-of-00063.safetensors", + "model.layers.31.self_attn.o_proj.weight": "model_00033-of-00063.safetensors", + "model.layers.31.self_attn.q_proj.weight": "model_00033-of-00063.safetensors", + "model.layers.31.self_attn.v_proj.weight": "model_00033-of-00063.safetensors", + "model.layers.32.input_layernorm.weight": "model_00034-of-00063.safetensors", + "model.layers.32.mlp.down_proj.weight": "model_00034-of-00063.safetensors", + "model.layers.32.mlp.gate_proj.weight": "model_00034-of-00063.safetensors", + "model.layers.32.mlp.up_proj.weight": "model_00034-of-00063.safetensors", + "model.layers.32.post_attention_layernorm.weight": "model_00034-of-00063.safetensors", + "model.layers.32.self_attn.k_proj.weight": "model_00034-of-00063.safetensors", + "model.layers.32.self_attn.o_proj.weight": "model_00034-of-00063.safetensors", + "model.layers.32.self_attn.q_proj.weight": "model_00034-of-00063.safetensors", + "model.layers.32.self_attn.v_proj.weight": "model_00034-of-00063.safetensors", + "model.layers.33.input_layernorm.weight": "model_00035-of-00063.safetensors", + "model.layers.33.mlp.down_proj.weight": "model_00035-of-00063.safetensors", + "model.layers.33.mlp.gate_proj.weight": "model_00035-of-00063.safetensors", + "model.layers.33.mlp.up_proj.weight": "model_00035-of-00063.safetensors", + "model.layers.33.post_attention_layernorm.weight": "model_00035-of-00063.safetensors", + "model.layers.33.self_attn.k_proj.weight": "model_00035-of-00063.safetensors", + "model.layers.33.self_attn.o_proj.weight": "model_00035-of-00063.safetensors", + "model.layers.33.self_attn.q_proj.weight": "model_00035-of-00063.safetensors", + "model.layers.33.self_attn.v_proj.weight": "model_00035-of-00063.safetensors", + "model.layers.34.input_layernorm.weight": "model_00036-of-00063.safetensors", + "model.layers.34.mlp.down_proj.weight": "model_00036-of-00063.safetensors", + "model.layers.34.mlp.gate_proj.weight": "model_00036-of-00063.safetensors", + "model.layers.34.mlp.up_proj.weight": "model_00036-of-00063.safetensors", + "model.layers.34.post_attention_layernorm.weight": "model_00036-of-00063.safetensors", + "model.layers.34.self_attn.k_proj.weight": "model_00036-of-00063.safetensors", + "model.layers.34.self_attn.o_proj.weight": "model_00036-of-00063.safetensors", + "model.layers.34.self_attn.q_proj.weight": "model_00036-of-00063.safetensors", + "model.layers.34.self_attn.v_proj.weight": "model_00036-of-00063.safetensors", + "model.layers.35.input_layernorm.weight": "model_00037-of-00063.safetensors", + "model.layers.35.mlp.down_proj.weight": "model_00037-of-00063.safetensors", + "model.layers.35.mlp.gate_proj.weight": "model_00037-of-00063.safetensors", + "model.layers.35.mlp.up_proj.weight": "model_00037-of-00063.safetensors", + "model.layers.35.post_attention_layernorm.weight": "model_00037-of-00063.safetensors", + "model.layers.35.self_attn.k_proj.weight": "model_00037-of-00063.safetensors", + "model.layers.35.self_attn.o_proj.weight": "model_00037-of-00063.safetensors", + "model.layers.35.self_attn.q_proj.weight": "model_00037-of-00063.safetensors", + "model.layers.35.self_attn.v_proj.weight": "model_00037-of-00063.safetensors", + "model.layers.36.input_layernorm.weight": "model_00038-of-00063.safetensors", + "model.layers.36.mlp.down_proj.weight": "model_00038-of-00063.safetensors", + "model.layers.36.mlp.gate_proj.weight": "model_00038-of-00063.safetensors", + "model.layers.36.mlp.up_proj.weight": "model_00038-of-00063.safetensors", + "model.layers.36.post_attention_layernorm.weight": "model_00038-of-00063.safetensors", + "model.layers.36.self_attn.k_proj.weight": "model_00038-of-00063.safetensors", + "model.layers.36.self_attn.o_proj.weight": "model_00038-of-00063.safetensors", + "model.layers.36.self_attn.q_proj.weight": "model_00038-of-00063.safetensors", + "model.layers.36.self_attn.v_proj.weight": "model_00038-of-00063.safetensors", + "model.layers.37.input_layernorm.weight": "model_00039-of-00063.safetensors", + "model.layers.37.mlp.down_proj.weight": "model_00039-of-00063.safetensors", + "model.layers.37.mlp.gate_proj.weight": "model_00039-of-00063.safetensors", + "model.layers.37.mlp.up_proj.weight": "model_00039-of-00063.safetensors", + "model.layers.37.post_attention_layernorm.weight": "model_00039-of-00063.safetensors", + "model.layers.37.self_attn.k_proj.weight": "model_00039-of-00063.safetensors", + "model.layers.37.self_attn.o_proj.weight": "model_00039-of-00063.safetensors", + "model.layers.37.self_attn.q_proj.weight": "model_00039-of-00063.safetensors", + "model.layers.37.self_attn.v_proj.weight": "model_00039-of-00063.safetensors", + "model.layers.38.input_layernorm.weight": "model_00040-of-00063.safetensors", + "model.layers.38.mlp.down_proj.weight": "model_00040-of-00063.safetensors", + "model.layers.38.mlp.gate_proj.weight": "model_00040-of-00063.safetensors", + "model.layers.38.mlp.up_proj.weight": "model_00040-of-00063.safetensors", + "model.layers.38.post_attention_layernorm.weight": "model_00040-of-00063.safetensors", + "model.layers.38.self_attn.k_proj.weight": "model_00040-of-00063.safetensors", + "model.layers.38.self_attn.o_proj.weight": "model_00040-of-00063.safetensors", + "model.layers.38.self_attn.q_proj.weight": "model_00040-of-00063.safetensors", + "model.layers.38.self_attn.v_proj.weight": "model_00040-of-00063.safetensors", + "model.layers.39.input_layernorm.weight": "model_00041-of-00063.safetensors", + "model.layers.39.mlp.down_proj.weight": "model_00041-of-00063.safetensors", + "model.layers.39.mlp.gate_proj.weight": "model_00041-of-00063.safetensors", + "model.layers.39.mlp.up_proj.weight": "model_00041-of-00063.safetensors", + "model.layers.39.post_attention_layernorm.weight": "model_00041-of-00063.safetensors", + "model.layers.39.self_attn.k_proj.weight": "model_00041-of-00063.safetensors", + "model.layers.39.self_attn.o_proj.weight": "model_00041-of-00063.safetensors", + "model.layers.39.self_attn.q_proj.weight": "model_00041-of-00063.safetensors", + "model.layers.39.self_attn.v_proj.weight": "model_00041-of-00063.safetensors", + "model.layers.4.input_layernorm.weight": "model_00006-of-00063.safetensors", + "model.layers.4.mlp.down_proj.weight": "model_00006-of-00063.safetensors", + "model.layers.4.mlp.gate_proj.weight": "model_00006-of-00063.safetensors", + "model.layers.4.mlp.up_proj.weight": "model_00006-of-00063.safetensors", + "model.layers.4.post_attention_layernorm.weight": "model_00006-of-00063.safetensors", + "model.layers.4.self_attn.k_proj.weight": "model_00006-of-00063.safetensors", + "model.layers.4.self_attn.o_proj.weight": "model_00006-of-00063.safetensors", + "model.layers.4.self_attn.q_proj.weight": "model_00006-of-00063.safetensors", + "model.layers.4.self_attn.v_proj.weight": "model_00006-of-00063.safetensors", + "model.layers.40.input_layernorm.weight": "model_00042-of-00063.safetensors", + "model.layers.40.mlp.down_proj.weight": "model_00042-of-00063.safetensors", + "model.layers.40.mlp.gate_proj.weight": "model_00042-of-00063.safetensors", + "model.layers.40.mlp.up_proj.weight": "model_00042-of-00063.safetensors", + "model.layers.40.post_attention_layernorm.weight": "model_00042-of-00063.safetensors", + "model.layers.40.self_attn.k_proj.weight": "model_00042-of-00063.safetensors", + "model.layers.40.self_attn.o_proj.weight": "model_00042-of-00063.safetensors", + "model.layers.40.self_attn.q_proj.weight": "model_00042-of-00063.safetensors", + "model.layers.40.self_attn.v_proj.weight": "model_00042-of-00063.safetensors", + "model.layers.41.input_layernorm.weight": "model_00043-of-00063.safetensors", + "model.layers.41.mlp.down_proj.weight": "model_00043-of-00063.safetensors", + "model.layers.41.mlp.gate_proj.weight": "model_00043-of-00063.safetensors", + "model.layers.41.mlp.up_proj.weight": "model_00043-of-00063.safetensors", + "model.layers.41.post_attention_layernorm.weight": "model_00043-of-00063.safetensors", + "model.layers.41.self_attn.k_proj.weight": "model_00043-of-00063.safetensors", + "model.layers.41.self_attn.o_proj.weight": "model_00043-of-00063.safetensors", + "model.layers.41.self_attn.q_proj.weight": "model_00043-of-00063.safetensors", + "model.layers.41.self_attn.v_proj.weight": "model_00043-of-00063.safetensors", + "model.layers.42.input_layernorm.weight": "model_00044-of-00063.safetensors", + "model.layers.42.mlp.down_proj.weight": "model_00044-of-00063.safetensors", + "model.layers.42.mlp.gate_proj.weight": "model_00044-of-00063.safetensors", + "model.layers.42.mlp.up_proj.weight": "model_00044-of-00063.safetensors", + "model.layers.42.post_attention_layernorm.weight": "model_00044-of-00063.safetensors", + "model.layers.42.self_attn.k_proj.weight": "model_00044-of-00063.safetensors", + "model.layers.42.self_attn.o_proj.weight": "model_00044-of-00063.safetensors", + "model.layers.42.self_attn.q_proj.weight": "model_00044-of-00063.safetensors", + "model.layers.42.self_attn.v_proj.weight": "model_00044-of-00063.safetensors", + "model.layers.43.input_layernorm.weight": "model_00045-of-00063.safetensors", + "model.layers.43.mlp.down_proj.weight": "model_00045-of-00063.safetensors", + "model.layers.43.mlp.gate_proj.weight": "model_00045-of-00063.safetensors", + "model.layers.43.mlp.up_proj.weight": "model_00045-of-00063.safetensors", + "model.layers.43.post_attention_layernorm.weight": "model_00045-of-00063.safetensors", + "model.layers.43.self_attn.k_proj.weight": "model_00045-of-00063.safetensors", + "model.layers.43.self_attn.o_proj.weight": "model_00045-of-00063.safetensors", + "model.layers.43.self_attn.q_proj.weight": "model_00045-of-00063.safetensors", + "model.layers.43.self_attn.v_proj.weight": "model_00045-of-00063.safetensors", + "model.layers.44.input_layernorm.weight": "model_00046-of-00063.safetensors", + "model.layers.44.mlp.down_proj.weight": "model_00046-of-00063.safetensors", + "model.layers.44.mlp.gate_proj.weight": "model_00046-of-00063.safetensors", + "model.layers.44.mlp.up_proj.weight": "model_00046-of-00063.safetensors", + "model.layers.44.post_attention_layernorm.weight": "model_00046-of-00063.safetensors", + "model.layers.44.self_attn.k_proj.weight": "model_00046-of-00063.safetensors", + "model.layers.44.self_attn.o_proj.weight": "model_00046-of-00063.safetensors", + "model.layers.44.self_attn.q_proj.weight": "model_00046-of-00063.safetensors", + "model.layers.44.self_attn.v_proj.weight": "model_00046-of-00063.safetensors", + "model.layers.45.input_layernorm.weight": "model_00047-of-00063.safetensors", + "model.layers.45.mlp.down_proj.weight": "model_00047-of-00063.safetensors", + "model.layers.45.mlp.gate_proj.weight": "model_00047-of-00063.safetensors", + "model.layers.45.mlp.up_proj.weight": "model_00047-of-00063.safetensors", + "model.layers.45.post_attention_layernorm.weight": "model_00047-of-00063.safetensors", + "model.layers.45.self_attn.k_proj.weight": "model_00047-of-00063.safetensors", + "model.layers.45.self_attn.o_proj.weight": "model_00047-of-00063.safetensors", + "model.layers.45.self_attn.q_proj.weight": "model_00047-of-00063.safetensors", + "model.layers.45.self_attn.v_proj.weight": "model_00047-of-00063.safetensors", + "model.layers.46.input_layernorm.weight": "model_00048-of-00063.safetensors", + "model.layers.46.mlp.down_proj.weight": "model_00048-of-00063.safetensors", + "model.layers.46.mlp.gate_proj.weight": "model_00048-of-00063.safetensors", + "model.layers.46.mlp.up_proj.weight": "model_00048-of-00063.safetensors", + "model.layers.46.post_attention_layernorm.weight": "model_00048-of-00063.safetensors", + "model.layers.46.self_attn.k_proj.weight": "model_00048-of-00063.safetensors", + "model.layers.46.self_attn.o_proj.weight": "model_00048-of-00063.safetensors", + "model.layers.46.self_attn.q_proj.weight": "model_00048-of-00063.safetensors", + "model.layers.46.self_attn.v_proj.weight": "model_00048-of-00063.safetensors", + "model.layers.47.input_layernorm.weight": "model_00049-of-00063.safetensors", + "model.layers.47.mlp.down_proj.weight": "model_00049-of-00063.safetensors", + "model.layers.47.mlp.gate_proj.weight": "model_00049-of-00063.safetensors", + "model.layers.47.mlp.up_proj.weight": "model_00049-of-00063.safetensors", + "model.layers.47.post_attention_layernorm.weight": "model_00049-of-00063.safetensors", + "model.layers.47.self_attn.k_proj.weight": "model_00049-of-00063.safetensors", + "model.layers.47.self_attn.o_proj.weight": "model_00049-of-00063.safetensors", + "model.layers.47.self_attn.q_proj.weight": "model_00049-of-00063.safetensors", + "model.layers.47.self_attn.v_proj.weight": "model_00049-of-00063.safetensors", + "model.layers.48.input_layernorm.weight": "model_00050-of-00063.safetensors", + "model.layers.48.mlp.down_proj.weight": "model_00050-of-00063.safetensors", + "model.layers.48.mlp.gate_proj.weight": "model_00050-of-00063.safetensors", + "model.layers.48.mlp.up_proj.weight": "model_00050-of-00063.safetensors", + "model.layers.48.post_attention_layernorm.weight": "model_00050-of-00063.safetensors", + "model.layers.48.self_attn.k_proj.weight": "model_00050-of-00063.safetensors", + "model.layers.48.self_attn.o_proj.weight": "model_00050-of-00063.safetensors", + "model.layers.48.self_attn.q_proj.weight": "model_00050-of-00063.safetensors", + "model.layers.48.self_attn.v_proj.weight": "model_00050-of-00063.safetensors", + "model.layers.49.input_layernorm.weight": "model_00051-of-00063.safetensors", + "model.layers.49.mlp.down_proj.weight": "model_00051-of-00063.safetensors", + "model.layers.49.mlp.gate_proj.weight": "model_00051-of-00063.safetensors", + "model.layers.49.mlp.up_proj.weight": "model_00051-of-00063.safetensors", + "model.layers.49.post_attention_layernorm.weight": "model_00051-of-00063.safetensors", + "model.layers.49.self_attn.k_proj.weight": "model_00051-of-00063.safetensors", + "model.layers.49.self_attn.o_proj.weight": "model_00051-of-00063.safetensors", + "model.layers.49.self_attn.q_proj.weight": "model_00051-of-00063.safetensors", + "model.layers.49.self_attn.v_proj.weight": "model_00051-of-00063.safetensors", + "model.layers.5.input_layernorm.weight": "model_00007-of-00063.safetensors", + "model.layers.5.mlp.down_proj.weight": "model_00007-of-00063.safetensors", + "model.layers.5.mlp.gate_proj.weight": "model_00007-of-00063.safetensors", + "model.layers.5.mlp.up_proj.weight": "model_00007-of-00063.safetensors", + "model.layers.5.post_attention_layernorm.weight": "model_00007-of-00063.safetensors", + "model.layers.5.self_attn.k_proj.weight": "model_00007-of-00063.safetensors", + "model.layers.5.self_attn.o_proj.weight": "model_00007-of-00063.safetensors", + "model.layers.5.self_attn.q_proj.weight": "model_00007-of-00063.safetensors", + "model.layers.5.self_attn.v_proj.weight": "model_00007-of-00063.safetensors", + "model.layers.50.input_layernorm.weight": "model_00052-of-00063.safetensors", + "model.layers.50.mlp.down_proj.weight": "model_00052-of-00063.safetensors", + "model.layers.50.mlp.gate_proj.weight": "model_00052-of-00063.safetensors", + "model.layers.50.mlp.up_proj.weight": "model_00052-of-00063.safetensors", + "model.layers.50.post_attention_layernorm.weight": "model_00052-of-00063.safetensors", + "model.layers.50.self_attn.k_proj.weight": "model_00052-of-00063.safetensors", + "model.layers.50.self_attn.o_proj.weight": "model_00052-of-00063.safetensors", + "model.layers.50.self_attn.q_proj.weight": "model_00052-of-00063.safetensors", + "model.layers.50.self_attn.v_proj.weight": "model_00052-of-00063.safetensors", + "model.layers.51.input_layernorm.weight": "model_00053-of-00063.safetensors", + "model.layers.51.mlp.down_proj.weight": "model_00053-of-00063.safetensors", + "model.layers.51.mlp.gate_proj.weight": "model_00053-of-00063.safetensors", + "model.layers.51.mlp.up_proj.weight": "model_00053-of-00063.safetensors", + "model.layers.51.post_attention_layernorm.weight": "model_00053-of-00063.safetensors", + "model.layers.51.self_attn.k_proj.weight": "model_00053-of-00063.safetensors", + "model.layers.51.self_attn.o_proj.weight": "model_00053-of-00063.safetensors", + "model.layers.51.self_attn.q_proj.weight": "model_00053-of-00063.safetensors", + "model.layers.51.self_attn.v_proj.weight": "model_00053-of-00063.safetensors", + "model.layers.52.input_layernorm.weight": "model_00054-of-00063.safetensors", + "model.layers.52.mlp.down_proj.weight": "model_00054-of-00063.safetensors", + "model.layers.52.mlp.gate_proj.weight": "model_00054-of-00063.safetensors", + "model.layers.52.mlp.up_proj.weight": "model_00054-of-00063.safetensors", + "model.layers.52.post_attention_layernorm.weight": "model_00054-of-00063.safetensors", + "model.layers.52.self_attn.k_proj.weight": "model_00054-of-00063.safetensors", + "model.layers.52.self_attn.o_proj.weight": "model_00054-of-00063.safetensors", + "model.layers.52.self_attn.q_proj.weight": "model_00054-of-00063.safetensors", + "model.layers.52.self_attn.v_proj.weight": "model_00054-of-00063.safetensors", + "model.layers.53.input_layernorm.weight": "model_00055-of-00063.safetensors", + "model.layers.53.mlp.down_proj.weight": "model_00055-of-00063.safetensors", + "model.layers.53.mlp.gate_proj.weight": "model_00055-of-00063.safetensors", + "model.layers.53.mlp.up_proj.weight": "model_00055-of-00063.safetensors", + "model.layers.53.post_attention_layernorm.weight": "model_00055-of-00063.safetensors", + "model.layers.53.self_attn.k_proj.weight": "model_00055-of-00063.safetensors", + "model.layers.53.self_attn.o_proj.weight": "model_00055-of-00063.safetensors", + "model.layers.53.self_attn.q_proj.weight": "model_00055-of-00063.safetensors", + "model.layers.53.self_attn.v_proj.weight": "model_00055-of-00063.safetensors", + "model.layers.54.input_layernorm.weight": "model_00056-of-00063.safetensors", + "model.layers.54.mlp.down_proj.weight": "model_00056-of-00063.safetensors", + "model.layers.54.mlp.gate_proj.weight": "model_00056-of-00063.safetensors", + "model.layers.54.mlp.up_proj.weight": "model_00056-of-00063.safetensors", + "model.layers.54.post_attention_layernorm.weight": "model_00056-of-00063.safetensors", + "model.layers.54.self_attn.k_proj.weight": "model_00056-of-00063.safetensors", + "model.layers.54.self_attn.o_proj.weight": "model_00056-of-00063.safetensors", + "model.layers.54.self_attn.q_proj.weight": "model_00056-of-00063.safetensors", + "model.layers.54.self_attn.v_proj.weight": "model_00056-of-00063.safetensors", + "model.layers.55.input_layernorm.weight": "model_00057-of-00063.safetensors", + "model.layers.55.mlp.down_proj.weight": "model_00057-of-00063.safetensors", + "model.layers.55.mlp.gate_proj.weight": "model_00057-of-00063.safetensors", + "model.layers.55.mlp.up_proj.weight": "model_00057-of-00063.safetensors", + "model.layers.55.post_attention_layernorm.weight": "model_00057-of-00063.safetensors", + "model.layers.55.self_attn.k_proj.weight": "model_00057-of-00063.safetensors", + "model.layers.55.self_attn.o_proj.weight": "model_00057-of-00063.safetensors", + "model.layers.55.self_attn.q_proj.weight": "model_00057-of-00063.safetensors", + "model.layers.55.self_attn.v_proj.weight": "model_00057-of-00063.safetensors", + "model.layers.56.input_layernorm.weight": "model_00058-of-00063.safetensors", + "model.layers.56.mlp.down_proj.weight": "model_00058-of-00063.safetensors", + "model.layers.56.mlp.gate_proj.weight": "model_00058-of-00063.safetensors", + "model.layers.56.mlp.up_proj.weight": "model_00058-of-00063.safetensors", + "model.layers.56.post_attention_layernorm.weight": "model_00058-of-00063.safetensors", + "model.layers.56.self_attn.k_proj.weight": "model_00058-of-00063.safetensors", + "model.layers.56.self_attn.o_proj.weight": "model_00058-of-00063.safetensors", + "model.layers.56.self_attn.q_proj.weight": "model_00058-of-00063.safetensors", + "model.layers.56.self_attn.v_proj.weight": "model_00058-of-00063.safetensors", + "model.layers.57.input_layernorm.weight": "model_00059-of-00063.safetensors", + "model.layers.57.mlp.down_proj.weight": "model_00059-of-00063.safetensors", + "model.layers.57.mlp.gate_proj.weight": "model_00059-of-00063.safetensors", + "model.layers.57.mlp.up_proj.weight": "model_00059-of-00063.safetensors", + "model.layers.57.post_attention_layernorm.weight": "model_00059-of-00063.safetensors", + "model.layers.57.self_attn.k_proj.weight": "model_00059-of-00063.safetensors", + "model.layers.57.self_attn.o_proj.weight": "model_00059-of-00063.safetensors", + "model.layers.57.self_attn.q_proj.weight": "model_00059-of-00063.safetensors", + "model.layers.57.self_attn.v_proj.weight": "model_00059-of-00063.safetensors", + "model.layers.58.input_layernorm.weight": "model_00060-of-00063.safetensors", + "model.layers.58.mlp.down_proj.weight": "model_00060-of-00063.safetensors", + "model.layers.58.mlp.gate_proj.weight": "model_00060-of-00063.safetensors", + "model.layers.58.mlp.up_proj.weight": "model_00060-of-00063.safetensors", + "model.layers.58.post_attention_layernorm.weight": "model_00060-of-00063.safetensors", + "model.layers.58.self_attn.k_proj.weight": "model_00060-of-00063.safetensors", + "model.layers.58.self_attn.o_proj.weight": "model_00060-of-00063.safetensors", + "model.layers.58.self_attn.q_proj.weight": "model_00060-of-00063.safetensors", + "model.layers.58.self_attn.v_proj.weight": "model_00060-of-00063.safetensors", + "model.layers.59.input_layernorm.weight": "model_00061-of-00063.safetensors", + "model.layers.59.mlp.down_proj.weight": "model_00061-of-00063.safetensors", + "model.layers.59.mlp.gate_proj.weight": "model_00061-of-00063.safetensors", + "model.layers.59.mlp.up_proj.weight": "model_00061-of-00063.safetensors", + "model.layers.59.post_attention_layernorm.weight": "model_00061-of-00063.safetensors", + "model.layers.59.self_attn.k_proj.weight": "model_00061-of-00063.safetensors", + "model.layers.59.self_attn.o_proj.weight": "model_00061-of-00063.safetensors", + "model.layers.59.self_attn.q_proj.weight": "model_00061-of-00063.safetensors", + "model.layers.59.self_attn.v_proj.weight": "model_00061-of-00063.safetensors", + "model.layers.6.input_layernorm.weight": "model_00008-of-00063.safetensors", + "model.layers.6.mlp.down_proj.weight": "model_00008-of-00063.safetensors", + "model.layers.6.mlp.gate_proj.weight": "model_00008-of-00063.safetensors", + "model.layers.6.mlp.up_proj.weight": "model_00008-of-00063.safetensors", + "model.layers.6.post_attention_layernorm.weight": "model_00008-of-00063.safetensors", + "model.layers.6.self_attn.k_proj.weight": "model_00008-of-00063.safetensors", + "model.layers.6.self_attn.o_proj.weight": "model_00008-of-00063.safetensors", + "model.layers.6.self_attn.q_proj.weight": "model_00008-of-00063.safetensors", + "model.layers.6.self_attn.v_proj.weight": "model_00008-of-00063.safetensors", + "model.layers.7.input_layernorm.weight": "model_00009-of-00063.safetensors", + "model.layers.7.mlp.down_proj.weight": "model_00009-of-00063.safetensors", + "model.layers.7.mlp.gate_proj.weight": "model_00009-of-00063.safetensors", + "model.layers.7.mlp.up_proj.weight": "model_00009-of-00063.safetensors", + "model.layers.7.post_attention_layernorm.weight": "model_00009-of-00063.safetensors", + "model.layers.7.self_attn.k_proj.weight": "model_00009-of-00063.safetensors", + "model.layers.7.self_attn.o_proj.weight": "model_00009-of-00063.safetensors", + "model.layers.7.self_attn.q_proj.weight": "model_00009-of-00063.safetensors", + "model.layers.7.self_attn.v_proj.weight": "model_00009-of-00063.safetensors", + "model.layers.8.input_layernorm.weight": "model_00010-of-00063.safetensors", + "model.layers.8.mlp.down_proj.weight": "model_00010-of-00063.safetensors", + "model.layers.8.mlp.gate_proj.weight": "model_00010-of-00063.safetensors", + "model.layers.8.mlp.up_proj.weight": "model_00010-of-00063.safetensors", + "model.layers.8.post_attention_layernorm.weight": "model_00010-of-00063.safetensors", + "model.layers.8.self_attn.k_proj.weight": "model_00010-of-00063.safetensors", + "model.layers.8.self_attn.o_proj.weight": "model_00010-of-00063.safetensors", + "model.layers.8.self_attn.q_proj.weight": "model_00010-of-00063.safetensors", + "model.layers.8.self_attn.v_proj.weight": "model_00010-of-00063.safetensors", + "model.layers.9.input_layernorm.weight": "model_00011-of-00063.safetensors", + "model.layers.9.mlp.down_proj.weight": "model_00011-of-00063.safetensors", + "model.layers.9.mlp.gate_proj.weight": "model_00011-of-00063.safetensors", + "model.layers.9.mlp.up_proj.weight": "model_00011-of-00063.safetensors", + "model.layers.9.post_attention_layernorm.weight": "model_00011-of-00063.safetensors", + "model.layers.9.self_attn.k_proj.weight": "model_00011-of-00063.safetensors", + "model.layers.9.self_attn.o_proj.weight": "model_00011-of-00063.safetensors", + "model.layers.9.self_attn.q_proj.weight": "model_00011-of-00063.safetensors", + "model.layers.9.self_attn.v_proj.weight": "model_00011-of-00063.safetensors", + "model.norm.weight": "model_00062-of-00063.safetensors" + } +} \ No newline at end of file diff --git a/model_00001-of-00063.safetensors b/model_00001-of-00063.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..e0beb69ebfd0ac2c3736ae234a8e30cb29d53936 --- /dev/null +++ b/model_00001-of-00063.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:e00894ce9ea605f09eb3a42cbd480ddc3a4ac5e92659c021ebf1660a2dcdfcbd +size 1059061904 diff --git a/model_00002-of-00063.safetensors b/model_00002-of-00063.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..1706e56f1e29aa0d5c47297a6030545b701ea8eb --- /dev/null +++ b/model_00002-of-00063.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:94b043c69e526b9243a794879909264b04f09212cf03a2baacc6599d1b549138 +size 634410008 diff --git a/model_00003-of-00063.safetensors b/model_00003-of-00063.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..24396f80adce8357b7c86f499956bdece32b9f95 --- /dev/null +++ b/model_00003-of-00063.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:755e9e3fb2f81c23649a0bc82557c7c28640ef0e59db5ab42ee43a666063d118 +size 634410008 diff --git a/model_00004-of-00063.safetensors b/model_00004-of-00063.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..4900b68d0f17cdd586bab9c023e5557e85ed1889 --- /dev/null +++ b/model_00004-of-00063.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:38a5b39bdb69cfb50b468cc3644f028dc7d5c0749ba64e7f5325b7e3a0df1a9d +size 634410008 diff --git a/model_00005-of-00063.safetensors b/model_00005-of-00063.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..8c9acfe2b1a0356cc9b547657b5daab493a817de --- /dev/null +++ b/model_00005-of-00063.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:a16a94935a082cfbd40f8ac64bd44d315294ab8f7fc308bbff57aef6e2d2d4c0 +size 634410008 diff --git a/model_00006-of-00063.safetensors b/model_00006-of-00063.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..ecd26554eb20fcc17754c7f8041b4fcd5b970cc0 --- /dev/null +++ b/model_00006-of-00063.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:b38f1f6248b9c324e115b91e1db9cf729f87a3cdc2d2c0e57f7ac6098ed7cfaa +size 634410008 diff --git a/model_00007-of-00063.safetensors b/model_00007-of-00063.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..4d6ce4042c795cf5d1c62760661981252fb28521 --- /dev/null +++ b/model_00007-of-00063.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:038b2e914b6f8a87e993807354c7c476a054fbff91cc4c58672034b886da0529 +size 634410008 diff --git a/model_00008-of-00063.safetensors b/model_00008-of-00063.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..b9f0fab91a231aab0efafe0fbc9fc8dae49500c3 --- /dev/null +++ b/model_00008-of-00063.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:c95d73048bd54d9d2aa62b7b4bc185ce04ca63766f61530741c7df20cd443aea +size 634410008 diff --git a/model_00009-of-00063.safetensors b/model_00009-of-00063.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..1e33fdfda74d96af081ba75ab2140f1e8eba28fa --- /dev/null +++ b/model_00009-of-00063.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:a37ab28bacac5eb06a170ba7a872de4dfe990769666c16b4fbf66ef13894a553 +size 634410008 diff --git a/model_00010-of-00063.safetensors b/model_00010-of-00063.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..67e2b8f35cee86088a541c2ac6cc14b847e38360 --- /dev/null +++ b/model_00010-of-00063.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:b6e5a2400d062c298c75981b351042c14f2a467a38f1b6ab94583f81203a6a22 +size 634410008 diff --git a/model_00011-of-00063.safetensors b/model_00011-of-00063.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..8c5c5c1151b331b73aa2e3f89548946ca1f76687 --- /dev/null +++ b/model_00011-of-00063.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:4a07a58e7917e75a060832cab78eb308c2c1baae692355234b4a0f38ac0ffdaa +size 634410008 diff --git a/model_00012-of-00063.safetensors b/model_00012-of-00063.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..1c6264735f7750b3221dec0ed10feabbf540e015 --- /dev/null +++ b/model_00012-of-00063.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:e71cb4c9bdc81392b4419ed9c294cc2ba2671bd82de2a3d95ed52789788331b4 +size 634410024 diff --git a/model_00013-of-00063.safetensors b/model_00013-of-00063.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..d782391d2a479caebbabab42c587bc8f30daf420 --- /dev/null +++ b/model_00013-of-00063.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:09156158b76dbcb6f1751a33e574e8155e51081b594f19c1d39b4632a6a4570f +size 634410024 diff --git a/model_00014-of-00063.safetensors b/model_00014-of-00063.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..33b084c836f7c7606c72a82c95fcb189aabe336f --- /dev/null +++ b/model_00014-of-00063.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:2e7c97a25b00d4007653e63047ad2caf83232a99d03e72df2108d953f1c3e0bd +size 634410024 diff --git a/model_00015-of-00063.safetensors b/model_00015-of-00063.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..2ee5c61ae8aa34c5ef4e1ac1ca659cca1f86456c --- /dev/null +++ b/model_00015-of-00063.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:6e73f7d074309e5643ea1ec7fc545bf2acfc302394a8eeed085952bdcb6b36f9 +size 634410024 diff --git a/model_00016-of-00063.safetensors b/model_00016-of-00063.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..20c335488c856f5e6f7c547f5baccb03fecd2633 --- /dev/null +++ b/model_00016-of-00063.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:d8d3ac81b0ec66645967c92449a6000245574e9c910a1264d3453a3e33875a12 +size 634410024 diff --git a/model_00017-of-00063.safetensors b/model_00017-of-00063.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..9ed4d3cee98d47090d451251fd52b5d39ef1c362 --- /dev/null +++ b/model_00017-of-00063.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:ade6622e14b4e784e4a4d64cd0d2c2a626ccd366fa609ea1eacc2962104940fb +size 634410024 diff --git a/model_00018-of-00063.safetensors b/model_00018-of-00063.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..cab112643b53980261b960c92cbc98afba3df3e2 --- /dev/null +++ b/model_00018-of-00063.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:353dc14df756a211d3b0c330738568f3e550abd9e1428bfae02ad1d7fc11ce83 +size 634410024 diff --git a/model_00019-of-00063.safetensors b/model_00019-of-00063.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..b7cb0f7c57480c68ea6077b4917c307ebc2ae49b --- /dev/null +++ b/model_00019-of-00063.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:fa6948ae4785235cb15c5b59e78f7b7a1c30c27b34eac0544df141bc09ca33ef +size 634410024 diff --git a/model_00020-of-00063.safetensors b/model_00020-of-00063.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..d4b54c35e509187f8f6bd63e51ecadaff24fe3f8 --- /dev/null +++ b/model_00020-of-00063.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:b2382f9be439028910064a5004d96e43fc0396d7601bddee4718ccdbff9749af +size 634410024 diff --git a/model_00021-of-00063.safetensors b/model_00021-of-00063.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..b82a6e0bbeb756236b9796ba54a2df25a6df8108 --- /dev/null +++ b/model_00021-of-00063.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:e8526638cce813a5a8e91ad3d863de8e5caaa14482e71f89484437741d2bf65c +size 634410024 diff --git a/model_00022-of-00063.safetensors b/model_00022-of-00063.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..5fab39d8617a59736fbd8890a51f2a05069bddfb --- /dev/null +++ b/model_00022-of-00063.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:f0b0348887735440127fc113712304d5f0e9b3e1c84fb364b52e28fd8eff777a +size 634410024 diff --git a/model_00023-of-00063.safetensors b/model_00023-of-00063.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..177dafd71afa90ff305faae50aa9f2afb60ab190 --- /dev/null +++ b/model_00023-of-00063.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:bef61a0cadc43f18308f01b3fc85a29ac9e6df11b3fdbd2342c00c4161bbea21 +size 634410024 diff --git a/model_00024-of-00063.safetensors b/model_00024-of-00063.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..4e70e3a83fc8fa2154f841a40777bb6168ea2166 --- /dev/null +++ b/model_00024-of-00063.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:1886b393fe0af19138332a348e6a970fe4fc8c261c3b9f7da0c45f25236ed92e +size 634410024 diff --git a/model_00025-of-00063.safetensors b/model_00025-of-00063.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..cfab3d4fd6e630a1bc0c429b4d63aa7a355afaf5 --- /dev/null +++ b/model_00025-of-00063.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:bc53b9d5d32b43792b2a5bf53daa9efe88d8d15e03fb09ae3a925b895140adb5 +size 634410024 diff --git a/model_00026-of-00063.safetensors b/model_00026-of-00063.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..f110f6abbe4cdf91c06bcc0a047fa84e3777c855 --- /dev/null +++ b/model_00026-of-00063.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:f4ed922d5715f3ff7eb2869503b97fdf875bb01ea920df0e4187f6aacac197d6 +size 634410024 diff --git a/model_00027-of-00063.safetensors b/model_00027-of-00063.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..b3746880ff6625fa7e0190614bffbb61f7ec56ec --- /dev/null +++ b/model_00027-of-00063.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:1bc4a5436073d406c4b5cb5a2120b88e7b6516f7ae29d1749f84f5bf5e11aeaa +size 634410024 diff --git a/model_00028-of-00063.safetensors b/model_00028-of-00063.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..533c81cd649dfeef07a48fb5f3d4e34eb7213d64 --- /dev/null +++ b/model_00028-of-00063.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:1b00db82e38d447d51cf355947eb24ce31934d23c818c2a3eba5e5cab9276a5f +size 634410024 diff --git a/model_00029-of-00063.safetensors b/model_00029-of-00063.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..ff7d207e500ee52a96f565ef35d23b39bcb39be8 --- /dev/null +++ b/model_00029-of-00063.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:91cc93e6980bd82ce6bbd6942a883a04ad4f4b85de1b4e3607eddc683dbde008 +size 634410024 diff --git a/model_00030-of-00063.safetensors b/model_00030-of-00063.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..ae090381cb3733ac8d11d2deea0e54fd24127e50 --- /dev/null +++ b/model_00030-of-00063.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:eeff0a8e3e712dc5613a8593b583099ecf72c867d668f3317f37826092dbc803 +size 634410024 diff --git a/model_00031-of-00063.safetensors b/model_00031-of-00063.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..42bb67f92ca69b21463b7de4a9334260d2241b9e --- /dev/null +++ b/model_00031-of-00063.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:00ec6abfb2147b571b749b6726025e99185efab448d0824e38f0a5442e3588ed +size 634410024 diff --git a/model_00032-of-00063.safetensors b/model_00032-of-00063.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..482aa2d0c37dea5410f259d47fdccf339d81b792 --- /dev/null +++ b/model_00032-of-00063.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:e8b42d4bdf927b61784770fdc35245102b6a00abe4f1d0d3fb83170e1998752d +size 634410024 diff --git a/model_00033-of-00063.safetensors b/model_00033-of-00063.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..589d421c6111d3d2d97f9f370d03ec376a4683f5 --- /dev/null +++ b/model_00033-of-00063.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:e1ab206c0147d2337fc44429d2f0bf214aecdd8652ad3657a9b058f4a57f51ae +size 634410024 diff --git a/model_00034-of-00063.safetensors b/model_00034-of-00063.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..976bbb35a4514f3dead77dee477d57519e075958 --- /dev/null +++ b/model_00034-of-00063.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:b3da02f6ab3d76473244c28ccc5dd05dc45e417a366b08788e55e361b156fe01 +size 634410024 diff --git a/model_00035-of-00063.safetensors b/model_00035-of-00063.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..1e8e04ab8e1042cfa67f7366fc1147e6c71340cf --- /dev/null +++ b/model_00035-of-00063.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:85461f1fb58db38cf9ba6ca6e79cd1b7b83ad12dbc09dd01a149fcf59a597e81 +size 634410024 diff --git a/model_00036-of-00063.safetensors b/model_00036-of-00063.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..1dda2aef550297e41ce1aafc213e942b6413026f --- /dev/null +++ b/model_00036-of-00063.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:75200978a9bb069829e8b976914ef64ec72e7515c113987125ee4ebf2096a84a +size 634410024 diff --git a/model_00037-of-00063.safetensors b/model_00037-of-00063.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..18eccdb46db3d0918e28089f7a8854d8eba63322 --- /dev/null +++ b/model_00037-of-00063.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:e6c61fd65d839a70fa8b2d4912fc6d1c264b05a7d46ff05c073ded9af6efd3b0 +size 634410024 diff --git a/model_00038-of-00063.safetensors b/model_00038-of-00063.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..92fda6a2558afadca6dfc061bdf6ea828164499c --- /dev/null +++ b/model_00038-of-00063.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:c0d332f5337bd9822437d422d863abcaf875b14948958129a4e84ac887586a01 +size 634410024 diff --git a/model_00039-of-00063.safetensors b/model_00039-of-00063.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..c6ce349934970dfae84cca621d65a5df8e8a67cb --- /dev/null +++ b/model_00039-of-00063.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:432610ab5ba7aa616a4645af509f5192c9b8820415a92af2efbfd2f52d948dc1 +size 634410024 diff --git a/model_00040-of-00063.safetensors b/model_00040-of-00063.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..67d4e7f83c22190db7e828e2413dd0713bdeb449 --- /dev/null +++ b/model_00040-of-00063.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:276d73b23a695fbcdc548626ea8b7c1b0c533479455304cdcb152e2df7c89420 +size 634410024 diff --git a/model_00041-of-00063.safetensors b/model_00041-of-00063.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..6b1f5b155f5b4410162aed8a736b0f508bda8563 --- /dev/null +++ b/model_00041-of-00063.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:bd21ab144cd22f47189bab85e331fa816a1e54f71b3ee6d9bd49f93d4b419117 +size 634410024 diff --git a/model_00042-of-00063.safetensors b/model_00042-of-00063.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..efc1ccac7adf6e1446f75940b21c3ac12b164c53 --- /dev/null +++ b/model_00042-of-00063.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:9b0b0ca66b9a42bf02525082fd4f24f2204f20ac416a0816423b6f67407d4220 +size 634410024 diff --git a/model_00043-of-00063.safetensors b/model_00043-of-00063.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..a2c4b0bb7c41facae9a6c94384dd99bf53e55d78 --- /dev/null +++ b/model_00043-of-00063.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:5893c329e32882180a4011b8dc478ac782538bd8eb088158601d9da887ab88eb +size 634410024 diff --git a/model_00044-of-00063.safetensors b/model_00044-of-00063.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..6bd9efb531db0905d67ae7ddc366d41543090657 --- /dev/null +++ b/model_00044-of-00063.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:1e4f3fa249beb6e39b1651dd76eba2e7ccf430298aba9b66666cc7a47cc44e29 +size 634410024 diff --git a/model_00045-of-00063.safetensors b/model_00045-of-00063.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..269bd821d3c430c0298388c7dc00e26f09d49f89 --- /dev/null +++ b/model_00045-of-00063.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:5683a7f4d4434be6a3ffab1883e5913508be50ccb5880b6579f4e4e0e924dd9a +size 634410024 diff --git a/model_00046-of-00063.safetensors b/model_00046-of-00063.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..c9a1889fe8ec438dcecc778b403e14e60fa5b14f --- /dev/null +++ b/model_00046-of-00063.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:146d3d5a01f7fe86baab092153f024a6e236cb3a1e086277a07a4e6af9aae865 +size 634410024 diff --git a/model_00047-of-00063.safetensors b/model_00047-of-00063.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..72b9923e3a30b4c3e4e6646c55f5fedec3d7fef6 --- /dev/null +++ b/model_00047-of-00063.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:dc67cbe870da82b451a49b3e0294486c82b4f742a5ca740d143d18b185f336a0 +size 634410024 diff --git a/model_00048-of-00063.safetensors b/model_00048-of-00063.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..a49fc5ac9ac73983b90a8cc6bc33af4a8d1e8bc4 --- /dev/null +++ b/model_00048-of-00063.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:db0c999501751c9a5073e12733fcf64d93663528ed70f3da5fa89c2c4352966c +size 634410024 diff --git a/model_00049-of-00063.safetensors b/model_00049-of-00063.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..ebd6321a2b43ab90a7a64616ae644e39a2f32c5b --- /dev/null +++ b/model_00049-of-00063.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:cc72abfad5c2172f7f0cd8e95f34815714c693d5572782fd1935f690968893c3 +size 634410024 diff --git a/model_00050-of-00063.safetensors b/model_00050-of-00063.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..347cbe6f08b14b307661ddbb9209b963afe23a2c --- /dev/null +++ b/model_00050-of-00063.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:b31f2c3cdd59287b38f387241933419da3167f1cb788ad2705b7f922ab6514e2 +size 634410024 diff --git a/model_00051-of-00063.safetensors b/model_00051-of-00063.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..89b4a529bc3820b46ba22670329986f998dbc75e --- /dev/null +++ b/model_00051-of-00063.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:a423592d03c4976127ecccbbab8b2858eb4da2ad4c68cc735c5fbaf5543ada27 +size 634410024 diff --git a/model_00052-of-00063.safetensors b/model_00052-of-00063.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..4315748091f17af545212a9fb3491b5c6a77bc6f --- /dev/null +++ b/model_00052-of-00063.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:beeb780fbf77024c6d4fc1d9b5f183e0c90b54e40ffd53caeb29419394a41f89 +size 634410024 diff --git a/model_00053-of-00063.safetensors b/model_00053-of-00063.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..0f06f07bc0d0ed4e3c83119ef8dbcc846a7554d6 --- /dev/null +++ b/model_00053-of-00063.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:ceab5f714a8c832afdd80d9d64c99fcd9b1aabfa2a8e41081db640a49bbc3aea +size 634410024 diff --git a/model_00054-of-00063.safetensors b/model_00054-of-00063.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..2278d7acf05417f8e958df8ff2114e1b8b465f1f --- /dev/null +++ b/model_00054-of-00063.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:1f28a4953dbca70f1ac0d7d183baf1ae49a50bfd9bd92a0c33ee733720ecc531 +size 634410024 diff --git a/model_00055-of-00063.safetensors b/model_00055-of-00063.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..3b516aeec1aeaa4c552c753addb6decd43b99133 --- /dev/null +++ b/model_00055-of-00063.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:5d330c6ab3fe12eb03ef2c5de50c06c2bd152f989ec4ab6cc6b6fd525948f54a +size 634410024 diff --git a/model_00056-of-00063.safetensors b/model_00056-of-00063.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..376869ba9e8ae69492df6eb22aa744520de4f7f2 --- /dev/null +++ b/model_00056-of-00063.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:3c6d20a8baf3864b052b8f9da5af03d311159fdc1f48d5494b460bf909abc9e6 +size 634410024 diff --git a/model_00057-of-00063.safetensors b/model_00057-of-00063.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..160db6f44f1d14d561e51a9f1dca7e1f061e6cf2 --- /dev/null +++ b/model_00057-of-00063.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:8c7eb8a924047b499f06c35c311a669ec3c7a67156864c548101d1fdcb8204eb +size 634410024 diff --git a/model_00058-of-00063.safetensors b/model_00058-of-00063.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..09820d35f91060703419229ca896137837e05093 --- /dev/null +++ b/model_00058-of-00063.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:b340e4686f18be3a51e28952eeb86b9cebc43e9e7e3efbf040861e9d1efe2df4 +size 634410024 diff --git a/model_00059-of-00063.safetensors b/model_00059-of-00063.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..ccd50031aed17123fb00bd2fc91b6909419afa2c --- /dev/null +++ b/model_00059-of-00063.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:797d79d45575335cba4f68f0d5cdd63bff7b4ddd05df84a20f64a850a7ebca34 +size 634410024 diff --git a/model_00060-of-00063.safetensors b/model_00060-of-00063.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..59e87f0e00c496f93018089b40391907c717849f --- /dev/null +++ b/model_00060-of-00063.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:322cb4e2dd1e7bc09e0b85506ba7f8fb254ec442961afa8e0983b467f574369b +size 634410024 diff --git a/model_00061-of-00063.safetensors b/model_00061-of-00063.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..479a8984cb0d770d79eff491db1106fae5aaad52 --- /dev/null +++ b/model_00061-of-00063.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:d3efaeb4a7dbc8ba52b9abd45bb99f313b2c842c74fb402b666c434b1d866141 +size 634410024 diff --git a/model_00062-of-00063.safetensors b/model_00062-of-00063.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..19d3af3aab9e2eb703c54c4217f2326e5b3b5114 --- /dev/null +++ b/model_00062-of-00063.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:a2e490fe631f25a000b418e8cf95510a06a286f17e32859fb62af5af919645c6 +size 10360 diff --git a/model_00063-of-00063.safetensors b/model_00063-of-00063.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..f0bbc1a6de4b242d929a27bc1be870c585b8e88d --- /dev/null +++ b/model_00063-of-00063.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:541820451af996da0948603f440c5894cf4d729afa7d894a05283933d8bb2af0 +size 1059061888