Upload folder using huggingface_hub
Browse filesThis view is limited to 50 files because it contains too many changes.
See raw diff
- .gitattributes +10 -0
- checkpoint/config.json +30 -0
- checkpoint/generation_config.json +6 -0
- checkpoint/pytorch_model.bin/key_to_filename.json +3 -0
- checkpoint/pytorch_model.bin/p0.model.embed_tokens.weight +3 -0
- checkpoint/pytorch_model.bin/p1.model.layers.0.self_attn.q_proj.weight +3 -0
- checkpoint/pytorch_model.bin/p10.model.layers.0.mlp.up_proj.weight +3 -0
- checkpoint/pytorch_model.bin/p100.model.layers.7.mlp.gate_proj.weight +3 -0
- checkpoint/pytorch_model.bin/p1000.model.layers.76.input_layernorm.weight +3 -0
- checkpoint/pytorch_model.bin/p1001.model.layers.76.post_attention_layernorm.weight +3 -0
- checkpoint/pytorch_model.bin/p1002.model.layers.77.self_attn.q_proj.weight +3 -0
- checkpoint/pytorch_model.bin/p1003.model.layers.77.self_attn.q_proj.bias +3 -0
- checkpoint/pytorch_model.bin/p1004.model.layers.77.self_attn.k_proj.weight +3 -0
- checkpoint/pytorch_model.bin/p1005.model.layers.77.self_attn.k_proj.bias +3 -0
- checkpoint/pytorch_model.bin/p1006.model.layers.77.self_attn.v_proj.weight +3 -0
- checkpoint/pytorch_model.bin/p1007.model.layers.77.self_attn.v_proj.bias +3 -0
- checkpoint/pytorch_model.bin/p1008.model.layers.77.self_attn.o_proj.weight +3 -0
- checkpoint/pytorch_model.bin/p1009.model.layers.77.self_attn.o_proj.bias +3 -0
- checkpoint/pytorch_model.bin/p101.model.layers.7.mlp.up_proj.weight +3 -0
- checkpoint/pytorch_model.bin/p1010.model.layers.77.mlp.gate_proj.weight +3 -0
- checkpoint/pytorch_model.bin/p1011.model.layers.77.mlp.up_proj.weight +3 -0
- checkpoint/pytorch_model.bin/p1012.model.layers.77.mlp.down_proj.weight +3 -0
- checkpoint/pytorch_model.bin/p1013.model.layers.77.input_layernorm.weight +3 -0
- checkpoint/pytorch_model.bin/p1014.model.layers.77.post_attention_layernorm.weight +3 -0
- checkpoint/pytorch_model.bin/p1015.model.layers.78.self_attn.q_proj.weight +3 -0
- checkpoint/pytorch_model.bin/p1016.model.layers.78.self_attn.q_proj.bias +3 -0
- checkpoint/pytorch_model.bin/p1017.model.layers.78.self_attn.k_proj.weight +3 -0
- checkpoint/pytorch_model.bin/p1018.model.layers.78.self_attn.k_proj.bias +3 -0
- checkpoint/pytorch_model.bin/p1019.model.layers.78.self_attn.v_proj.weight +3 -0
- checkpoint/pytorch_model.bin/p102.model.layers.7.mlp.down_proj.weight +3 -0
- checkpoint/pytorch_model.bin/p1020.model.layers.78.self_attn.v_proj.bias +3 -0
- checkpoint/pytorch_model.bin/p1021.model.layers.78.self_attn.o_proj.weight +3 -0
- checkpoint/pytorch_model.bin/p1022.model.layers.78.self_attn.o_proj.bias +3 -0
- checkpoint/pytorch_model.bin/p1023.model.layers.78.mlp.gate_proj.weight +3 -0
- checkpoint/pytorch_model.bin/p1024.model.layers.78.mlp.up_proj.weight +3 -0
- checkpoint/pytorch_model.bin/p1025.model.layers.78.mlp.down_proj.weight +3 -0
- checkpoint/pytorch_model.bin/p1026.model.layers.78.input_layernorm.weight +3 -0
- checkpoint/pytorch_model.bin/p1027.model.layers.78.post_attention_layernorm.weight +3 -0
- checkpoint/pytorch_model.bin/p1028.model.layers.79.self_attn.q_proj.weight +3 -0
- checkpoint/pytorch_model.bin/p1029.model.layers.79.self_attn.q_proj.bias +3 -0
- checkpoint/pytorch_model.bin/p103.model.layers.7.input_layernorm.weight +3 -0
- checkpoint/pytorch_model.bin/p1030.model.layers.79.self_attn.k_proj.weight +3 -0
- checkpoint/pytorch_model.bin/p1031.model.layers.79.self_attn.k_proj.bias +3 -0
- checkpoint/pytorch_model.bin/p1032.model.layers.79.self_attn.v_proj.weight +3 -0
- checkpoint/pytorch_model.bin/p1033.model.layers.79.self_attn.v_proj.bias +3 -0
- checkpoint/pytorch_model.bin/p1034.model.layers.79.self_attn.o_proj.weight +3 -0
- checkpoint/pytorch_model.bin/p1035.model.layers.79.self_attn.o_proj.bias +3 -0
- checkpoint/pytorch_model.bin/p1036.model.layers.79.mlp.gate_proj.weight +3 -0
- checkpoint/pytorch_model.bin/p1037.model.layers.79.mlp.up_proj.weight +3 -0
- checkpoint/pytorch_model.bin/p1038.model.layers.79.mlp.down_proj.weight +3 -0
.gitattributes
CHANGED
@@ -33,3 +33,13 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
|
|
33 |
*.zip filter=lfs diff=lfs merge=lfs -text
|
34 |
*.zst filter=lfs diff=lfs merge=lfs -text
|
35 |
*tfevents* filter=lfs diff=lfs merge=lfs -text
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
33 |
*.zip filter=lfs diff=lfs merge=lfs -text
|
34 |
*.zst filter=lfs diff=lfs merge=lfs -text
|
35 |
*tfevents* filter=lfs diff=lfs merge=lfs -text
|
36 |
+
compiled/39671fa6153a417c0059.neff filter=lfs diff=lfs merge=lfs -text
|
37 |
+
compiled/4004277e31e89dbc2cd2.neff filter=lfs diff=lfs merge=lfs -text
|
38 |
+
compiled/8273c4799b9ddf453584.neff filter=lfs diff=lfs merge=lfs -text
|
39 |
+
compiled/86668046903bd034a5a6.neff filter=lfs diff=lfs merge=lfs -text
|
40 |
+
compiled/891ea69f293f2939545b.neff filter=lfs diff=lfs merge=lfs -text
|
41 |
+
compiled/8c4b90826998854c36a0.neff filter=lfs diff=lfs merge=lfs -text
|
42 |
+
compiled/933237200bbf8dc1417b.neff filter=lfs diff=lfs merge=lfs -text
|
43 |
+
compiled/93e48a55161931f3aee7.neff filter=lfs diff=lfs merge=lfs -text
|
44 |
+
compiled/cf6f85a4d52f11817865.neff filter=lfs diff=lfs merge=lfs -text
|
45 |
+
compiled/d3b6c95e6bb346ee4b43.neff filter=lfs diff=lfs merge=lfs -text
|
checkpoint/config.json
ADDED
@@ -0,0 +1,30 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"_name_or_path": "Smaug-72B-v0.1/",
|
3 |
+
"architectures": [
|
4 |
+
"LlamaForCausalLM"
|
5 |
+
],
|
6 |
+
"attention_bias": true,
|
7 |
+
"attention_dropout": 0.0,
|
8 |
+
"bos_token_id": 1,
|
9 |
+
"eos_token_id": 151643,
|
10 |
+
"hidden_act": "silu",
|
11 |
+
"hidden_size": 8192,
|
12 |
+
"initializer_range": 0.02,
|
13 |
+
"intermediate_size": 24576,
|
14 |
+
"max_position_embeddings": 32768,
|
15 |
+
"model_type": "llama",
|
16 |
+
"num_attention_heads": 64,
|
17 |
+
"num_hidden_layers": 80,
|
18 |
+
"num_key_value_heads": 64,
|
19 |
+
"pad_token_id": 151643,
|
20 |
+
"pretraining_tp": 1,
|
21 |
+
"rms_norm_eps": 1e-06,
|
22 |
+
"rope_scaling": null,
|
23 |
+
"rope_theta": 1000000,
|
24 |
+
"seq_length": 32768,
|
25 |
+
"tie_word_embeddings": false,
|
26 |
+
"torch_dtype": "float32",
|
27 |
+
"transformers_version": "4.36.2",
|
28 |
+
"use_cache": true,
|
29 |
+
"vocab_size": 152064
|
30 |
+
}
|
checkpoint/generation_config.json
ADDED
@@ -0,0 +1,6 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"_from_model_config": true,
|
3 |
+
"eos_token_id": 151643,
|
4 |
+
"pad_token_id": 151643,
|
5 |
+
"transformers_version": "4.36.2"
|
6 |
+
}
|
checkpoint/pytorch_model.bin/key_to_filename.json
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:96b043d79e15b14221ff9be0360598ae29d60dc3bdff1d41ec9a92ca94ae65c7
|
3 |
+
size 94632
|
checkpoint/pytorch_model.bin/p0.model.embed_tokens.weight
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:cf2ba4dff83df47c51598892b511a6d85d77ae3f711c042c648a28ec0dd6a3c8
|
3 |
+
size 4982834045
|
checkpoint/pytorch_model.bin/p1.model.layers.0.self_attn.q_proj.weight
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:c38ac059caca2da46fd6f9832f39c1e79a0c7e383f0127b72899e6a9fa8cfc38
|
3 |
+
size 268436348
|
checkpoint/pytorch_model.bin/p10.model.layers.0.mlp.up_proj.weight
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:7dd88f35543815914d0b39b61962ffebe2ae768a182374126e71686ad5f626e8
|
3 |
+
size 805307248
|
checkpoint/pytorch_model.bin/p100.model.layers.7.mlp.gate_proj.weight
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:92c50b02de0869b7c80cb16953b4c314a88fc1b1314f1193bf9a443d30c93d67
|
3 |
+
size 805307257
|
checkpoint/pytorch_model.bin/p1000.model.layers.76.input_layernorm.weight
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:19bb3c6146e250c2407a4722815201388976a1df1c46af1c661fca8b2f0bc81a
|
3 |
+
size 33669
|
checkpoint/pytorch_model.bin/p1001.model.layers.76.post_attention_layernorm.weight
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:0d4d800221ad6143f6ab9251efe3a9730451a02b093332da0c9bbd861649040a
|
3 |
+
size 33696
|
checkpoint/pytorch_model.bin/p1002.model.layers.77.self_attn.q_proj.weight
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:13035ca89111cda145c5399987ff6b9b4b492e5006f1cc53b5b7bc95f24a0c4b
|
3 |
+
size 268436360
|
checkpoint/pytorch_model.bin/p1003.model.layers.77.self_attn.q_proj.bias
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:02ffe7f1e60cd04f776a237de2cce85eb10789c31a2de02d8574d1ccdcbe4f97
|
3 |
+
size 33672
|
checkpoint/pytorch_model.bin/p1004.model.layers.77.self_attn.k_proj.weight
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:edbcf6ebc07dd0acfdf1819841ded195e5e1fb8cc3d9eafe9d87f32b41f39582
|
3 |
+
size 268436360
|
checkpoint/pytorch_model.bin/p1005.model.layers.77.self_attn.k_proj.bias
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:34900538f22980f303d54a6e2b5b1e49ea36ed985be637af73d31c8e0b45d852
|
3 |
+
size 33672
|
checkpoint/pytorch_model.bin/p1006.model.layers.77.self_attn.v_proj.weight
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:ca578550b4d4d961f2d7bba0e076a50c4dbe5cbb8b16d99594e2cc41dc654c8a
|
3 |
+
size 268436360
|
checkpoint/pytorch_model.bin/p1007.model.layers.77.self_attn.v_proj.bias
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:2f1b0f95d39db390c677155ba2ec6f9a5a1537043759e14453658462e0ac2fcf
|
3 |
+
size 33672
|
checkpoint/pytorch_model.bin/p1008.model.layers.77.self_attn.o_proj.weight
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:4e38d3cf2e779558f117236f05b86e52e58c9ba5aa103fd2181d086dc1184659
|
3 |
+
size 268436360
|
checkpoint/pytorch_model.bin/p1009.model.layers.77.self_attn.o_proj.bias
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:7ee348eeb2703a35bf6d9e4b170fb0855ac8ad719cb48119baaef3e11b3f3dd8
|
3 |
+
size 33672
|
checkpoint/pytorch_model.bin/p101.model.layers.7.mlp.up_proj.weight
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:c052eceb6937fae9c8ff34c51f5534ea36944a2a59965cfd812306d1f535e13a
|
3 |
+
size 805307251
|
checkpoint/pytorch_model.bin/p1010.model.layers.77.mlp.gate_proj.weight
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:6a08e113ca55fb0937034ad67cd08409a25d011d3632a84b964d54de656e0aeb
|
3 |
+
size 805307263
|
checkpoint/pytorch_model.bin/p1011.model.layers.77.mlp.up_proj.weight
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:84e5a2a121d13c58e96f97c6d963915b99b5870d95ba8375a71cfea36bb930bc
|
3 |
+
size 805307257
|
checkpoint/pytorch_model.bin/p1012.model.layers.77.mlp.down_proj.weight
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:7042573bac0b64307d1e52da23c52c1d94a135c8756b030220776237700a50db
|
3 |
+
size 805307263
|
checkpoint/pytorch_model.bin/p1013.model.layers.77.input_layernorm.weight
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:c6556366826ef92f791fd854c422a837e23ab12002b47b7a52767b6498f40448
|
3 |
+
size 33669
|
checkpoint/pytorch_model.bin/p1014.model.layers.77.post_attention_layernorm.weight
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:0e92b8afcb8454b148b8698f0ed41166b05b84440e9600c234005e5e2f7f1f82
|
3 |
+
size 33696
|
checkpoint/pytorch_model.bin/p1015.model.layers.78.self_attn.q_proj.weight
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:def16bdc409447f5b8686dc44c9564cc693b480a1b355b234c9a25ced4f44865
|
3 |
+
size 268436360
|
checkpoint/pytorch_model.bin/p1016.model.layers.78.self_attn.q_proj.bias
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:288532a9332460dba0cf6d702ec2c6df7f6cefa106b55a6aed4be88cd0265cd5
|
3 |
+
size 33672
|
checkpoint/pytorch_model.bin/p1017.model.layers.78.self_attn.k_proj.weight
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:fd07a7f67337533e714c131af0b8115960637398d83d790ba14d31ac4c10ce39
|
3 |
+
size 268436360
|
checkpoint/pytorch_model.bin/p1018.model.layers.78.self_attn.k_proj.bias
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:2cb8b18d4e40ec94b3117826e682b09ae98e6f5ac1f4c29037ee184239f91fad
|
3 |
+
size 33672
|
checkpoint/pytorch_model.bin/p1019.model.layers.78.self_attn.v_proj.weight
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:f6a0370e5e8ae0c85c895579f5efe50865e5835bd381049ca4b02cbd34e57380
|
3 |
+
size 268436360
|
checkpoint/pytorch_model.bin/p102.model.layers.7.mlp.down_proj.weight
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:d1b784afda55f8ecfb32d7c2090f0f50d9fd2a480f5c8d061342e0540314da9e
|
3 |
+
size 805307257
|
checkpoint/pytorch_model.bin/p1020.model.layers.78.self_attn.v_proj.bias
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:ff2f99a64c5dddc47157f2bb39658d560701649f0fe6c1b5c6f7abd023675bd9
|
3 |
+
size 33672
|
checkpoint/pytorch_model.bin/p1021.model.layers.78.self_attn.o_proj.weight
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:13604c6e361891aefc71f1378a38a29a554ad923be116abfa346e3e2bbde9242
|
3 |
+
size 268436360
|
checkpoint/pytorch_model.bin/p1022.model.layers.78.self_attn.o_proj.bias
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:657aadf2e69572adddc88a4a3240708ad7d555e90ab71578ceef0d5fa193b2ea
|
3 |
+
size 33672
|
checkpoint/pytorch_model.bin/p1023.model.layers.78.mlp.gate_proj.weight
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:88e146b9dd48639c765f574d168ca3531b5b2302e55f52fb54cf9ecd7faeba97
|
3 |
+
size 805307263
|
checkpoint/pytorch_model.bin/p1024.model.layers.78.mlp.up_proj.weight
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:27c65d77479b717511b92c0510c78606edd2c095089e6ada45029ebedb0470f4
|
3 |
+
size 805307257
|
checkpoint/pytorch_model.bin/p1025.model.layers.78.mlp.down_proj.weight
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:7f652d93bd7527139330d16859b522494ab62f2bca0b48b9f5353fed77e2873a
|
3 |
+
size 805307263
|
checkpoint/pytorch_model.bin/p1026.model.layers.78.input_layernorm.weight
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:1c059e62110f6ffd08cc1877f1bd873cefaab8b3fc7fa614870c87f8708c92ac
|
3 |
+
size 33669
|
checkpoint/pytorch_model.bin/p1027.model.layers.78.post_attention_layernorm.weight
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:9390d4913d8d9a7f6eb4a52ed6066bd99d4c8a0cde28db624947c39b5092cea0
|
3 |
+
size 33696
|
checkpoint/pytorch_model.bin/p1028.model.layers.79.self_attn.q_proj.weight
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:fc4203e66420cf3af9ec409d3dbc4275bbe51efe5ae5b13e4113acefa612753a
|
3 |
+
size 268436360
|
checkpoint/pytorch_model.bin/p1029.model.layers.79.self_attn.q_proj.bias
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:f65142b19918cafb31ff7999188e0cae5ce6f3bd3bc10798a96354e473499a39
|
3 |
+
size 33672
|
checkpoint/pytorch_model.bin/p103.model.layers.7.input_layernorm.weight
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:a96748c23d540fce2a2b825e4f1fd377fa3b91fa6ef77f30bb30b51a9bba53b9
|
3 |
+
size 33663
|
checkpoint/pytorch_model.bin/p1030.model.layers.79.self_attn.k_proj.weight
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:1587ba215131195fe4829a95b91dd3405e1f707b58d1bf44300641c74fc011a9
|
3 |
+
size 268436360
|
checkpoint/pytorch_model.bin/p1031.model.layers.79.self_attn.k_proj.bias
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:01cd37eb94c86a33b304edec230116649740f6d1929067245455c542c8b49f81
|
3 |
+
size 33672
|
checkpoint/pytorch_model.bin/p1032.model.layers.79.self_attn.v_proj.weight
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:6b7044897d53e08ed91cd4d64145ec7bfe45b2c206bc8323e453d5fd11efbf29
|
3 |
+
size 268436360
|
checkpoint/pytorch_model.bin/p1033.model.layers.79.self_attn.v_proj.bias
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:f8cc4dc8b3d5697a55e5e178c13feffc6151dff98c662cd41ed9271a95d1dc1a
|
3 |
+
size 33672
|
checkpoint/pytorch_model.bin/p1034.model.layers.79.self_attn.o_proj.weight
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:6183c14a67e5c09c0d57611b6d04f1e61e0cc92025b372dfe88aafb931abfe99
|
3 |
+
size 268436360
|
checkpoint/pytorch_model.bin/p1035.model.layers.79.self_attn.o_proj.bias
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:4943793ccbafe4b0dfe082b6b7b9aff21973d7934a2fb5810230b9eaa1466665
|
3 |
+
size 33672
|
checkpoint/pytorch_model.bin/p1036.model.layers.79.mlp.gate_proj.weight
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:e01b3446baf8c407683141d85e48bcd12b7de1dc862af39e9d34460d560bc683
|
3 |
+
size 805307263
|
checkpoint/pytorch_model.bin/p1037.model.layers.79.mlp.up_proj.weight
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:7877e35babcb3a57f8a0fe8270cd2dc8bc019b68667d95d8505a7ae2d30146e6
|
3 |
+
size 805307257
|
checkpoint/pytorch_model.bin/p1038.model.layers.79.mlp.down_proj.weight
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:5cc4dd463df45331a9159d4d22229d520e41bc96400e164a87da3adad62fbf09
|
3 |
+
size 805307263
|