haydn-jones
commited on
Commit
•
059cdd9
1
Parent(s):
9f62b36
Upload folder using huggingface_hub
Browse files- adapter_0.pt +3 -0
- adapter_1.pt +3 -0
- adapter_10.pt +3 -0
- adapter_11.pt +3 -0
- adapter_12.pt +3 -0
- adapter_13.pt +3 -0
- adapter_14.pt +3 -0
- adapter_15.pt +3 -0
- adapter_16.pt +3 -0
- adapter_17.pt +3 -0
- adapter_18.pt +3 -0
- adapter_19.pt +3 -0
- adapter_2.pt +3 -0
- adapter_20.pt +3 -0
- adapter_21.pt +3 -0
- adapter_22.pt +3 -0
- adapter_23.pt +3 -0
- adapter_3.pt +3 -0
- adapter_4.pt +3 -0
- adapter_5.pt +3 -0
- adapter_6.pt +3 -0
- adapter_7.pt +3 -0
- adapter_8.pt +3 -0
- adapter_9.pt +3 -0
- adapter_config.json +1 -0
- adapter_model.bin +3 -0
- config.json +1 -0
- pytorch_adapter.bin +3 -0
- recipe_state.pt +3 -0
adapter_0.pt
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:d587b9df00f281a729abee841e79c352522d9fb47148d6c832f6acb012ed3e90
|
3 |
+
size 42083770
|
adapter_1.pt
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:ab6e7747089e61a9496b17f6ec1c4ecd4ac4e51b656b46685a1bb87bc10db496
|
3 |
+
size 42083770
|
adapter_10.pt
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:61d573687e96fdeae9cf896373ef771c72f2e14094ac64b30c3eab3257c69068
|
3 |
+
size 42084222
|
adapter_11.pt
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:19e9ba055ee2def06337675394a90a29dba813cf61fe931d7f223d6ac06daf48
|
3 |
+
size 42084222
|
adapter_12.pt
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:d52a34579750916301107e34110223ca9addba554c7e39aef2566a4af4c5ad72
|
3 |
+
size 42084222
|
adapter_13.pt
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:563844eb33a9f8bd56a598d83aa84e689512d3f644b4ced94c5a85c9e61c3b42
|
3 |
+
size 42084222
|
adapter_14.pt
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:8ca98fce8d59341bf85ee4703b6bd682b7ef54551aea7b1f0025dc23c12492cd
|
3 |
+
size 42084222
|
adapter_15.pt
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:f2498f6b7c18908643877631d0352c20d2f01e786f03189580c977a6642b41fe
|
3 |
+
size 42084222
|
adapter_16.pt
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:eb58a01bda7bb8861a964747755d733d28f623dd163acb082a4d39107367f306
|
3 |
+
size 42084222
|
adapter_17.pt
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:bcfba4d17fa795500ad4e8ec068aed6a647e8d97bf3b529f8731e46195c6f713
|
3 |
+
size 42084222
|
adapter_18.pt
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:419af8765fa3cc9576014d4ddb33b7251a641e1cacb696f7ca64946f8f7b5963
|
3 |
+
size 42084222
|
adapter_19.pt
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:e7abc03a8f80d8c2c1aa03be9a031f19e2ff9246feb88b1748c27be0097f68a5
|
3 |
+
size 42084222
|
adapter_2.pt
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:88044ffe330998fb4e7cb43a85c73d203f722456cd6774a6bc6395571743ca1f
|
3 |
+
size 42083770
|
adapter_20.pt
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:7309f4f1d6cf95df6552f716d74c75a955823f20b7bf949a48f2290ce6c2f46b
|
3 |
+
size 42084222
|
adapter_21.pt
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:6dea9a9146cebc6082aa229b2ac569be16b39f5be738893b754a763012c62f34
|
3 |
+
size 42084222
|
adapter_22.pt
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:74c89d90de1caaf5a5486ec528957a09dbf8c5a4fb81f7fcf12790e92ece937f
|
3 |
+
size 42084222
|
adapter_23.pt
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:2c05b6d43dc3fd50b2c89dc1d006f3aebf79cb7fccac71cf6d00f48320963997
|
3 |
+
size 42084222
|
adapter_3.pt
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:3526e4f947974034825667f9b013ec97ed2722adbb2a82e353b50f5c17651105
|
3 |
+
size 42083770
|
adapter_4.pt
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:f58ca7145aff86e9a50223351105f908902da1234858479a1a708474703b5e58
|
3 |
+
size 42083770
|
adapter_5.pt
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:26eef3dd35ef682059507a38f12d870b9026c3750885ab877e7746f8892fbe60
|
3 |
+
size 42083770
|
adapter_6.pt
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:a5f91ec1903a1e0e7955a4314988f13f5019a6f02165816b44b20806f1bf047c
|
3 |
+
size 42083770
|
adapter_7.pt
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:4c091079b49231ea26d23b97d35b6f56eff5824579e88b7ea786a0cec63c7202
|
3 |
+
size 42083770
|
adapter_8.pt
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:6df44fc5f52d62b5ada14086ec7d1b6cfa8805ac2e3239a5103552f1b7eb64d3
|
3 |
+
size 42083770
|
adapter_9.pt
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:c73e9c89798aab9aa178b11eca3ab92e40954890cc4ef9a3de203bfaf8f78f32
|
3 |
+
size 42083770
|
adapter_config.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"r": 8, "lora_alpha": 16, "target_modules": ["q_proj", "v_proj", "k_proj", "o_proj", "gate_proj", "down_proj", "up_proj"], "peft_type": "LORA"}
|
adapter_model.bin
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:05d09cce93b07429468f1b197d43f46da0c433626ee6aefc877ffb79d7f4691f
|
3 |
+
size 42098058
|
config.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"architectures": ["LlamaForCausalLM"], "attention_bias": false, "attention_dropout": 0.0, "bos_token_id": 128000, "eos_token_id": [128001, 128008, 128009], "hidden_act": "silu", "hidden_size": 4096, "initializer_range": 0.02, "intermediate_size": 14336, "max_position_embeddings": 131072, "mlp_bias": false, "model_type": "llama", "num_attention_heads": 32, "num_hidden_layers": 32, "num_key_value_heads": 8, "pretraining_tp": 1, "rms_norm_eps": 1e-05, "rope_scaling": {"factor": 8.0, "low_freq_factor": 1.0, "high_freq_factor": 4.0, "original_max_position_embeddings": 8192, "rope_type": "llama3"}, "rope_theta": 500000.0, "tie_word_embeddings": false, "torch_dtype": "bfloat16", "transformers_version": "4.42.3", "use_cache": true, "vocab_size": 128256}
|
pytorch_adapter.bin
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:2c05b6d43dc3fd50b2c89dc1d006f3aebf79cb7fccac71cf6d00f48320963997
|
3 |
+
size 42084222
|
recipe_state.pt
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:0f9e280fd351e7f341eb569e21c9f554b1dc03cbd66feada57bb787e902072a8
|
3 |
+
size 84274334
|