quantumaikr
commited on
Commit
·
94eb90f
1
Parent(s):
791fa62
Upload LlamaForCausalLM
Browse files
config.json
CHANGED
@@ -18,6 +18,7 @@
|
|
18 |
"pretraining_tp": 1,
|
19 |
"rms_norm_eps": 1e-06,
|
20 |
"rope_scaling": null,
|
|
|
21 |
"tie_word_embeddings": false,
|
22 |
"torch_dtype": "float32",
|
23 |
"transformers_version": "4.33.0.dev0",
|
|
|
18 |
"pretraining_tp": 1,
|
19 |
"rms_norm_eps": 1e-06,
|
20 |
"rope_scaling": null,
|
21 |
+
"rope_theta": 10000.0,
|
22 |
"tie_word_embeddings": false,
|
23 |
"torch_dtype": "float32",
|
24 |
"transformers_version": "4.33.0.dev0",
|
model-00001-of-00004.safetensors
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 2996969872
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:873e0a7c7eea1c703e48a2c1f9887d7804cc0fdc10e7615508830fd6b4a35b72
|
3 |
size 2996969872
|
model-00002-of-00004.safetensors
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 2915181104
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:768bd139a2a50d7e08da84029c35c052c80e0c9f01a034012fa17ef5a293f814
|
3 |
size 2915181104
|
model-00003-of-00004.safetensors
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 2948752400
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:8d791e889c8d13bf18e0105467844d62cf27283fc6243c25bd723e4136378ff2
|
3 |
size 2948752400
|
model-00004-of-00004.safetensors
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 2468477680
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:ca7bec1f007e9cb762f86a16ab1e92e7118a582242b2b2581979ea5b9d651750
|
3 |
size 2468477680
|