quantumaikr commited on
Commit
94eb90f
·
1 Parent(s): 791fa62

Upload LlamaForCausalLM

Browse files
config.json CHANGED
@@ -18,6 +18,7 @@
18
  "pretraining_tp": 1,
19
  "rms_norm_eps": 1e-06,
20
  "rope_scaling": null,
 
21
  "tie_word_embeddings": false,
22
  "torch_dtype": "float32",
23
  "transformers_version": "4.33.0.dev0",
 
18
  "pretraining_tp": 1,
19
  "rms_norm_eps": 1e-06,
20
  "rope_scaling": null,
21
+ "rope_theta": 10000.0,
22
  "tie_word_embeddings": false,
23
  "torch_dtype": "float32",
24
  "transformers_version": "4.33.0.dev0",
model-00001-of-00004.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:ee7fe926b74f17c1c60ee67b155d32cd3509f82f5ffda39540d245cda10167b6
3
  size 2996969872
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:873e0a7c7eea1c703e48a2c1f9887d7804cc0fdc10e7615508830fd6b4a35b72
3
  size 2996969872
model-00002-of-00004.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:488e31fa878e32f6e66fe5d95d709ef9f1733a233e5cb4bda27cf077836311a6
3
  size 2915181104
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:768bd139a2a50d7e08da84029c35c052c80e0c9f01a034012fa17ef5a293f814
3
  size 2915181104
model-00003-of-00004.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:6842c40992958854510e94cf03906c67a684db8b7ed464df6ea85c3a476a5c65
3
  size 2948752400
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8d791e889c8d13bf18e0105467844d62cf27283fc6243c25bd723e4136378ff2
3
  size 2948752400
model-00004-of-00004.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:0c07afe0a0f19ff44f5bce06457aa954302acb07855d35a7912fcbf6838d32d8
3
  size 2468477680
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ca7bec1f007e9cb762f86a16ab1e92e7118a582242b2b2581979ea5b9d651750
3
  size 2468477680