Upload folder using huggingface_hub (#2)
Browse files- 07b3cf58c1862764a90fecaf9f608b47723441706444ce23258a10c6355d86f9 (21a42a7438fca2c9b70dd672e4023e4ae9ee28ad)
- bd8bb1703c52fbc27b7e7b703f791bc91635a73eca1cfeb18125b64281de028d (63eb46ffd9dcc7748f2d8d08bd3ca32abb81ac7d)
- 0bc263b4215f034358c93324d4e34084ae32daff4b21fa6f096dc358867aed20 (ed7c04bb69f7f0961b14490c7f4b827ab55dadbb)
- fb75ce3d8c137bef8edecafd778ad754a21c0244a304825002b816f6002784f9 (e99fcdae07e9667cf4f34bed9849b07b6bec5a31)
- a7b55e4dd2c9dfddc5ba2dbc9dc1c9e8691b02d6f81bf75d03d3d5cfac66f9bf (cc8c835188b6a99054404d587477267468c1ffa8)
- 1ce5035b135261e8d2a4fd708a47ca66266de3131e324c582d3a932df9be29b2 (1169a655225734e39ae096d7024acae57dfc56bd)
- 994c93e70ebc319cbef858d856ea23b083b8be32e2d8dd9628c91aa79df7e432 (b6e6b4d58a981876abdbb119b096f860d818238a)
- c7fec94d7fe1386d8a03b1df4b93967f165e3bea0ca5249909523faa3fc696bf (2bddbcbe199a6daa92d9459dd5b90f6e14f12617)
- 63770611d12348b12abd16a94b63c03ed7220e717659e4f7ee51cb973e7dc81f (a7f36682f8e423b5459ea97226ea6367364c9293)
- e35207287542a3e89b8c476e0e7a51a2434a7a9913b099566104d236fc1b116c (2e178bc4e3a2130222f98aab6f813fe8e828e147)
- eeecca46d478625dac5d17574120522c6975e18eb8269eaf8bc9dbcd8113de28 (77619691f0148ab619362b462c9a7524a263887e)
- a73aef439e689e5bfd233dcb4c959c1d03f7b69f7d0be38950c8f526d4bbea61 (ab7db8d0ec76c2a773d9b5c5f16a4c89a9d88417)
- 38d3509ce766513f2f506d576c12c8ee6150c9208cc80644b969f9b9a2f7d5d5 (8c6897e4c1f63387b3a5707291debb78da4060e6)
- .gitattributes +2 -0
- README.md +8 -7
- tulu-2-dpo-13b.IQ3_XS.gguf +3 -0
- tulu-2-dpo-13b.IQ4_XS.gguf +3 -0
- tulu-2-dpo-13b.Q2_K.gguf +2 -2
- tulu-2-dpo-13b.Q3_K_L.gguf +2 -2
- tulu-2-dpo-13b.Q3_K_M.gguf +2 -2
- tulu-2-dpo-13b.Q3_K_S.gguf +2 -2
- tulu-2-dpo-13b.Q4_K_M.gguf +2 -2
- tulu-2-dpo-13b.Q4_K_S.gguf +2 -2
- tulu-2-dpo-13b.Q5_K_M.gguf +2 -2
- tulu-2-dpo-13b.Q5_K_S.gguf +2 -2
- tulu-2-dpo-13b.Q6_K.gguf +2 -2
- tulu-2-dpo-13b.Q8_0.gguf +2 -2
@@ -43,3 +43,5 @@ tulu-2-dpo-13b.Q5_K_M.gguf filter=lfs diff=lfs merge=lfs -text
|
|
43 |
tulu-2-dpo-13b.Q5_K_S.gguf filter=lfs diff=lfs merge=lfs -text
|
44 |
tulu-2-dpo-13b.Q6_K.gguf filter=lfs diff=lfs merge=lfs -text
|
45 |
tulu-2-dpo-13b.Q8_0.gguf filter=lfs diff=lfs merge=lfs -text
|
|
|
|
|
|
43 |
tulu-2-dpo-13b.Q5_K_S.gguf filter=lfs diff=lfs merge=lfs -text
|
44 |
tulu-2-dpo-13b.Q6_K.gguf filter=lfs diff=lfs merge=lfs -text
|
45 |
tulu-2-dpo-13b.Q8_0.gguf filter=lfs diff=lfs merge=lfs -text
|
46 |
+
tulu-2-dpo-13b.IQ3_XS.gguf filter=lfs diff=lfs merge=lfs -text
|
47 |
+
tulu-2-dpo-13b.IQ4_XS.gguf filter=lfs diff=lfs merge=lfs -text
|
@@ -10,6 +10,7 @@ tags:
|
|
10 |
- GGUF
|
11 |
- transformers
|
12 |
- pytorch
|
|
|
13 |
- llama
|
14 |
- text-generation
|
15 |
- en
|
@@ -98,7 +99,7 @@ pip3 install huggingface-hub
|
|
98 |
Then you can download any individual model file to the current directory, at high speed, with a command like this:
|
99 |
|
100 |
```shell
|
101 |
-
huggingface-cli download MaziyarPanahi/tulu-2-dpo-13b-GGUF tulu-2-dpo-13b
|
102 |
```
|
103 |
</details>
|
104 |
<details>
|
@@ -121,7 +122,7 @@ pip3 install hf_transfer
|
|
121 |
And set environment variable `HF_HUB_ENABLE_HF_TRANSFER` to `1`:
|
122 |
|
123 |
```shell
|
124 |
-
HF_HUB_ENABLE_HF_TRANSFER=1 huggingface-cli download MaziyarPanahi/tulu-2-dpo-13b-GGUF tulu-2-dpo-13b
|
125 |
```
|
126 |
|
127 |
Windows Command Line users: You can set the environment variable by running `set HF_HUB_ENABLE_HF_TRANSFER=1` before the download command.
|
@@ -132,7 +133,7 @@ Windows Command Line users: You can set the environment variable by running `set
|
|
132 |
Make sure you are using `llama.cpp` from commit [d0cee0d](https://github.com/ggerganov/llama.cpp/commit/d0cee0d36d5be95a0d9088b674dbb27354107221) or later.
|
133 |
|
134 |
```shell
|
135 |
-
./main -ngl 35 -m tulu-2-dpo-13b
|
136 |
{system_message}<|im_end|>
|
137 |
<|im_start|>user
|
138 |
{prompt}<|im_end|>
|
@@ -149,7 +150,7 @@ For other parameters and how to use them, please refer to [the llama.cpp documen
|
|
149 |
|
150 |
## How to run in `text-generation-webui`
|
151 |
|
152 |
-
Further instructions can be found in the text-generation-webui documentation, here: [text-generation-webui/docs/04 ‐ Model Tab.md](https://github.com/oobabooga/text-generation-webui/blob/main/docs/04%20
|
153 |
|
154 |
## How to run from Python code
|
155 |
|
@@ -157,7 +158,7 @@ You can use GGUF models from Python using the [llama-cpp-python](https://github.
|
|
157 |
|
158 |
### How to load this model in Python code, using llama-cpp-python
|
159 |
|
160 |
-
For full documentation, please see: [llama-cpp-python docs](https://
|
161 |
|
162 |
#### First install the package
|
163 |
|
@@ -189,7 +190,7 @@ from llama_cpp import Llama
|
|
189 |
|
190 |
# Set gpu_layers to the number of layers to offload to GPU. Set to 0 if no GPU acceleration is available on your system.
|
191 |
llm = Llama(
|
192 |
-
model_path="./tulu-2-dpo-13b
|
193 |
n_ctx=32768, # The max sequence length to use - note that longer sequence lengths require much more resources
|
194 |
n_threads=8, # The number of CPU threads to use, tailor to your system and the resulting performance
|
195 |
n_gpu_layers=35 # The number of layers to offload to GPU, if you have GPU acceleration available
|
@@ -209,7 +210,7 @@ output = llm(
|
|
209 |
|
210 |
# Chat Completion API
|
211 |
|
212 |
-
llm = Llama(model_path="./tulu-2-dpo-13b
|
213 |
llm.create_chat_completion(
|
214 |
messages = [
|
215 |
{"role": "system", "content": "You are a story writing assistant."},
|
|
|
10 |
- GGUF
|
11 |
- transformers
|
12 |
- pytorch
|
13 |
+
- safetensors
|
14 |
- llama
|
15 |
- text-generation
|
16 |
- en
|
|
|
99 |
Then you can download any individual model file to the current directory, at high speed, with a command like this:
|
100 |
|
101 |
```shell
|
102 |
+
huggingface-cli download MaziyarPanahi/tulu-2-dpo-13b-GGUF tulu-2-dpo-13b.Q4_K_M.gguf --local-dir . --local-dir-use-symlinks False
|
103 |
```
|
104 |
</details>
|
105 |
<details>
|
|
|
122 |
And set environment variable `HF_HUB_ENABLE_HF_TRANSFER` to `1`:
|
123 |
|
124 |
```shell
|
125 |
+
HF_HUB_ENABLE_HF_TRANSFER=1 huggingface-cli download MaziyarPanahi/tulu-2-dpo-13b-GGUF tulu-2-dpo-13b.Q4_K_M.gguf --local-dir . --local-dir-use-symlinks False
|
126 |
```
|
127 |
|
128 |
Windows Command Line users: You can set the environment variable by running `set HF_HUB_ENABLE_HF_TRANSFER=1` before the download command.
|
|
|
133 |
Make sure you are using `llama.cpp` from commit [d0cee0d](https://github.com/ggerganov/llama.cpp/commit/d0cee0d36d5be95a0d9088b674dbb27354107221) or later.
|
134 |
|
135 |
```shell
|
136 |
+
./main -ngl 35 -m tulu-2-dpo-13b.Q4_K_M.gguf --color -c 32768 --temp 0.7 --repeat_penalty 1.1 -n -1 -p "<|im_start|>system
|
137 |
{system_message}<|im_end|>
|
138 |
<|im_start|>user
|
139 |
{prompt}<|im_end|>
|
|
|
150 |
|
151 |
## How to run in `text-generation-webui`
|
152 |
|
153 |
+
Further instructions can be found in the text-generation-webui documentation, here: [text-generation-webui/docs/04 ‐ Model Tab.md](https://github.com/oobabooga/text-generation-webui/blob/main/docs/04%20-%20Model%20Tab.md#llamacpp).
|
154 |
|
155 |
## How to run from Python code
|
156 |
|
|
|
158 |
|
159 |
### How to load this model in Python code, using llama-cpp-python
|
160 |
|
161 |
+
For full documentation, please see: [llama-cpp-python docs](https://github.com/abetlen/llama-cpp-python/).
|
162 |
|
163 |
#### First install the package
|
164 |
|
|
|
190 |
|
191 |
# Set gpu_layers to the number of layers to offload to GPU. Set to 0 if no GPU acceleration is available on your system.
|
192 |
llm = Llama(
|
193 |
+
model_path="./tulu-2-dpo-13b.Q4_K_M.gguf", # Download the model file first
|
194 |
n_ctx=32768, # The max sequence length to use - note that longer sequence lengths require much more resources
|
195 |
n_threads=8, # The number of CPU threads to use, tailor to your system and the resulting performance
|
196 |
n_gpu_layers=35 # The number of layers to offload to GPU, if you have GPU acceleration available
|
|
|
210 |
|
211 |
# Chat Completion API
|
212 |
|
213 |
+
llm = Llama(model_path="./tulu-2-dpo-13b.Q4_K_M.gguf", chat_format="llama-2") # Set chat_format according to the model you are using
|
214 |
llm.create_chat_completion(
|
215 |
messages = [
|
216 |
{"role": "system", "content": "You are a story writing assistant."},
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:7a4179d21f4435c64d8fff0b7bd65cb1086169c10f7c84d2a6daccf7adfbaf8c
|
3 |
+
size 5361610944
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:ffbee25bd18bda9bf6f409ad876fae1bd30de5b53c3d7764978fd02554fae058
|
3 |
+
size 7019518144
|
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
-
size
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:9ae74fb7d1bc709a348b650134f50a66e541d64c31006f526b106733016857ee
|
3 |
+
size 4854270144
|
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
-
size
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:f0fc9d4b2846fdff1d2a0a39258cd1572ce51ee5ae0ff17b9185975cbfbc1326
|
3 |
+
size 6929559744
|
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
-
size
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:86977b49e1cd0b5b7dcc8fbfbb3667ad7d39b671865e8adb37c3216a73518fc8
|
3 |
+
size 6337769664
|
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
-
size
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:d8b6b66fc333972aa64380427d5e3362fbe91e3588d2827c7856ee8dc0e8f5a3
|
3 |
+
size 5658980544
|
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
-
size
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:59e253478d970b952d8204d0ad52c3f87959d027f7b240c1c00ba5107e8d36d5
|
3 |
+
size 7865956544
|
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
-
size
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:53347c88ca23c6f636421f45d538f7a7e47a1b8be10966b01fcbb488b37c1fb3
|
3 |
+
size 7423178944
|
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
-
size
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:1dd54cb5f6b88b72dba5348c5d1da3ce20b140b380fa3d84aaf300cd9031b213
|
3 |
+
size 9229924544
|
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
-
size
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:4d7b4264fa1e76563274dc6936dc3e2dcaa600eca10c9bb1f2145e5eb8cd9570
|
3 |
+
size 8972286144
|
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
-
size
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:70d83833cf611dbdbb423b605a3381b91129a519cc2fbbf0b13185b898c73716
|
3 |
+
size 10679140544
|
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
-
size
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:6fbb27802565c81c9b18678359e1cf6a930601e1d3ab0a33607606aa4ef99b07
|
3 |
+
size 13831319744
|