Initial GPTQ model commit.
Browse files
README.md
CHANGED
@@ -5,17 +5,19 @@ datasets:
|
|
5 |
inference: false
|
6 |
---
|
7 |
|
8 |
-
# WizardLM
|
9 |
|
10 |
-
These files are GPTQ 4bit model files for [Eric Hartford's
|
11 |
|
12 |
It is the result of quantising to 4bit using [GPTQ-for-LLaMa](https://github.com/qwopqwop200/GPTQ-for-LLaMa).
|
13 |
|
|
|
|
|
14 |
## Other repositories available
|
15 |
|
16 |
-
* [4bit GPTQ
|
17 |
-
* [
|
18 |
-
* [Eric's unquantised model in
|
19 |
|
20 |
## How to easily download and use this model in text-generation-webui
|
21 |
|
@@ -43,7 +45,7 @@ This will work with all versions of GPTQ-for-LLaMa. It has maximum compatibility
|
|
43 |
|
44 |
It was created without the `--act-order` parameter. It may have slightly lower inference quality compared to the other file, but is guaranteed to work on all versions of GPTQ-for-LLaMa and text-generation-webui.
|
45 |
|
46 |
-
* `
|
47 |
* Works with all versions of GPTQ-for-LLaMa code, both Triton and CUDA branches
|
48 |
* Works with AutoGPTQ. Use `strict=False` to load.
|
49 |
* Works with text-generation-webui one-click-installers
|
|
|
5 |
inference: false
|
6 |
---
|
7 |
|
8 |
+
# WizardLM - uncensored: An Instruction-following LLM Using Evol-Instruct
|
9 |
|
10 |
+
These files are GPTQ 4bit model files for [Eric Hartford's 'uncensored' version of WizardLM](https://huggingface.co/ehartford/WizardLM-30B-Uncensored).
|
11 |
|
12 |
It is the result of quantising to 4bit using [GPTQ-for-LLaMa](https://github.com/qwopqwop200/GPTQ-for-LLaMa).
|
13 |
|
14 |
+
Eric did a fresh 7B training using the WizardLM method, on [a dataset edited to remove all the "I'm sorry.." type ChatGPT responses](https://huggingface.co/datasets/ehartford/WizardLM_alpaca_evol_instruct_70k_unfiltered).
|
15 |
+
|
16 |
## Other repositories available
|
17 |
|
18 |
+
* [4bit GPTQ models for GPU inference](https://huggingface.co/TheBloke/WizardLM-30B-uncensored-GPTQ)
|
19 |
+
* [4bit and 5bit GGML models for CPU inference](https://huggingface.co/TheBloke/WizardLM-30B-uncensored-GGML)
|
20 |
+
* [Eric's unquantised model in HF format](https://huggingface.co/ehartford/WizardLM-30B-Uncensored)
|
21 |
|
22 |
## How to easily download and use this model in text-generation-webui
|
23 |
|
|
|
45 |
|
46 |
It was created without the `--act-order` parameter. It may have slightly lower inference quality compared to the other file, but is guaranteed to work on all versions of GPTQ-for-LLaMa and text-generation-webui.
|
47 |
|
48 |
+
* `wizard-vicuna-13B-GPTQ-4bit.compat.no-act-order.safetensors`
|
49 |
* Works with all versions of GPTQ-for-LLaMa code, both Triton and CUDA branches
|
50 |
* Works with AutoGPTQ. Use `strict=False` to load.
|
51 |
* Works with text-generation-webui one-click-installers
|