Update README.md
Browse files
README.md
CHANGED
@@ -38,7 +38,7 @@ Open the text-generation-webui UI as normal.
|
|
38 |
|
39 |
This will work with all versions of GPTQ-for-LLaMa. It has maximum compatibility.
|
40 |
|
41 |
-
It was created without `--act-order` to
|
42 |
|
43 |
* `wizard-mega-13B-GPTQ-4bit-128g.safetensors`
|
44 |
* Works with all versions of GPTQ-for-LLaMa code, both Triton and CUDA branches
|
@@ -46,7 +46,7 @@ It was created without `--act-order` to increase quantisation quality, but witho
|
|
46 |
* Parameters: Groupsize = 128. No act-order.
|
47 |
* Command used to create the GPTQ:
|
48 |
```
|
49 |
-
python llama.py /workspace/models/openaccess-ai-collective_wizard-mega-13b c4 --wbits 4 --
|
50 |
```
|
51 |
|
52 |
|
|
|
38 |
|
39 |
This will work with all versions of GPTQ-for-LLaMa. It has maximum compatibility.
|
40 |
|
41 |
+
It was created without `--act-order` to ensure compatibility with all UIs out there.
|
42 |
|
43 |
* `wizard-mega-13B-GPTQ-4bit-128g.safetensors`
|
44 |
* Works with all versions of GPTQ-for-LLaMa code, both Triton and CUDA branches
|
|
|
46 |
* Parameters: Groupsize = 128. No act-order.
|
47 |
* Command used to create the GPTQ:
|
48 |
```
|
49 |
+
python llama.py /workspace/models/openaccess-ai-collective_wizard-mega-13b c4 --wbits 4 --true-sequential --groupsize 128 --save_safetensors /workspace/wizard-mega-13b/gptq/wizard-mega-13B-GPTQ-4bit-128g.no-act-order.safetensors
|
50 |
```
|
51 |
|
52 |
|