Update README.md
Browse files
README.md
CHANGED
@@ -118,6 +118,8 @@ git clone --single-branch --branch gptq-3bit--1g-actorder_True https://huggingfa
|
|
118 |
<!-- README_GPTQ.md-text-generation-webui start -->
|
119 |
## How to easily download and use this model in [text-generation-webui](https://github.com/oobabooga/text-generation-webui).
|
120 |
|
|
|
|
|
121 |
Please make sure you're using the latest version of [text-generation-webui](https://github.com/oobabooga/text-generation-webui).
|
122 |
|
123 |
It is strongly recommended to use the text-generation-webui one-click-installers unless you're sure you know how to make a manual install.
|
@@ -128,12 +130,13 @@ It is strongly recommended to use the text-generation-webui one-click-installers
|
|
128 |
- see Provided Files above for the list of branches for each option.
|
129 |
3. Click **Download**.
|
130 |
4. The model will start downloading. Once it's finished it will say "Done".
|
131 |
-
5.
|
132 |
-
6. In the
|
133 |
-
7.
|
134 |
-
8.
|
|
|
135 |
* Note that you do not need to and should not set manual GPTQ parameters any more. These are set automatically from the file `quantize_config.json`.
|
136 |
-
|
137 |
<!-- README_GPTQ.md-text-generation-webui end -->
|
138 |
|
139 |
<!-- README_GPTQ.md-use-from-python start -->
|
|
|
118 |
<!-- README_GPTQ.md-text-generation-webui start -->
|
119 |
## How to easily download and use this model in [text-generation-webui](https://github.com/oobabooga/text-generation-webui).
|
120 |
|
121 |
+
**NOTE**: I have not tested this model with Text Generation Webui. It *should* work through the Transformers Loader. It will *not* work through the AutoGPTQ loader, due to the files being sharded.
|
122 |
+
|
123 |
Please make sure you're using the latest version of [text-generation-webui](https://github.com/oobabooga/text-generation-webui).
|
124 |
|
125 |
It is strongly recommended to use the text-generation-webui one-click-installers unless you're sure you know how to make a manual install.
|
|
|
130 |
- see Provided Files above for the list of branches for each option.
|
131 |
3. Click **Download**.
|
132 |
4. The model will start downloading. Once it's finished it will say "Done".
|
133 |
+
5. Choose Loader: Transformers
|
134 |
+
6. In the top left, click the refresh icon next to **Model**.
|
135 |
+
7. In the **Model** dropdown, choose the model you just downloaded: `Falcon-180B-Chat-GPTQ`
|
136 |
+
8. The model will automatically load, and is now ready for use!
|
137 |
+
9. If you want any custom settings, set them and then click **Save settings for this model** followed by **Reload the Model** in the top right.
|
138 |
* Note that you do not need to and should not set manual GPTQ parameters any more. These are set automatically from the file `quantize_config.json`.
|
139 |
+
10. Once you're ready, click the **Text Generation tab** and enter a prompt to get started!
|
140 |
<!-- README_GPTQ.md-text-generation-webui end -->
|
141 |
|
142 |
<!-- README_GPTQ.md-use-from-python start -->
|