TheBloke commited on
Commit
229b765
1 Parent(s): 2b78e44

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +8 -5
README.md CHANGED
@@ -118,6 +118,8 @@ git clone --single-branch --branch gptq-3bit--1g-actorder_True https://huggingfa
118
  <!-- README_GPTQ.md-text-generation-webui start -->
119
  ## How to easily download and use this model in [text-generation-webui](https://github.com/oobabooga/text-generation-webui).
120
 
 
 
121
  Please make sure you're using the latest version of [text-generation-webui](https://github.com/oobabooga/text-generation-webui).
122
 
123
  It is strongly recommended to use the text-generation-webui one-click-installers unless you're sure you know how to make a manual install.
@@ -128,12 +130,13 @@ It is strongly recommended to use the text-generation-webui one-click-installers
128
  - see Provided Files above for the list of branches for each option.
129
  3. Click **Download**.
130
  4. The model will start downloading. Once it's finished it will say "Done".
131
- 5. In the top left, click the refresh icon next to **Model**.
132
- 6. In the **Model** dropdown, choose the model you just downloaded: `Falcon-180B-Chat-GPTQ`
133
- 7. The model will automatically load, and is now ready for use!
134
- 8. If you want any custom settings, set them and then click **Save settings for this model** followed by **Reload the Model** in the top right.
 
135
  * Note that you do not need to and should not set manual GPTQ parameters any more. These are set automatically from the file `quantize_config.json`.
136
- 9. Once you're ready, click the **Text Generation tab** and enter a prompt to get started!
137
  <!-- README_GPTQ.md-text-generation-webui end -->
138
 
139
  <!-- README_GPTQ.md-use-from-python start -->
 
118
  <!-- README_GPTQ.md-text-generation-webui start -->
119
  ## How to easily download and use this model in [text-generation-webui](https://github.com/oobabooga/text-generation-webui).
120
 
121
+ **NOTE**: I have not tested this model with Text Generation Webui. It *should* work through the Transformers Loader. It will *not* work through the AutoGPTQ loader, due to the files being sharded.
122
+
123
  Please make sure you're using the latest version of [text-generation-webui](https://github.com/oobabooga/text-generation-webui).
124
 
125
  It is strongly recommended to use the text-generation-webui one-click-installers unless you're sure you know how to make a manual install.
 
130
  - see Provided Files above for the list of branches for each option.
131
  3. Click **Download**.
132
  4. The model will start downloading. Once it's finished it will say "Done".
133
+ 5. Choose Loader: Transformers
134
+ 6. In the top left, click the refresh icon next to **Model**.
135
+ 7. In the **Model** dropdown, choose the model you just downloaded: `Falcon-180B-Chat-GPTQ`
136
+ 8. The model will automatically load, and is now ready for use!
137
+ 9. If you want any custom settings, set them and then click **Save settings for this model** followed by **Reload the Model** in the top right.
138
  * Note that you do not need to and should not set manual GPTQ parameters any more. These are set automatically from the file `quantize_config.json`.
139
+ 10. Once you're ready, click the **Text Generation tab** and enter a prompt to get started!
140
  <!-- README_GPTQ.md-text-generation-webui end -->
141
 
142
  <!-- README_GPTQ.md-use-from-python start -->