Update README.md
Browse files
README.md
CHANGED
@@ -24,13 +24,10 @@ base_model: THUDM/glm-4v-9b
|
|
24 |
## Usage
|
25 |
This model is quantized using [AutoGPTQ](https://github.com/AutoGPTQ/AutoGPTQ) for [THUDM/glm-4v-9b](https://huggingface.co/THUDM/glm-4v-9b).
|
26 |
|
27 |
-
|
28 |
|
29 |
(The quantization script will be released later)
|
30 |
|
31 |
-
```bash
|
32 |
-
pip install auto-gptq
|
33 |
-
```
|
34 |
|
35 |
Since the original auto-gptq library does not support the quantization of chatglm models, manual import (hack) is required.
|
36 |
```python
|
|
|
24 |
## Usage
|
25 |
This model is quantized using [AutoGPTQ](https://github.com/AutoGPTQ/AutoGPTQ) for [THUDM/glm-4v-9b](https://huggingface.co/THUDM/glm-4v-9b).
|
26 |
|
27 |
+
It is recommended to install AutoGPTQ by compiling from the source code.
|
28 |
|
29 |
(The quantization script will be released later)
|
30 |
|
|
|
|
|
|
|
31 |
|
32 |
Since the original auto-gptq library does not support the quantization of chatglm models, manual import (hack) is required.
|
33 |
```python
|