Update README.md
Browse files
README.md
CHANGED
@@ -24,6 +24,8 @@ pip install auto-gptq
|
|
24 |
Since the original auto-gptq library does not support the quantization of chatglm models, manual import (hack) is required.
|
25 |
```python
|
26 |
from auto_gptq.modeling._base import BaseGPTQForCausalLM
|
|
|
|
|
27 |
from auto_gptq.modeling.auto import GPTQ_CAUSAL_LM_MODEL_MAP
|
28 |
|
29 |
class ChatGLMGPTQForCausalLM(BaseGPTQForCausalLM):
|
@@ -42,6 +44,7 @@ class ChatGLMGPTQForCausalLM(BaseGPTQForCausalLM):
|
|
42 |
]
|
43 |
|
44 |
GPTQ_CAUSAL_LM_MODEL_MAP['chatglm'] = ChatGLMGPTQForCausalLM
|
|
|
45 |
```
|
46 |
|
47 |
The complete model import code is as follows:
|
@@ -61,6 +64,8 @@ from auto_gptq import AutoGPTQForCausalLM
|
|
61 |
|
62 |
|
63 |
from auto_gptq.modeling._base import BaseGPTQForCausalLM
|
|
|
|
|
64 |
from auto_gptq.modeling.auto import GPTQ_CAUSAL_LM_MODEL_MAP
|
65 |
|
66 |
class ChatGLMGPTQForCausalLM(BaseGPTQForCausalLM):
|
@@ -79,6 +84,7 @@ class ChatGLMGPTQForCausalLM(BaseGPTQForCausalLM):
|
|
79 |
]
|
80 |
|
81 |
GPTQ_CAUSAL_LM_MODEL_MAP['chatglm'] = ChatGLMGPTQForCausalLM
|
|
|
82 |
|
83 |
device = 'cuda:0'
|
84 |
quantized_model_dir = 'alexwww94/glm-4v-9b-gptq'
|
|
|
24 |
Since the original auto-gptq library does not support the quantization of chatglm models, manual import (hack) is required.
|
25 |
```python
|
26 |
from auto_gptq.modeling._base import BaseGPTQForCausalLM
|
27 |
+
from auto_gptq.modeling._const import SUPPORTED_MODELS
|
28 |
+
|
29 |
from auto_gptq.modeling.auto import GPTQ_CAUSAL_LM_MODEL_MAP
|
30 |
|
31 |
class ChatGLMGPTQForCausalLM(BaseGPTQForCausalLM):
|
|
|
44 |
]
|
45 |
|
46 |
GPTQ_CAUSAL_LM_MODEL_MAP['chatglm'] = ChatGLMGPTQForCausalLM
|
47 |
+
SUPPORTED_MODELS = SUPPORTED_MODELS.append('chatglm')
|
48 |
```
|
49 |
|
50 |
The complete model import code is as follows:
|
|
|
64 |
|
65 |
|
66 |
from auto_gptq.modeling._base import BaseGPTQForCausalLM
|
67 |
+
from auto_gptq.modeling._const import SUPPORTED_MODELS
|
68 |
+
|
69 |
from auto_gptq.modeling.auto import GPTQ_CAUSAL_LM_MODEL_MAP
|
70 |
|
71 |
class ChatGLMGPTQForCausalLM(BaseGPTQForCausalLM):
|
|
|
84 |
]
|
85 |
|
86 |
GPTQ_CAUSAL_LM_MODEL_MAP['chatglm'] = ChatGLMGPTQForCausalLM
|
87 |
+
SUPPORTED_MODELS = SUPPORTED_MODELS.append('chatglm')
|
88 |
|
89 |
device = 'cuda:0'
|
90 |
quantized_model_dir = 'alexwww94/glm-4v-9b-gptq'
|