Update README.md
Browse files
README.md
CHANGED
@@ -2,4 +2,15 @@
|
|
2 |
license: openrail
|
3 |
model_creator: axiong
|
4 |
model_name: PMC_LLaMA_13B
|
5 |
-
---
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
2 |
license: openrail
|
3 |
model_creator: axiong
|
4 |
model_name: PMC_LLaMA_13B
|
5 |
+
---
|
6 |
+
# PMC-LLaMA-13B - AWQ
|
7 |
+
- Model creator: [axiong](https://huggingface.co/axiong)
|
8 |
+
- Original model: [PMC_LLaMA_13B](https://huggingface.co/axiong/PMC_LLaMA_13B)
|
9 |
+
|
10 |
+
## Description
|
11 |
+
|
12 |
+
This repo contains AWQ model files for [PMC_LLaMA_13B](https://huggingface.co/axiong/PMC_LLaMA_13B).
|
13 |
+
|
14 |
+
### About AWQ
|
15 |
+
|
16 |
+
AWQ is an efficient, accurate and blazing-fast low-bit weight quantization method, currently supporting 4-bit quantization. Compared to GPTQ, it offers faster Transformers-based inference with equivalent or better quality compared to the most commonly used GPTQ settings.
|