Update README.md
Browse files
README.md
CHANGED
@@ -7,6 +7,8 @@ inference: false
|
|
7 |
|
8 |
Quantised 4bit and 2bit GGMLs of [changsung's alpaca-lora-65B](https://huggingface.co/chansung/alpaca-lora-65b) for CPU inference with [llama.cpp](https://github.com/ggerganov/llama.cpp).
|
9 |
|
|
|
|
|
10 |
## Provided files
|
11 |
| Bits | Size | RAM required | Name |
|
12 |
| ---- | ---- | ---- | ---- |
|
|
|
7 |
|
8 |
Quantised 4bit and 2bit GGMLs of [changsung's alpaca-lora-65B](https://huggingface.co/chansung/alpaca-lora-65b) for CPU inference with [llama.cpp](https://github.com/ggerganov/llama.cpp).
|
9 |
|
10 |
+
I also have 4bit GPTQ files for GPU inference available here: [TheBloke/alpaca-lora-65B-GPTQ-4bit](https://huggingface.co/TheBloke/alpaca-lora-65B-GPTQ-4bit).
|
11 |
+
|
12 |
## Provided files
|
13 |
| Bits | Size | RAM required | Name |
|
14 |
| ---- | ---- | ---- | ---- |
|