Update README.md
Browse files
README.md
CHANGED
@@ -20,6 +20,8 @@ This is GGML format quantised 4-bit, 5-bit and 8-bit [MosaicML's MPT-7B](https:/
|
|
20 |
|
21 |
This repo is the result of converting to GGML and quantising.
|
22 |
|
|
|
|
|
23 |
## Repositories available
|
24 |
|
25 |
* [MPT-7B: 4-bit, 5-bit and 8-bit GGML models for CPU (+CUDA) inference](https://huggingface.co/TheBloke/MPT-7B-GGML).
|
|
|
20 |
|
21 |
This repo is the result of converting to GGML and quantising.
|
22 |
|
23 |
+
Please note that these MPT GGMLs are **not compatbile with llama.cpp**. Right now they can only be used in a couple of tools. This is bound to improve in the near future.
|
24 |
+
|
25 |
## Repositories available
|
26 |
|
27 |
* [MPT-7B: 4-bit, 5-bit and 8-bit GGML models for CPU (+CUDA) inference](https://huggingface.co/TheBloke/MPT-7B-GGML).
|