mradermacher
commited on
Commit
•
8a60569
1
Parent(s):
8115011
auto-patch README.md
Browse files
README.md
CHANGED
@@ -28,6 +28,7 @@ static quants of https://huggingface.co/aaditya/OpenBioLLM-Llama3-8B
|
|
28 |
You should use `--override-kv tokenizer.ggml.pre=str:llama3` and a current llama.cpp version to work around a bug in llama.cpp that made these quants.
|
29 |
|
30 |
<!-- provided-files -->
|
|
|
31 |
## Usage
|
32 |
|
33 |
If you are unsure how to use GGUF files, refer to one of [TheBloke's
|
|
|
28 |
You should use `--override-kv tokenizer.ggml.pre=str:llama3` and a current llama.cpp version to work around a bug in llama.cpp that made these quants.
|
29 |
|
30 |
<!-- provided-files -->
|
31 |
+
weighted/imatrix quants seem not to be available (by me) at this time. If they do not show up a week or so after the static ones, I have probably not planned for them. Feel free to request them by opening a Community Discussion.
|
32 |
## Usage
|
33 |
|
34 |
If you are unsure how to use GGUF files, refer to one of [TheBloke's
|