Update README.md
Browse files
README.md
CHANGED
@@ -23,7 +23,7 @@ tags:
|
|
23 |
- Requires latest llama.cpp master;
|
24 |
- quants done with an importance matrix for improved quantization loss
|
25 |
- gguf & imatrix generated from bf16 for "optimal" accuracy loss (some say this is snake oil, but it can't hurt)
|
26 |
-
-
|
27 |
- Quantized with [llama.cpp](https://github.com/ggerganov/llama.cpp) commit [201cc11afa0a1950e1f632390b2ac6c937a0d8f0](https://github.com/ggerganov/llama.cpp/commit/201cc11afa0a1950e1f632390b2ac6c937a0d8f0)
|
28 |
- Imatrix generated with [this](https://github.com/ggerganov/llama.cpp/discussions/5263#discussioncomment-8395384) multi-purpose dataset.
|
29 |
```
|
|
|
23 |
- Requires latest llama.cpp master;
|
24 |
- quants done with an importance matrix for improved quantization loss
|
25 |
- gguf & imatrix generated from bf16 for "optimal" accuracy loss (some say this is snake oil, but it can't hurt)
|
26 |
+
- Wide coverage of different gguf quant types from Q\_8\_0 down to IQ1\_S
|
27 |
- Quantized with [llama.cpp](https://github.com/ggerganov/llama.cpp) commit [201cc11afa0a1950e1f632390b2ac6c937a0d8f0](https://github.com/ggerganov/llama.cpp/commit/201cc11afa0a1950e1f632390b2ac6c937a0d8f0)
|
28 |
- Imatrix generated with [this](https://github.com/ggerganov/llama.cpp/discussions/5263#discussioncomment-8395384) multi-purpose dataset.
|
29 |
```
|