Update README.md
Browse files
README.md
CHANGED
@@ -29,7 +29,7 @@ language:
|
|
29 |
- quants done with an importance matrix for improved quantization loss
|
30 |
- K & IQ quants in basically all variants from Q6_K down to IQ1_S
|
31 |
- Quantized with [llama.cpp](https://github.com/ggerganov/llama.cpp) commit [8d608a81b7bd170f700648f8214e6f3279d4d715](https://github.com/ggerganov/llama.cpp/commit/8d608a81b7bd170f700648f8214e6f3279d4d715) (master from 2024-05-01)
|
32 |
-
-
|
33 |
```
|
34 |
./imatrix -c 512 -m $model_name-f16.gguf -f $llama_cpp_path/groups_merged.txt -o $out_path/imat-f16-gmerged.dat
|
35 |
```
|
|
|
29 |
- quants done with an importance matrix for improved quantization loss
|
30 |
- K & IQ quants in basically all variants from Q6_K down to IQ1_S
|
31 |
- Quantized with [llama.cpp](https://github.com/ggerganov/llama.cpp) commit [8d608a81b7bd170f700648f8214e6f3279d4d715](https://github.com/ggerganov/llama.cpp/commit/8d608a81b7bd170f700648f8214e6f3279d4d715) (master from 2024-05-01)
|
32 |
+
- Imatrix generated with [this](https://github.com/ggerganov/llama.cpp/discussions/5263#discussioncomment-8395384) dataset.
|
33 |
```
|
34 |
./imatrix -c 512 -m $model_name-f16.gguf -f $llama_cpp_path/groups_merged.txt -o $out_path/imat-f16-gmerged.dat
|
35 |
```
|