iandennismiller
commited on
Commit
·
f87e3a7
1
Parent(s):
207bad5
3-bit quant
Browse files- LLama-2-MedText-13b-Q3_K_L.gguf +3 -0
- LLama-2-MedText-13b-Q6_K.gguf +3 -0
- README.md +1 -0
LLama-2-MedText-13b-Q3_K_L.gguf
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:93371f4e513bdffdc3a7cc164068f9b18ce3287326af67675def93abeab8e2fb
|
3 |
+
size 6929559424
|
LLama-2-MedText-13b-Q6_K.gguf
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:34602cc7158d458c82cec54c838e9bdf7691b95ce9256657b6a493b07886f91b
|
3 |
+
size 10679140224
|
README.md
CHANGED
@@ -85,6 +85,7 @@ Then quantize f32 GGUF to lower bit resolutions
|
|
85 |
|
86 |
```bash
|
87 |
llama.cpp/build/bin/quantize LLama-2-MedText-13b-f32.gguf LLama-2-MedText-13b-Q3_K_L.gguf Q3_K_L
|
|
|
88 |
```
|
89 |
|
90 |
### Distributing model through huggingface
|
|
|
85 |
|
86 |
```bash
|
87 |
llama.cpp/build/bin/quantize LLama-2-MedText-13b-f32.gguf LLama-2-MedText-13b-Q3_K_L.gguf Q3_K_L
|
88 |
+
llama.cpp/build/bin/quantize LLama-2-MedText-13b-f32.gguf LLama-2-MedText-13b-Q6_K.gguf Q6_K
|
89 |
```
|
90 |
|
91 |
### Distributing model through huggingface
|