Commit
·
d9ea9e2
1
Parent(s):
4ae68f2
Update README
Browse files- Fix URL to the model
- List acrastt & Bohan Du
README.md
CHANGED
|
@@ -9,7 +9,7 @@ pipeline_tag: text-generation
|
|
| 9 |
---
|
| 10 |
|
| 11 |
# Puma 3B - GGUF
|
| 12 |
-
- Model creator: [Bohan Du](https://huggingface.co/acrastt)
|
| 13 |
- Original model: [Puma 3B](https://huggingface.co/acrastt/puma-3b)
|
| 14 |
|
| 15 |
<!-- description start -->
|
|
@@ -75,9 +75,9 @@ For other parameters and how to use them, please refer to [the llama.cpp documen
|
|
| 75 |
|
| 76 |
| Name | Quant method | Bits | Size | Max RAM required | Use case |
|
| 77 |
| ---- | ---- | ---- | ---- | ---- | ----- |
|
| 78 |
-
| [puma-3b.q4_1.gguf](https://huggingface.co/
|
| 79 |
|
| 80 |
## Thanks
|
| 81 |
|
| 82 |
-
- to [Bohan Du](https://huggingface.co/acrastt) for the Puma model
|
| 83 |
- to [TheBloke](https://huggingface.co/TheBloke) for all the quantized models and this model card template
|
|
|
|
| 9 |
---
|
| 10 |
|
| 11 |
# Puma 3B - GGUF
|
| 12 |
+
- Model creator: [Bohan Du / acrastt](https://huggingface.co/acrastt)
|
| 13 |
- Original model: [Puma 3B](https://huggingface.co/acrastt/puma-3b)
|
| 14 |
|
| 15 |
<!-- description start -->
|
|
|
|
| 75 |
|
| 76 |
| Name | Quant method | Bits | Size | Max RAM required | Use case |
|
| 77 |
| ---- | ---- | ---- | ---- | ---- | ----- |
|
| 78 |
+
| [puma-3b.q4_1.gguf](https://huggingface.co/alexedelsburg/Puma-3b-GGUF/blob/main/puma-3b.q4_1.gguf) | q4_1 | 4 | 2.14 GB| 4.64 GB | Original quant method, 4-bit. Higher accuracy than q4_0 but not as high as q5_0. However has quicker inference than q5 models. |
|
| 79 |
|
| 80 |
## Thanks
|
| 81 |
|
| 82 |
+
- to [Bohan Du / acrastt](https://huggingface.co/acrastt) for the Puma model
|
| 83 |
- to [TheBloke](https://huggingface.co/TheBloke) for all the quantized models and this model card template
|