Update README.md
Browse files
README.md
CHANGED
@@ -7,6 +7,12 @@ This repository contains the most relevant quantizations of Stability AI's
|
|
7 |
in GGUF format - ready to be used with
|
8 |
[llama.cpp](https://github.com/ggerganov/llama.cpp) and similar applications.
|
9 |
|
|
|
|
|
|
|
|
|
|
|
|
|
10 |
Right now, the following quantizations are available:
|
11 |
|
12 |
* [stablelm-3b-4e1t-Q4_K_M](https://huggingface.co/rozek/StableLM-3B-4E1T_GGUF/blob/main/stablelm-3b-4e1t-Q4_K_M.bin)
|
@@ -17,6 +23,13 @@ Right now, the following quantizations are available:
|
|
17 |
These files are presented here with the written permission of Stability AI (although
|
18 |
access to the model itself is "gated").
|
19 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
20 |
|
21 |
The chosen license is the same as that
|
22 |
of the original model.
|
|
|
7 |
in GGUF format - ready to be used with
|
8 |
[llama.cpp](https://github.com/ggerganov/llama.cpp) and similar applications.
|
9 |
|
10 |
+
> For this model, Stability AI claims: "_StableLM-3B-4E1T achieves
|
11 |
+
> state-of-the-art performance (September 2023) at the 3B parameter scale
|
12 |
+
> for open-source models and is competitive with many of the popular
|
13 |
+
> contemporary 7B models, even outperforming our most recent 7B
|
14 |
+
> StableLM-Base-Alpha-v2._"
|
15 |
+
|
16 |
Right now, the following quantizations are available:
|
17 |
|
18 |
* [stablelm-3b-4e1t-Q4_K_M](https://huggingface.co/rozek/StableLM-3B-4E1T_GGUF/blob/main/stablelm-3b-4e1t-Q4_K_M.bin)
|
|
|
23 |
These files are presented here with the written permission of Stability AI (although
|
24 |
access to the model itself is "gated").
|
25 |
|
26 |
+
Any model details can be found on the
|
27 |
+
[original model card](https://huggingface.co/stabilityai/stablelm-3b-4e1t) and in
|
28 |
+
a paper on [StableLM-3B-4E1T](https://stability.wandb.io/stability-llm/stable-lm/reports/StableLM-3B-4E1T--VmlldzoyMjU4?accessToken=u3zujipenkx5g7rtcj9qojjgxpconyjktjkli2po09nffrffdhhchq045vp0wyfo).
|
29 |
+
The most important ones are
|
30 |
+
|
31 |
+
* context length is 4096
|
32 |
+
|
33 |
|
34 |
The chosen license is the same as that
|
35 |
of the original model.
|