rozek commited on
Commit
132eb31
·
1 Parent(s): 282d944

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +13 -0
README.md CHANGED
@@ -7,6 +7,12 @@ This repository contains the most relevant quantizations of Stability AI's
7
  in GGUF format - ready to be used with
8
  [llama.cpp](https://github.com/ggerganov/llama.cpp) and similar applications.
9
 
 
 
 
 
 
 
10
  Right now, the following quantizations are available:
11
 
12
  * [stablelm-3b-4e1t-Q4_K_M](https://huggingface.co/rozek/StableLM-3B-4E1T_GGUF/blob/main/stablelm-3b-4e1t-Q4_K_M.bin)
@@ -17,6 +23,13 @@ Right now, the following quantizations are available:
17
  These files are presented here with the written permission of Stability AI (although
18
  access to the model itself is "gated").
19
 
 
 
 
 
 
 
 
20
 
21
  The chosen license is the same as that
22
  of the original model.
 
7
  in GGUF format - ready to be used with
8
  [llama.cpp](https://github.com/ggerganov/llama.cpp) and similar applications.
9
 
10
+ > For this model, Stability AI claims: "_StableLM-3B-4E1T achieves
11
+ > state-of-the-art performance (September 2023) at the 3B parameter scale
12
+ > for open-source models and is competitive with many of the popular
13
+ > contemporary 7B models, even outperforming our most recent 7B
14
+ > StableLM-Base-Alpha-v2._"
15
+
16
  Right now, the following quantizations are available:
17
 
18
  * [stablelm-3b-4e1t-Q4_K_M](https://huggingface.co/rozek/StableLM-3B-4E1T_GGUF/blob/main/stablelm-3b-4e1t-Q4_K_M.bin)
 
23
  These files are presented here with the written permission of Stability AI (although
24
  access to the model itself is "gated").
25
 
26
+ Any model details can be found on the
27
+ [original model card](https://huggingface.co/stabilityai/stablelm-3b-4e1t) and in
28
+ a paper on [StableLM-3B-4E1T](https://stability.wandb.io/stability-llm/stable-lm/reports/StableLM-3B-4E1T--VmlldzoyMjU4?accessToken=u3zujipenkx5g7rtcj9qojjgxpconyjktjkli2po09nffrffdhhchq045vp0wyfo).
29
+ The most important ones are
30
+
31
+ * context length is 4096
32
+
33
 
34
  The chosen license is the same as that
35
  of the original model.