GGUF
Inference Endpoints
conversational
Michielo commited on
Commit
77cfecf
·
verified ·
1 Parent(s): d6eaec2

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +3 -0
README.md CHANGED
@@ -7,6 +7,9 @@ license: apache-2.0
7
  **This repo contains the humanized 360M SmolLM2 model in the GGUF Format**
8
  - Quantization: q2_K, q3_K_S, q3_K_M, q3_K_L, q4_0, q4_K_S, q4_K_M, q5_0, q5_K_S, q5_K_M, q6_K, q8_0
9
 
 
 
 
10
  ## Quickstart
11
 
12
  We advise you to clone [`llama.cpp`](https://github.com/ggerganov/llama.cpp) and install it following the official guide. We follow the latest version of llama.cpp.
 
7
  **This repo contains the humanized 360M SmolLM2 model in the GGUF Format**
8
  - Quantization: q2_K, q3_K_S, q3_K_M, q3_K_L, q4_0, q4_K_S, q4_K_M, q5_0, q5_K_S, q5_K_M, q6_K, q8_0
9
 
10
+ **More about this model**
11
+ - We released a 135M, 360M and 1.7B parameter version of this model. For more information, view our [report](https://www.assistantslab.com/research/smollm2-report).
12
+
13
  ## Quickstart
14
 
15
  We advise you to clone [`llama.cpp`](https://github.com/ggerganov/llama.cpp) and install it following the official guide. We follow the latest version of llama.cpp.