Update README.md
Browse files
README.md
CHANGED
|
@@ -7,6 +7,9 @@ license: apache-2.0
|
|
| 7 |
**This repo contains the humanized 360M SmolLM2 model in the GGUF Format**
|
| 8 |
- Quantization: q2_K, q3_K_S, q3_K_M, q3_K_L, q4_0, q4_K_S, q4_K_M, q5_0, q5_K_S, q5_K_M, q6_K, q8_0
|
| 9 |
|
|
|
|
|
|
|
|
|
|
| 10 |
## Quickstart
|
| 11 |
|
| 12 |
We advise you to clone [`llama.cpp`](https://github.com/ggerganov/llama.cpp) and install it following the official guide. We follow the latest version of llama.cpp.
|
|
|
|
| 7 |
**This repo contains the humanized 360M SmolLM2 model in the GGUF Format**
|
| 8 |
- Quantization: q2_K, q3_K_S, q3_K_M, q3_K_L, q4_0, q4_K_S, q4_K_M, q5_0, q5_K_S, q5_K_M, q6_K, q8_0
|
| 9 |
|
| 10 |
+
**More about this model**
|
| 11 |
+
- We released a 135M, 360M and 1.7B parameter version of this model. For more information, view our [report](https://www.assistantslab.com/research/smollm2-report).
|
| 12 |
+
|
| 13 |
## Quickstart
|
| 14 |
|
| 15 |
We advise you to clone [`llama.cpp`](https://github.com/ggerganov/llama.cpp) and install it following the official guide. We follow the latest version of llama.cpp.
|