Update README.md
Browse files
README.md
CHANGED
|
@@ -24,7 +24,10 @@ nexaml/Qwen2.5-Omni-3B-GGUF
|
|
| 24 |
#### Available Quantizations
|
| 25 |
| Filename | Quant type | File Size | Split | Description |
|
| 26 |
| -------- | ---------- | --------- | ----- | ----------- |
|
| 27 |
-
|
|
|
|
|
|
|
|
|
|
|
| 28 |
|
| 29 |
## Overview
|
| 30 |
|
|
|
|
| 24 |
#### Available Quantizations
|
| 25 |
| Filename | Quant type | File Size | Split | Description |
|
| 26 |
| -------- | ---------- | --------- | ----- | ----------- |
|
| 27 |
+
| [Qwen2.5-Omni-3B-4bit.gguf](https://huggingface.co/nexaml/Qwen2.5-Omni-3B-GGUF/blob/main/Qwen2.5-Omni-3B-4bit.gguf) | 4bit | 2.1 GB | false | Lightweight 4-bit quant for fast inference. |
|
| 28 |
+
| [Qwen2.5-Omni-3B-Q8_0.gguf](https://huggingface.co/nexaml/Qwen2.5-Omni-3B-GGUF/blob/main/Qwen2.5-Omni-3B-Q8_0.gguf) | Q8_0 | 3.62 GB | false | High-quality 8-bit quantization. |
|
| 29 |
+
| [Qwen2.5-Omni-3Bq2_k.gguf](https://huggingface.co/nexaml/Qwen2.5-Omni-3B-GGUF/blob/main/Qwen2.5-Omni-3Bq2_k.gguf) | Q2_K | 4 Bytes | false | 2-bit quant. Best for extreme low-resource use. |
|
| 30 |
+
| [mmproj-Qwen2.5-Omni-3B-Q8_0.gguf](https://huggingface.co/nexaml/Qwen2.5-Omni-3B-GGUF/blob/main/mmproj-Qwen2.5-Omni-3B-Q8_0.gguf) | Q8_0 | 1.54 GB | false | Required vision adapter for Q8_0 model. |
|
| 31 |
|
| 32 |
## Overview
|
| 33 |
|