Update README.md
Browse files
README.md
CHANGED
@@ -24,10 +24,10 @@ NexaAI/Qwen2.5-Omni-3B-GGUF
|
|
24 |
#### Available Quantizations
|
25 |
| Filename | Quant type | File Size | Split | Description |
|
26 |
| -------- | ---------- | --------- | ----- | ----------- |
|
27 |
-
| [Qwen2.5-Omni-3B-4bit.gguf](https://huggingface.co/
|
28 |
-
| [Qwen2.5-Omni-3B-Q8_0.gguf](https://huggingface.co/
|
29 |
-
| [Qwen2.5-Omni-3Bq2_k.gguf](https://huggingface.co/
|
30 |
-
| [mmproj-Qwen2.5-Omni-3B-Q8_0.gguf](https://huggingface.co/
|
31 |
|
32 |
## Overview
|
33 |
|
|
|
24 |
#### Available Quantizations
|
25 |
| Filename | Quant type | File Size | Split | Description |
|
26 |
| -------- | ---------- | --------- | ----- | ----------- |
|
27 |
+
| [Qwen2.5-Omni-3B-4bit.gguf](https://huggingface.co/NexaAI/Qwen2.5-Omni-3B-GGUF/blob/main/Qwen2.5-Omni-3B-4bit.gguf) | 4bit | 2.1 GB | false | Lightweight 4-bit quant for fast inference. |
|
28 |
+
| [Qwen2.5-Omni-3B-Q8_0.gguf](https://huggingface.co/NexaAI/Qwen2.5-Omni-3B-GGUF/blob/main/Qwen2.5-Omni-3B-Q8_0.gguf) | Q8_0 | 3.62 GB | false | High-quality 8-bit quantization. |
|
29 |
+
| [Qwen2.5-Omni-3Bq2_k.gguf](https://huggingface.co/NexaAI/Qwen2.5-Omni-3B-GGUF/blob/main/Qwen2.5-Omni-3Bq2_k.gguf) | Q2_K | 4 Bytes | false | 2-bit quant. Best for extreme low-resource use. |
|
30 |
+
| [mmproj-Qwen2.5-Omni-3B-Q8_0.gguf](https://huggingface.co/NexaAI/Qwen2.5-Omni-3B-GGUF/blob/main/mmproj-Qwen2.5-Omni-3B-Q8_0.gguf) | Q8_0 | 1.54 GB | false | Required vision adapter for Q8_0 model. |
|
31 |
|
32 |
## Overview
|
33 |
|