nexaml commited on
Commit
c54b0b7
·
verified ·
1 Parent(s): 2b800e4

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +4 -1
README.md CHANGED
@@ -24,7 +24,10 @@ nexaml/Qwen2.5-Omni-3B-GGUF
24
  #### Available Quantizations
25
  | Filename | Quant type | File Size | Split | Description |
26
  | -------- | ---------- | --------- | ----- | ----------- |
27
- | | | | | |
 
 
 
28
 
29
  ## Overview
30
 
 
24
  #### Available Quantizations
25
  | Filename | Quant type | File Size | Split | Description |
26
  | -------- | ---------- | --------- | ----- | ----------- |
27
+ | [Qwen2.5-Omni-3B-4bit.gguf](https://huggingface.co/nexaml/Qwen2.5-Omni-3B-GGUF/blob/main/Qwen2.5-Omni-3B-4bit.gguf) | 4bit | 2.1 GB | false | Lightweight 4-bit quant for fast inference. |
28
+ | [Qwen2.5-Omni-3B-Q8_0.gguf](https://huggingface.co/nexaml/Qwen2.5-Omni-3B-GGUF/blob/main/Qwen2.5-Omni-3B-Q8_0.gguf) | Q8_0 | 3.62 GB | false | High-quality 8-bit quantization. |
29
+ | [Qwen2.5-Omni-3Bq2_k.gguf](https://huggingface.co/nexaml/Qwen2.5-Omni-3B-GGUF/blob/main/Qwen2.5-Omni-3Bq2_k.gguf) | Q2_K | 4 Bytes | false | 2-bit quant. Best for extreme low-resource use. |
30
+ | [mmproj-Qwen2.5-Omni-3B-Q8_0.gguf](https://huggingface.co/nexaml/Qwen2.5-Omni-3B-GGUF/blob/main/mmproj-Qwen2.5-Omni-3B-Q8_0.gguf) | Q8_0 | 1.54 GB | false | Required vision adapter for Q8_0 model. |
31
 
32
  ## Overview
33