Model Summary

This repository hosts quantized versions of the Llama-3.2-3B-Instruct model.

Format: GGUF
Converter: llama.cpp 70392f1f81470607ba3afef04aa56c9f65587664
Quantizer: LM-Kit.NET 2024.8.4

For more detailed information on the base model, please visit the following link

Downloads last month
38
GGUF
Model size
3B params
Architecture
llama
Hardware compatibility
Log In to add your hardware

2-bit

3-bit

4-bit

5-bit

6-bit

8-bit

16-bit

Inference Providers NEW
This model isn't deployed by any Inference Provider. ๐Ÿ™‹ Ask for provider support