-
-
-
-
-
-
Inference Providers
Active filters:
GPTQ
RedHatAI/Kimi-K2-Instruct-quantized.w4a16
Text Generation
•
Updated
•
5.82k
•
7
DanielAWrightGabrielAI/pygmalion-7b-4bit-128g-cuda-2048Token
Text Generation
•
Updated
•
13
•
15
mlabonne/gpt2-GPTQ-4bit
Text Generation
•
Updated
•
3
CalderaAI/13B-Ouroboros-GPTQ4bit-128g-CUDA
Text Generation
•
Updated
•
2
daedalus314/Griffin-3B-GPTQ
Text Generation
•
0.7B
•
Updated
•
4
Sanrove/gpt2-GPTQ-4b
Text Generation
•
Updated
•
8
daedalus314/Marx-3B-V2-GPTQ
Text Generation
•
Updated
•
4
TKDKid1000/pythia-2.8b-deduped-GPTQ
Text Generation
•
Updated
•
9
Trelis/Yi-34B-200K-Llamafied-chat-SFT-function-calling-v2-GPTQ
Text Generation
•
Updated
Inferless/deciLM-7B-GPTQ
Text Generation
•
Updated
•
6
•
1
Inferless/SOLAR-10.7B-Instruct-v1.0-GPTQ
Text Generation
•
Updated
•
6
•
2
Inferless/Mixtral-8x7B-v0.1-int8-GPTQ
Text Generation
•
Updated
•
6
•
2
Masterjp123/SnowyRP-FinalV1-L2-13B-GPTQ
Text Generation
•
2B
•
Updated
•
12
•
4
bigquant/Senku-70B-GPTQ-4bit
Text Generation
•
Updated
•
5
•
1
twhoool02/Llama-2-7b-hf-AutoGPTQ
Text Generation
•
1B
•
Updated
•
4
Dmitriy007/rugpt2_gen_news-gptq-4bit
Text Generation
•
0.1B
•
Updated
•
3
SwastikM/Llama-2-7B-Chat-text2code
Text Generation
•
Updated
•
12
•
4
adriabama06/Llama-3.2-1B-Instruct-GPTQ-8bit-128g
Text Generation
•
0.5B
•
Updated
•
5
•
1
NightForger/saiga_nemo_12b-GPTQ
Text Generation
•
Updated
•
30
NaomiBTW/L3-8B-Lunaris-v1-GPTQ
Text Generation
•
Updated
GusPuffy/Llama-3.1-70B-ArliAI-RPMax-v1.3-GPTQ
11B
•
Updated
•
18
AXERA-TECH/DeepSeek-R1-Distill-Qwen-1.5B-GPTQ-Int4
AXERA-TECH/DeepSeek-R1-Distill-Qwen-7B-GPTQ-Int4
AXERA-TECH/Qwen2.5-1.5B-Instruct-GPTQ-Int4
Text Generation
•
Updated
•
3
AXERA-TECH/Qwen2.5-3B-Instruct-GPTQ-Int4
AXERA-TECH/Qwen2.5-0.5B-Instruct-GPTQ-Int4
AXERA-TECH/Qwen2.5-7B-Instruct-GPTQ-Int4
RedHatAI/DeepSeek-R1-quantized.w4a16
Text Generation
•
Updated
•
55
•
7
iqbalamo93/Phi-4-mini-instruct-GPTQ-4bit
Text Generation
•
1B
•
Updated
•
12
iqbalamo93/Phi-4-mini-instruct-GPTQ-8bit
Text Generation
•
1B
•
Updated
•
124k