Hugging Face's logo Hugging Face
  • Models
  • Datasets
  • Spaces
  • Docs
  • Enterprise
  • Pricing

  • Log In
  • Sign Up

RedHatAI
/
Llama-4-Scout-17B-16E-Instruct-FP8-dynamic

Image-Text-to-Text
Safetensors
PyTorch
vllm
llama4
facebook
meta
llama
neuralmagic
redhat
llmcompressor
quantized
FP8
conversational
compressed-tensors
Model card Files Files and versions
xet
Community
8
New discussion
Resources
  • PR & discussions documentation
  • Code of Conduct
  • Hub documentation

πŸͺ„ InferenceService name updated

#8 opened 14 days ago by
ckavili

change-name

#7 opened 18 days ago by
robertgshaw

Overview states 109b, should be 17b

#6 opened 20 days ago by
jcordes

Failing to quantize using your method

#4 opened 3 months ago by
redd2dead

VLLM launch parametrs

πŸ‘ 3
#3 opened 3 months ago by
Clutchkin

Why not FP8 with static and per-tensor quantization?

πŸ‘ 1
2
#2 opened 4 months ago by
wanzhenchn

Thank you uploading this.

❀️ 6
#1 opened 4 months ago by
chriswritescode
Company
TOS Privacy About Jobs
Website
Models Datasets Spaces Pricing Docs