Model does not run with VLLM
2
#3 opened 11 months ago
by
aswad546
Any idea when the evaluation data will be in for this model? would like to know how the performance differ from unquantized version of the model
#2 opened about 1 year ago
by
jahhs0n
Any chance your team is working on a 4-bit Llama-3.2-90B-Vision-Instruct-quantized.w4a16 version?
2
#1 opened about 1 year ago
by
mrhendrey