Error loading meta-llama/Llama-4-Maverick-17B-128E-Instruct-FP8: Files not found (OSError)
#27 opened about 1 month ago
by
hiroyoshi1644

get_Llama4_Maverick_17B_FP8
#25 opened about 1 month ago
by
duckingsimsen
Quantizer: Running into an error with quantization "TypeError: 'dict' object is not callable"
4
#24 opened 2 months ago
by
AaronVogler

Support for FP8 + Fused MoE layers in vLLM?
2
#23 opened 2 months ago
by
szlevi
is it w8a16 or w8a8?
👍
➕
1
#19 opened 2 months ago
by
ehartford

[request for feedback] faster downloads with xet
#18 opened 2 months ago
by
clem
