issue with vLLM inference
1
#7 opened 4 days ago
by
rohitg

MLX version please?
1
#5 opened 13 days ago
by
Narutoouz

Running into issue when trying to run this model with vllm
1
#4 opened 19 days ago
by
Travisjw25
Are there any quantization models, such as GGUF? Can it run with 16GB of VRAM?
๐
2
4
#2 opened 27 days ago
by
yoolv
Thank you!
๐ค
1
2
#1 opened 27 days ago
by
googlefan
