Running into an issue that was resolved in the Instruct model
#10 opened 7 days ago
by
orr-tzafon

Please Support Pipeline Parallism
2
#9 opened 10 days ago
by
VivekMalipatel23
issue with vLLM inference
1
#7 opened about 1 month ago
by
rohitg

MLX version please?
2
#5 opened about 1 month ago
by
Narutoouz

Running into issue when trying to run this model with vllm
4
#4 opened about 2 months ago
by
Travisjw25
Are there any quantization models, such as GGUF? Can it run with 16GB of VRAM?
๐
2
4
#2 opened about 2 months ago
by
yoolv
Thank you!
๐ค
1
2
#1 opened about 2 months ago
by
googlefan
