/v1/chat/completions endpoint not working
#32
by
PremkumarChandak
- opened
We have successfully load and serve model with vllm.
When we are try to communicate with model /chat/completions , we are not getting any response.(continuously loading)
PremkumarChandak
changed discussion title from
/v1C endpoint not working
to /v1/chat/completions endpoint not working
I have a same problem, do you solve problem ?
same error
I was able to generate output using dtype=float32.
i test api v1/chat/completions successfully with vllm. But url images not working. https://cdn-uploads.huggingface.co/production/uploads/66e3abda596fcff3e4d0b06b/eFCNNQ9galifwYZ8ebHAf.jpeg Can you help me ? Thanks.