Unable to depploy this model using vLLM

#1
by hdnh2006 - opened

Hi, I am just following the suggestion from huggingface and I am running:

vllm serve "bnb-community/Llama-4-Scout-17B-16E-Instruct-bnb-4bit"

Unfortunately I got this error:

ERROR 04-08 07:32:28 [core.py:390] OSError: bnb-community/Llama-4-Scout-17B-16E-Instruct-bnb-4bit does not appear to have a file named preprocessor_config.json. Checkout 'https://huggingface.co/bnb-community/Llama-4-Scout-17B-16E-Instruct-bnb-4bit/tree/main'for available files.

Any idea?

Bitsandbytes Community org

thanks for reporting ! Yes I updated the processor files on the repo, I think this should solve the problem

got assertion error

ERROR 04-09 13:25:06 [core.py:390]     assert self.quant_method is not None
ERROR 04-09 13:25:06 [core.py:390] AssertionError
Bitsandbytes Community org

Hi @george-arta ,can you share the trace please ?

Your need to confirm your account before you can post a new comment.

Sign up or log in to comment