Error with vLLM docker image

#10
by mhwang093 - opened

Hi I tried to host this model with vLLM docker, tried different vLLM images versions, like latest, v0.9.0, v0.8.5, but all gave me different errors, examples below

ValueError: Unknown version: v11 in /llm/Magistral-Small-2506/tekken.json. Make sure to use a valid version string: ['v1', 'v2', 'v3', 'v7']

api_server.py: error: argument --load-format: invalid choice: 'mistral' (choose from 'auto', 'pt', 'safetensors', 'npcache', 'dummy', 'tensorizer', 'bitsandbytes')

api_server.py: error: argument --tokenizer-mode: invalid choice: 'mistral' (choose from 'auto', 'slow')

Here is the docker command I used

docker run -d --name Magistral-Small-2506 --gpus "device=0" -p 9080:8000 --ipc=host -v /opt/llm:/llm vllm/vllm-openai:v0.8.5 --model /llm/Magistral-Small-2506 --tokenizer_mode mistral --config_format mistral --load_format mistral --tool-call-parser mistral --enable-auto-tool-choice --tensor-parallel-size 2

Could you kindly confirm the correct version of vLLM image to use? Thanks

Hi,

Currently we recommend installing vLLM to have the latest features as follows:

pip install -U vllm \
    --pre \
    --extra-index-url https://wheels.vllm.ai/nightly

As soon as a new release of vLLM comes out, their released docker image will support mistral-common with tokenizer v11 !

Thank you so much for your quick response, Julien!

Sign up or log in to comment