Issue serving YannQi/R-4B with official vLLM Docker image

#3
by vm7608 - opened

Hello, and thank you for your fantastic work on this model!

I'm encountering an issue when trying to serve the YannQi/R-4B model using the official vllm/vllm-openai Docker image. I've tried both the :latest and :v0.10.1.1 tags, but unfortunately, both result in an error.

I used the following command:

docker run --rm \
    --gpus all \
    --name R-4B \
    -v ~/.cache/huggingface:/root/.cache/huggingface \
    -p 8003:8000 \
    --ipc=host \
    vllm/vllm-openai:v0.10.1.1 \
    --model YannQi/R-4B \
    --gpu-memory-utilization 0.9 \
    --trust-remote-code

The error message:

ValueError: There is no module or parameter named 'model.language_model' in TransformersForMultimodalLM

Could you possibly provide some guidance on how to resolve this? Thanks in advance for your help!

This comment has been hidden (marked as Off-Topic)

Currently, R-4B only supports deployment from vLLM source code. After vLLM is updated to v0.10.2 https://github.com/vllm-project/vllm/releases/tag/v0.10.2rc1 , it can be deployed from the official website or the newest docker.

YannQi changed discussion status to closed

Sign up or log in to comment