Issue serving YannQi/R-4B with official vLLM Docker image
#3
by
vm7608
- opened
Hello, and thank you for your fantastic work on this model!
I'm encountering an issue when trying to serve the YannQi/R-4B model using the official vllm/vllm-openai
Docker image. I've tried both the :latest
and :v0.10.1.1
tags, but unfortunately, both result in an error.
I used the following command:
docker run --rm \
--gpus all \
--name R-4B \
-v ~/.cache/huggingface:/root/.cache/huggingface \
-p 8003:8000 \
--ipc=host \
vllm/vllm-openai:v0.10.1.1 \
--model YannQi/R-4B \
--gpu-memory-utilization 0.9 \
--trust-remote-code
The error message:
ValueError: There is no module or parameter named 'model.language_model' in TransformersForMultimodalLM
Could you possibly provide some guidance on how to resolve this? Thanks in advance for your help!
This comment has been hidden (marked as Off-Topic)
Currently, R-4B only supports deployment from vLLM source code. After vLLM is updated to v0.10.2 https://github.com/vllm-project/vllm/releases/tag/v0.10.2rc1 , it can be deployed from the official website or the newest docker.
YannQi
changed discussion status to
closed