[Usage]: When I installed version 0.4.1 and started vllm.entrypoints.openai.api_server
with the --engine-use-ray
parameter, I encountered some issues.
#4367
Labels
usage
How to use vllm
Your current environment
How would you like to use vllm
When I start VLLM with the --worker-use-ray parameter using the following command:
python3 -m vllm.entrypoints.openai.api_server --model /input2/ --host 0.0.0.0 --port 8080 --engine-use-ray --worker-use-ray --max-num-seqs 16 --tensor-parallel-size 1 --max-model-len 8192
I encounter the following issues:
These issues do not occur when I start VLLM without the --worker-use-ray parameter or when I downgrade to version 0.3.3.
The text was updated successfully, but these errors were encountered: