-
Notifications
You must be signed in to change notification settings - Fork 77
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
[Bug]: VLMs shutting down MQLLMEngine with more than 1 concurrent requests #841
Comments
Hi @akarX23, thank you for the ticket. Vllm version from your collect_env output seems to not reflect release 1.19 ("vLLM Version: 0.6.3.dev2006+gc3ce6ae77" and should be something like "0.6.4...").
https://github.com/HabanaAI/vllm-fork/releases/tag/v0.6.6.post1%2BGaudi-1.20.0 |
Thank you for directing me to the latest stable branch. I installed vLLM from that branch and reran the steps. For concurrency 1, everything works well. However, for concurrency 2, I am still getting the same error: Output of collect_env:
|
For llava-hf/llava-v1.6-mistral-7b-hf, I start the vllm server using this command:
The error with 2 concurrent requests is different than what I get with llama:
|
Your current environment
The output of `python collect_env.py`
🐛 Describe the bug
When I try to use VLMs like llava-hf/llava-v1.6-mistral-7b-hf or Llama-3.2-11B-Vision-Instruct, the vLLM server starts without issues. Command used:
Now I have a Python script that sends requests to this server using images in a particular folder. I can specify the total number of requests that need to be completed and the concurrency with which requests are sent.
When I specify concurrency as 1, everything seems to run fine and the vLLM server output looks like this:
However, when I make the concurrency to 2, first few requests go through and then the MQLLMEngine crashes, giving this output:
The traceback looks like this in the vLLM server:
This concurrency issue didn't exist last month. I had tested this around 24th Jan 2025, and everything worked that time. This same issue is seen with LLava as well. Please assist, thank you!
I am on the v1.20.0 branch currently.
Before submitting a new issue...
The text was updated successfully, but these errors were encountered: