forked from vllm-project/vllm
-
Notifications
You must be signed in to change notification settings - Fork 1
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
[Bug]: Llama-3.2-11B-Vision-Instruct: block table already exists #55
Labels
bug
Something isn't working
Comments
I also see hangs during prefill sometimes with the same workload on N300, see below I bumped the VLLM_RPC_TIMEOUT to 15 minutes to be sure:
|
Occurred again on https://github.com/tenstorrent/vllm/tree/b9564bf364e95a3850619fc7b2ed968cc71e30b7
|
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Your current environment
Model: Llama-3.2-11B-Vision-Instruct
TT device: N300
MESH_DEVICE=N300
Docker image: ghcr.io/tenstorrent/tt-inference-server/tt-metal-llama3-70b-src-base-vllm-ubuntu-22.04-amd64:v0.0.1-47fb1a2fb6e0-2f33504bad49
tt-metal branch: main (last verified commit: 47fb1a2)
vLLM branch: dev (last verified commit: 2f33504)
Model Input Dumps
No response
🐛 Describe the bug
I got this error a few times when running Llama-3.2-11B-Vision-Instruct in vLLM with successive batches of requests.
repro script: https://github.com/tenstorrent/tt-inference-server/blob/tstesco/dev/utils/prompt_client_cli.py
vLLM logs
Before submitting a new issue...
The text was updated successfully, but these errors were encountered: