You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
When loading the mistralai/Mixtral-8x22B-Instruct-v0.1 model using the LLM function, I keep running into a FileNotFound error where one VllmWorkerProcess appears to try to access a temporary triton cache file of another VllmWorkerProcess, which doesn't exist. My code ran perfectly fine the first time I ran it with the same LLM, but after I killed the first run and tried to run the same code again, I've been having this error. I was able to load another language model meta-llama/Meta-Llama-3-70B-Instruct using the same code. I've tried clearing triton and python caches, but the error persisted. I wasn't able to find any information about this error online, so I'm reaching out for help. I'd appreciate it very much if anyone could share any ideas for fixing this issue. I've attached the full log below.
Thanks,
Jing-Jing Li
🐛 Describe the bug
transformers counterpart, which finished running without an error:
Thanks for the pointer! This does look like a related solution, but it doesn't seem to apply to my setup because I'm not using vLLM from the docker image. However, through the issues linked in this post, I found out that someone worked around this bug by building triton from the source (since this appeared to be a triton bug and they recently pushed code to resolve it). Upgrading to the latest nightly release of triton solved my issue.
Your current environment
Hi,
When loading the mistralai/Mixtral-8x22B-Instruct-v0.1 model using the LLM function, I keep running into a FileNotFound error where one VllmWorkerProcess appears to try to access a temporary triton cache file of another VllmWorkerProcess, which doesn't exist. My code ran perfectly fine the first time I ran it with the same LLM, but after I killed the first run and tried to run the same code again, I've been having this error. I was able to load another language model meta-llama/Meta-Llama-3-70B-Instruct using the same code. I've tried clearing triton and python caches, but the error persisted. I wasn't able to find any information about this error online, so I'm reaching out for help. I'd appreciate it very much if anyone could share any ideas for fixing this issue. I've attached the full log below.
Thanks,
Jing-Jing Li
🐛 Describe the bug
transformers counterpart, which finished running without an error:
Minimal code that reproduces the error:
Log:
The text was updated successfully, but these errors were encountered: