-
Notifications
You must be signed in to change notification settings - Fork 1.3k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Frequent Crashes with Ollama on ARC770 and i7-14700K #12555
Comments
Similar to #12550, you may try the latest version of ollama with |
Hi @sirlegendary, could you please provide us more information with the following script: #12550 (comment) |
We have re-checked and refactored the device decision logic based on your error log, it would be nice if you can try this tomorrow and we can see if it solves your problem. |
|
Hi @sirlegendary , you may try our latest ollama version via |
Description
Ollama keeps crashing on my system when I attempt to run models or after interacting with them for 1-2 prompts. Even using q4_K_M, which I understand is designed to be resource-efficient, the crashes persist.
Examples of Crashing Models:
llama3.2:3b-instruct-q4_K_M
: Crashes after 1-2 prompts.qwen2.5-coder:3b-instruct-q4_K_M
: Also crashes after a few prompts.Models That Do Not Crash:
gemma:2b-instruct-q5_K_M
: Runs without issues but is not suitable for my needs.Additionally, I am unable to run
sycl-ls
to troubleshoot further. When I attempt it, I receive:Environment
Steps to Reproduce
podman exec -it ollama-intel-gpu /bin/bash
llama3.2:3b-instruct-q4_K_M
orqwen2.5-coder:3b-instruct-q4_K_M
.Logs
ollama_intel_gpu_logs.txt
ollama_intel_gpu_2_logs.txt
The text was updated successfully, but these errors were encountered: