-
Notifications
You must be signed in to change notification settings - Fork 537
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Problem running jetson-containers run $(autotag ollama) command on Jetson AGX Orin 32GB #814
Comments
I have the same issue on my jetson orin nano when running:
or
The container is starting but it seems like that the ollama service itself is not starting correctly. If I attach the shell and run inside the container |
same here, but also looks like ollama models like llama3.2 is running on CPUs and not on GPUs |
I also recognized that the model inference is really slow. Can you reopen this issue? |
sorry @kalustian @makoit , my guess is this is related to the build system changes in ollama from this week, was trying to fix this in 51e5449 My guess is @tokk-nv had been looking at these patches and can take another look this week |
EDIT: try overriding |
One thing I saw is that jetson-containers really wants to be cloned to the root of your SSD. Once I figured that out, things went smoother. (Not sure if this is your issue, but I just sorted it out on my new Nano, so I thought I'd throw it out.) |
Seeing the same on Orin Nano Super:
|
it is not problem from jetson-containers. Ollama is broken in all plattforms. |
I tried out different things on my nvidia jetson nano (8GB): 1. Using jetson-containers cli:
-> container starts but in container if using ollama list I see this message: Error: could not connect to ollama app, is it running?
-> container is not starting and following error occurs:
-> this is working and starts the container and ollama
-> this is working and starts the container, ollama and runs a model 2. Running the container with docker:
-> this is not working container starts and crashes
-> this is working, starts the container and the ollama service
-> this is working, starts the container and ollama service and runs a model
** -> this is working, it starts the container, ollama and runs the model Summary: If using a bash command when running a container or a service in compose you are able to run ollama (and optional directly a model). But I'm not shure if this is best practise or only a workaround. Comment: It would be nice if we could run a container and ollama starts automatically as service and we could optional pass model_names which are run and also embedding_model_names which automatically are pulled. |
I'm having the same issue, (Jetson ORIN NX 16GB) but it just started yesterday. The problem, that I was able to solve was, create the logs directory inside the data directory that the docker command mounts in my home.
After that, everything run smoothly . |
Can you try with last commit in master? |
This commit did not solve the issue as I reported initially, well at list on my AGX JP6.2. I fresh installed git clone https://github.com/dusty-nv/jetson-containers. Here is a screenshot: ![]() |
If you create the logs directory inside the data directory that's on your home, it doesn't work ? |
well...it in a way works...but, let me ask ask a question...when running an inference using any model...did you see GPU utilization ? like looking into jtop at the same time doing inference. For me, when inferencing, all workload is going into the CPU and not GPU. |
Did you do? CUDA_VERSION=12.6 jetson-containers build ollama jetson-containers run (name that generate) |
Yes, it works. If it's not working, you probably need to install nvidia-container-toolkit in the Jetson (not the container) . |
@johnnynunez - The logs directory has been created and fixed. No need to create the "logs" folder manually, but still not working for me yet as inference is being loaded into the CPU. Let me reflash my AGX again with JP 6.2. |
do you modify docker json ? |
Yes I did and I double checked is running as "nvidia". Let me reflash and start over. |
@johnnynunez , hi Johnny, I have flashed my AGX ( jp6.2)
![]() |
@kalustian Experiencing the same, and also with DEBUG logs:
/etc/docker/daemon.json is set to nvidia. |
Same here, reverted to to |
@kalustian why did you closed the issue? The core issue is not fixed!? |
@dusty-nv @johnnynunez is there any update? I have seen that you pushed changes on docker hub on r36.4.0. I pulled it and my workarounds to run ollama and load the model is not working anymore. Would be nice if there is a stable tag which I can use otherwise each change will break everything again. |
It’s fixed. Only compile docker by yourself. |
Amazing news! What do you mean with compile docker myself? You mean building the image locally myself? |
On point 2, you can try CUDA_VERSION=12.8, now it's in master |
YES !! ..Now is working as expected for me, now I see GPU been used instead CPU as previously. I used CUDA_VERSION=12.6 jetson-containers build ollama. Thanks a lot for your help. |
@kalustian you can test cuda 12.8. I upload for you https://hub.docker.com/r/johnnync/ollama/tags |
Ok here is my overall test. Again, this is on my AGX Orin 32Gb with JP 6.2:
|
If I may suggest to updated the https://github.com/dusty-nv/jetson-containers documentation to include CUDA_VERISON=12.6 jetson-containers build ollama. My 2 Cents. |
Amazing, thanks a lot for your effort! I tested it also on my jetson nano (8GB) and this worked also for me. Was able to build the image with cuda 12.6. The container started and also ollama service was running automatically: Do you also plan to release the image on docker hub with a kind of stable release tag? In my case I have 15 jetson devices and do not want to build on each on them the image manually. |
Yeah for sure! @dusty-nv |
recently I decided to flash agan my AGX orin , and reinstall ollama with above commands. This time when using jetson-containers run $(autotag ollama) I got:
when trying to run any command like "ollama ls "... or "ollama run llama3.2"...etc, I got:
I never had this issue before...any idea what the issue is ?
NOTE: I have Jeston 6.2
The text was updated successfully, but these errors were encountered: