Replies: 1 comment
-
It looks like you are using the GPU but the utilization is not great. You can double check with:
Now why it's under-utilized is a harder question to answer. One thing that is very odd is the way your toks/sec are periodic: high, then low. Could you share some details about the dataset? Are there highly variable sequence lengths? |
Beta Was this translation helpful? Give feedback.
0 replies
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
-
Issue: Long Training Times and GPU Utilization Problem
Description:
Hi everyone,
I am currently working on fine-tuning an LLM using the MLX library on my MacBook Pro M1 with 16GB of RAM. While the training process works, it is taking an exceptionally long time to complete. From what I observe, the training is not utilizing the GPU and is instead maxing out the Performance-CPUs at 100%.
Here are the key details of my setup and issue:
Symptoms:
Questions:
Additional Information:
I have verified that the necessary packages and dependencies are installed correctly. Any insights or guidance on how to resolve this issue would be greatly appreciated.
Thanks in advance for your help!
Best regards,
Meir
Beta Was this translation helpful? Give feedback.
All reactions