-
Notifications
You must be signed in to change notification settings - Fork 107
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Llama.cpp fails on Fedora AMD - ROCm error #732
Comments
I think that GPU is gfx1103 . Can you check if the relevant file in in the container in /opt ? (It should have gfx1103 in the filename) |
You could just be simply running out of VRAM, how much VRAM does your GPU have? |
If: llama3.2:1b works, you are likely running out of VRAM, I think by default it's 3b. |
It's an iGPU and seems like it gets 50% of teh RAM I am pulling qwen2.5:1.5b to check a smaller model, but my internet sucks, so it is gonna take a few more minutes....:D |
qwen still fails with the same error Not sure if these are the files you were looking for?
|
I also noticed this
and since I am not specifying |
There is something generally weird happening - I checked BIOS and there was actually only 3GB VRAM assigned, so I bumped it up to 16GB (out of 32GB RAM total in the machine) I can see the VRAM availeble now But it seems both Again, this is my first experience with GPUs in general, so it is all very confusing - feel free to send me somewhere else:D |
Curious if the memory error might be related to this: ollama/ollama#5471 (comment) |
Hi folks:-)
I am not saying this is a Ramalama issue, but I would appreciate your help/guidance, because thi is my first endeavour with local GPUs:-)
I just got my GMKtec K11 machine (https://www.gmktec.com/products/amd-ryzen%E2%84%A2-9-8945hs-nucbox-k11) and installed Fedora 41 on it + podman + ramalama (Installed via the
curl .. | sh
method from README)This is a result of
ramalama run
It successfully finds the GPU (very cool)
but tehn fails when I try to prompt the model
Any ideas/thoughts are appreciated:) Happy to file an issue against llama.cpp, just wanted to make sure I am not missing something obvious (like packages installed or something)
The text was updated successfully, but these errors were encountered: