-
Hi, thank you for developing llamafile, it's such a wonderful tool. For some time now, llama.cpp on Linux has had support for unified memory architecture (UMA for AMD APU) to share main memory between the CPU and integrated GPU. This requires compiling llama.cpp with I'm trying to compile llamafile with this additional setting for the llama.cpp, but I'm having some problems. Could you point me in the right direction? I'm using Ubuntu 22.04 with AMD 5600G APU and ROCm 6.1. When I compile llama.cpp I use I tried adding
And in
Then I compiled llamafile:
Unfortunately, when I launched the same model that I use with llama.cpp the
When I run llama.cpp compiled with UMA support, everything works fine:
Should I add |
Beta Was this translation helpful? Give feedback.
Replies: 1 comment 1 reply
-
Did not see this question... create a issuse #439 for that... But before make a PR need to know what the best to activate it. |
Beta Was this translation helpful? Give feedback.
some test #468
(always use UMA if not enough VRAM)