-
Notifications
You must be signed in to change notification settings - Fork 1.1k
After installing with CMAKE_ARGS="-DLLAMA_OPENBLAS=on" FORCE_CMAKE=1, BLAS = 0 on model load #357
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Comments
I have the same issue after upgrading to llama-cpp-python-0.1.62. Previous (llama-cpp-python-0.1.61):
with llama-cpp-python-0.1.62 version:
maybe something upstream with llamacpp if the python bindings pickup latest from llamacpp? |
ok maybe my issue was different, but I didn't set the env settings correctly for powershell (windows). this works:
and this doesn't:
|
It is a different issue. CUBLAS flag works, OPENBLAS does not seem to work. |
I can confirm this issue as well |
I tested and confirmed the
|
Just however confirmed that:
does work, but:
does not. |
May I add, I guess it's ok to have linux-only instructions on a cross-platform project, but at least say so. In case anyone is interested, on windows I solved this by doing a recursive checkout of the repo and then having a cmd file that contains: set CMAKE_ARGS=-DLLAMA_CUBLAS=on Doing pip uninstall llama-cpp-python multiple times before running that also helped in the past. For the record, my system has all the dev stuff installed that could be needed, I am not saying that is all one needs to do. |
I tried this (I'm on windows as well) and was having some difficulty figuring out what they where even referring to when talking about their Environment Variable. I went digging and ended up find a file called CMakeLists.Txt from ggerganov's repo and on line 70 changed
I then completely re-installed llama-cpp-python and I've been able to get it to use the GPU. |
Could be related to ggml-org/llama.cpp#1830 in which case should be fixed shortly. |
I wrote the above issue; I think flags in llama-cpp-python are not correct - trying to find time to make PR for llama-cpp-python. |
Closing. Please reopen if the problem is reproducible with the latest |
Expected Behavior
CMAKE_ARGS="-DLLAMA_OPENBLAS=on" FORCE_CMAKE=1 pip install llama-cpp-python --force-reinstall --upgrade --no-cache-dir
I may be misunderstanding the status output but after making sure that
OpenBLAS
is installed on my system and testing the build with llama.cpp I would expect to see in the instructions/architecture used this after the model has loadedBLAS = 1
Current Behavior
BLAS = 0
Environment and Context
$ lscpu
AMD Ryzen 9 3900XT 12-Core Processor$ uname -a
DESKTOP-1TO72R9 5.15.68.1-microsoft-standard-WSL2+ #2 SMPOpenBLAS built from source and installed in default paths
llama.cpp built with OpenBLAS and tested
Example environment info:
The text was updated successfully, but these errors were encountered: