-
Notifications
You must be signed in to change notification settings - Fork 11.4k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
CUDA error 12 : invalid pitch argument #1388
Comments
~/llama.cpp$ nvcc --version My CUDA information. CUDA error 2 at ggml-cuda.cu:623: out of memory And this is the error message that occurred when I tried to run it in a Linux environment through WSL2. I have 8GB of VRAM. |
I was having the same issue with long prompts, resulting in the same error message, also using cuBLAS. |
Thank you for your response. Fortunately, my issue was resolved automatically after installing the latest version. So I will consider this issue closed. |
CUDA error 12 at C:\Users\USER\llama.cpp\ggml-cuda.cu:527: invalid pitch argument
Until yesterday, the program was running normally, but today I suddenly encountered the above error, causing the model to crash. I am using Windows 10, and I compiled it with CMAKE to use cuBLAS. The program runs without any issues when a short prompt is provided, but it immediately stops working when a long prompt is entered.
I created a virtual environment and performed a clean installation and compilation, but I still encountered the same error. I would greatly appreciate it if you could provide a solution to this issue.
The text was updated successfully, but these errors were encountered: