-
Notifications
You must be signed in to change notification settings - Fork 901
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
ggml_new_tensor_impl: not enough space in the context's memory pool (needed 1178743392, available 1106773248) #130
Comments
Ya, I'm seeing a repeatable crash similar to this after about 10 prompts complete. I'm using the 30B model with default parameters. This is on a PC with 128GB of RAM. So it should not be having out of memory issues. Is there param that increases the available memory pool? |
I have the same problem. This has been addressed in the Llama model but no patch is available yet (read the last comment from @ggerganov) |
This should be resolved by ggml-org#626 . |
Here is his fix in the llama.cpp code. Looks easy enough of a fix for alpaca.cpp. I believe the code bases are still similar enough. |
Agreed, should be rather easy to merge, I have no working |
I just tried merging these specific changed and it created an assert in another part of the code when the model loaded. |
Hello, I'm trying to run this on Windows 11, and after fixing the cmake errors with #106 by adding the definitions, when I try to load the 30B model I get a lot of title errors followed by a Segmentation fault:
I have 64Gb of memory and 53Gb are free when I try, so I'm guessing it should be enough? I see the same issue popping for someone in llama ggml-org#153 so it might be related.
I have also tested with 13B model and it won't load either:
So I guess it's not a matter of me lacking memory:
The text was updated successfully, but these errors were encountered: