-
Notifications
You must be signed in to change notification settings - Fork 11.2k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Not having enough memory just causes a segfault or something #257
Comments
Performing the quantization step from f16 down to q4_0 significantly helps with the memory usage (I am dumb but eventually figured this out even though I was supposed to do it anyway) It does make the model extra drunk though (which I guess is to be expected from the 7B version) |
I was wondering why the q4_0 was giving such a wierd responses |
q4_0 will say 2 + 2 = 10, then proceeds to explain how that's because they are in their "respective ranges" of "-5 through 8" by not being "greater than 9 or lesser than -4", but that's also an off-by-one error, not like it matters because what? It's actually hilarious |
Detecting when a stack overflow is going to happen in a portable way is tricky. If we could use |
Is this a stack overflow? |
You're right, my bad. In that case it should be easier to deal with. |
There is now an assert that checks Closing this as it's quite old, please re-open if you still encounter the problem with a recent revision. |
Yep, that looks like it would fix it~
Actually you'd be closing it because it's solved lol |
So. I'm trying to build with CMake on Windows 11 and the thing just stops after it's done loading the model.
And apparently, this is a segfault.
Yay yay yyayy yyayay
this is a memory allocation failure it seems, from me not having enough memory. not like llama.cpp Tells Me That lmao, it just segfaults
(
ctx->mem_buffer
is nullptr which probably means the malloc just failed)The text was updated successfully, but these errors were encountered: