Skip to content

[User] Bus error (core dumped) on a 65B model #666

New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Closed
mastr-ch13f opened this issue Mar 31, 2023 · 2 comments
Closed

[User] Bus error (core dumped) on a 65B model #666

mastr-ch13f opened this issue Mar 31, 2023 · 2 comments

Comments

@mastr-ch13f
Copy link

I tried running the a 65B model that was converted using the unversioned .py conversion script then migrated from an 8-file ggml .bin to a single-file ggjt .bin. Tried to run the model and I get a Bus error then the program ends.

user@ubuntu: ~/Desktop/llama.cpp$ ./main -m ./models/ggjt-model-model-q4_0.bin -t 7 -i
main: seed = 1680306291
llama_model_load: loading model from './models/ggjt-model-model-q4_0.bin' - please wait ...
llama_model_load: n_vocab = 32000
llama_model_load: n_ctx   = 512
llama_model_load: n_embd  = 8192
llama_model_load: n_mult  = 256
llama_model_load: n_head  = 64
llama_model_load: n_layer = 80
llama_model_load: n_rot   = 128
llama_model_load: f16     = 2
llama_model_load: n_ff    = 22016
llama_model_load: n_parts = 8
llama_model_load: type    = 4
llama_model_load: ggml map size = 38917.99 MB
llama_model_load: ggml ctx size = 201.25 KB
llama_model_load: mem required  = 41478.18 MB (+ 5120.00 MB per state)
llama_model_load: loading tensors from './models/ggjt-model-model-q4_0.bin'
llama_model_load: model size = 38917.53 MB / num tensors = 723
llama_init_from_file: kv self size  = 1280.00 MB

system_info: n_threads = 7 / 8 | AVX = 1 | AVX2 = 1 | AVX512 = 0 | FMA = 1 | NEON = 0 | ARM_FMA = 0 | F16C = 1 | FP16_VA = 0 | WASM_SIMD = 0 | BLAS = 0 | SSE3 = 1 | VSX = 0 | 
main: interactive mode on.
sampling: temp = 0.800000, top_k = 40, top_p = 0.950000, repeat_last_n = 64, repeat_penalty = 1.100000
generate: n_ctx = 512, n_batch = 8, n_predict = 128, n_keep = 0


== Running in interactive mode. ==
 - Press Ctrl+C to interject at any time.
 - Press Return to return control to LLaMa.
 - If you want to submit another line, end your input in '\'.

 Bus error (core dumped)
@FNsi
Copy link
Contributor

FNsi commented Apr 1, 2023

Maybe should Try with -t 8.
I think it's because hyper threading.
Try any even may work.

@Martin-HZK
Copy link

Same problem! But I encounter this when I try to reproduce the SYCL backend sample for Ubuntu Linux 24.04. I do not know why

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

4 participants