-
Notifications
You must be signed in to change notification settings - Fork 10.2k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
[bug] ROCm segfault when running multi-gpu inference. #3451
Comments
Can you better describe the |
I had to start it up with the -lv option when using multiple cards. Kind of sucks though as it makes it a lot slower. This was on a windows machine. Worked fine without the option on my linux server. |
What's the output of On Windows I think the equivalent would be resizable BAR, and AMD has that functionality limited to certain GPUs and hardware configurations. |
My motherboard has no support for ReBar or above 4g decoding, so this should be the case. Closing issue for now. |
Try this for making ReBar GA😄 |
Prerequisites
Please answer the following questions for yourself before submitting an issue.
Expected Behavior
Expected Tensor split to leverage multi gpus.
Current Behavior
Segfault after model loading when using multi-gpu. Correct inference when using either GPU(two vega-56s installed) and HIP_VISIBLE_DEVICES to force single GPU inference.
Environment and Context
Please provide detailed information about your computer setup. This is important in case the issue is not reproducible except for under certain specific conditions.
Ryzen 1700x
Vega-56 8G*2
Linux jerryxu-Inspiron-5675 6.2.0-33-generic #33~22.04.1-Ubuntu SMP PREEMPT_DYNAMIC Thu Sep 7 10:33:52 UTC 2 x86_64 x86_64 x86_64 GNU/Linux
Failure Information (for bugs)
See logs.
Steps to Reproduce.
Failure Logs
llama.cpp log:
GDB stacktrace on segfault:
The text was updated successfully, but these errors were encountered: