-
Notifications
You must be signed in to change notification settings - Fork 656
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
RuntimeError: mat1 and mat2 shapes cannot be multiplied #162
Comments
Exact same error while fine-tuning BLOOMZ: hyperonym/basaran#5 |
Same error while fine-tuning on https://github.com/lvwerra/trl/tree/main/examples/sentiment/scripts/gpt-neox-20b_peft step1.SFT with LoRA. Have you guys ever resolved it? @chenmingjiong @peakji |
I haven't found a solution yet. Now I'm using half-precision instead of int8. |
Same error when doing inference when there are multiple threads involved. h2oai/h2ogpt#104 |
Anyone solve this problem? I'm having this same issue when loading a LoRA'd checkpoint for Falcon 7B. |
same issue with llama 33b inference, any update here? |
This issue has been automatically marked as stale because it has not had recent activity. If you think this still needs to be addressed please comment on this thread. |
Was anyone able to solve this issue, getting this issue when fine-tuning with LORA and BitsandBytesConfig |
I got this error when finetuning "EleutherAI/gpt-j-6B" using load_in_8bit and LoRA on 8×2080ti:
RuntimeError: mat1 and mat2 shapes cannot be multiplied
I'm using data parallelism and not using model parallelism.
The code runs normally when training with 1 gpu, but failed when using accelerate on 8 gpu.
Reproduce steps:
clone this repo: https://github.com/CarperAI/trlx
modify the script: examples/summarize_rlhf/sft/train_gptj_summarize.py
and run:
accelerate launch --num_processes 8 examples/summarize_rlhf/sft/train_gptj_summarize.py
Full error logs:
The text was updated successfully, but these errors were encountered: