Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

[BUG] OOM errors during fine-tuning with Polyglot 12.8B models #330

Closed
yum-yeom opened this issue Aug 2, 2023 · 1 comment
Closed

[BUG] OOM errors during fine-tuning with Polyglot 12.8B models #330

yum-yeom opened this issue Aug 2, 2023 · 1 comment
Labels
type/bug Bug in code

Comments

@yum-yeom
Copy link

yum-yeom commented Aug 2, 2023

🐛 Bug

hello. I'm trying to train Polyglot 12.8B model using h2o llm-studio, but I'm getting an OOM error.

The GPU I am using is Nvidia A5000 24GB 4 sheets, and the params I used are as follows zip file.

Is there any solution?

@yum-yeom yum-yeom added the type/bug Bug in code label Aug 2, 2023
@pascal-pfeiffer
Copy link
Collaborator

There are multiple ways to bring down memory consumption.
Mainly:

  • int4 training
  • smaller batchsize
  • lower token length for question and answer

There is also an ongoing effort to split the model weights across multiple GPUs during training in this PR: #288

@yum-yeom yum-yeom closed this as completed Aug 4, 2023
@yum-yeom yum-yeom reopened this Aug 4, 2023
@psinger psinger closed this as completed Aug 18, 2023
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
type/bug Bug in code
Projects
None yet
Development

No branches or pull requests

3 participants