Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Add eval_batch_size for evaluation #12

Merged
merged 1 commit into from
May 7, 2023

Conversation

NanoCode012
Copy link
Collaborator

@NanoCode012 NanoCode012 commented May 6, 2023

Problem:

I would OOM during the evaluation stage despite setting micro_batch_size to be low. Turns out that batch size for eval using a different parameter.

Proposal:

New config for eval stage for lower VRAM GPUs.

Edit:

Since it's now a somewhat required argument, I'm not sure whether we should

  • assert value exist
  • set some default if not in config

Copy link
Collaborator

@winglian winglian left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

thanks!

@winglian winglian merged commit a15d823 into axolotl-ai-cloud:main May 7, 2023
@NanoCode012 NanoCode012 deleted the feat/eval_config branch May 7, 2023 05:48
djsaunde pushed a commit that referenced this pull request Dec 17, 2024
Add eval_batch_size for evaluation
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

Successfully merging this pull request may close these issues.

2 participants