Skip to content

Finetuning Llama-2-13B with 1x a100 80gb? torch.cuda.OutOfMemoryError #356

Closed Answered by haotian-liu
chigkim asked this question in Q&A
Discussion options

You must be logged in to vote

Full-finetuning 13B model may be too tough for a single A100 80GB 😭

You can try QLoRA, which is optimized for low VRAM usage.

To further save the memory, you can try zero3_offload, and see explanations here

Replies: 5 comments 2 replies

Comment options

You must be logged in to vote
0 replies
Answer selected by chigkim
Comment options

You must be logged in to vote
0 replies
Comment options

You must be logged in to vote
0 replies
Comment options

You must be logged in to vote
1 reply
@haotian-liu
Comment options

Comment options

You must be logged in to vote
1 reply
@nighting0le01
Comment options

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Category
Q&A
Labels
None yet
3 participants