-
Notifications
You must be signed in to change notification settings - Fork 59
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Wired CUDA memory utilization #16
Comments
Hi! Thanks for your interest. Have you tried accelerate? That worked for us! The python way also works, but is very slow. Definitely try accelerate, but if you don’t want to I’d at least switch to 4 A100 80gb GPUs. |
Hi, thanks for your reply. Is there any script I can refer to if I want to try accelerate? Also, do you mean that the python mode runs faster with 4*a100 than 8*a100?---- Replied Message ----FromAriel N. ***@***.***>Date08/24/2023 21:39 ***@***.***> ***@***.***>***@***.***>SubjectRe: [arielnlee/Platypus] Wired CUDA memory utilization (Issue #16)
Hi! Thanks for your interest. Have you tried accelerate? That worked for us! The python way also works, but is very slow. Definitely try accelerate, but if you don’t want to I’d at least switch to 4 A100 80gb GPUs.
—Reply to this email directly, view it on GitHub, or unsubscribe.You are receiving this because you authored the thread.Message ID: ***@***.***>
[
{
***@***.***": "http://schema.org",
***@***.***": "EmailMessage",
"potentialAction": {
***@***.***": "ViewAction",
"target": "#16 (comment)",
"url": "#16 (comment)",
"name": "View Issue"
},
"description": "View this Issue on GitHub",
"publisher": {
***@***.***": "Organization",
"name": "GitHub",
"url": "https://github.com"
}
}
]
|
First run To clarify, running |
I just tried running the script with accelerate launch(8*a100-80gb), but it went CUDA OOM during model loading. Any advice? |
same problem. I solve this by reinstall the python package with the version in requirement.txt,i think this is relate with the peft package. |
Hi, I am using the python launch to lora-finetune Llama2-70b, the training is doing good. But, it seems a bit wired that the memory utilization is quite low, less than 18G. Also, the training speed is relatively slow compared to the codebase in llama-recipes.
The command is:
The gpu status during training is:
The text was updated successfully, but these errors were encountered: