-
Notifications
You must be signed in to change notification settings - Fork 36
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
CUDA out of memory error. #240
Comments
Hi @seemavishal, could you paste the whole stack trace of the error? From the looks of it, provided that the attention attribution method does not perform any additional action beyond the forward passes required for generation, I think the error might get raised at the generation stage with If this is the case, the only solution would be to use a smaller model, or get access to a machine with more GPU memory! |
Hi @gsarti Pasting the error trace below.
|
Hello,
I am using inseq to generate explanation for text summarization problem of long length input article.
While the code works well with short length article it throws an error if the length of the article is increased.
I am using 48 GB GPU memory on cloud to execute this.
Following is the code i have used.
I get following error.
CUDA out of memory. Tried to allocate 15.02 GiB (GPU 0; 47.54 GiB total capacity; 32.75 GiB already allocated; 13.67 GiB free; 32.87 GiB reserved in total by PyTorch) If reserved memory is >> allocated memory try setting max_split_size_mb to avoid fragmentation. See documentation for Memory Management and PYTORCH_CUDA_ALLOC_CONF
I have tried with defining max_split_size_mb = 256 as environment variable, cleared cache as well but nothing worked.
The text was updated successfully, but these errors were encountered: