You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
You verified that this is a bug and not a feature request or question by asking in the discord?
Yes
Describe the bug
As a newcomer, I have trained out the lora weight on the gradio UI. Intuitively, I tried to run inference on the UI. However, there is no way to inference on it.
Then, I follow the README in the directory of the lora weight step by step,
torch.OutOfMemoryError: CUDA out of memory. Tried to allocate 90.00 MiB. GPU 1 has a total capacity of 23.55 GiB of which 51.69 MiB is free. Including non-PyTorch memory, this process has 23.49 GiB memory in use. Of the allocated memory 23.11 GiB is allocated by PyTorch, and 12.93 MiB is reserved by PyTorch but unallocated. If reserved but unallocated memory is large try setting PYTORCH_CUDA_ALLOC_CONF=expandable_segments:True to avoid fragmentation. See documentation for Memory Management (https://pytorch.org/docs/stable/notes/cuda.html#environment-variables)
My gpu is gtx4090 with 24g vram in.
Now I lose the direction. Beg for help from anyone.
The text was updated successfully, but these errors were encountered:
There's less information about FLUX inference on consumer hardware than I expected, I had some luck running it using optimum-quanto. Could you try pip install optimum-quanto and then the following code?
This is for bugs only
Did you already ask in the discord?
Yes
You verified that this is a bug and not a feature request or question by asking in the discord?
Yes
Describe the bug
As a newcomer, I have trained out the lora weight on the gradio UI. Intuitively, I tried to run inference on the UI. However, there is no way to inference on it.
Then, I follow the README in the directory of the lora weight step by step,
.
Then I got
My gpu is gtx4090 with 24g vram in.
Now I lose the direction. Beg for help from anyone.
The text was updated successfully, but these errors were encountered: