Closed
Description
Hi, @pcuenca @patil-suraj @anton-l
I'm trying to fully fine-tune SD u-net on my own dataset, including about 1M image-text pairs.
I'm following the script in examples/text_to_image
.
However, it's blocked at load_dataset with very slow processing. It has been 12 hrs but the preparation of these 1M data is still not ready.
diffusers/examples/text_to_image/train_text_to_image.py
Lines 424 to 431 in ef3fcbb
I tried another smaller dataset with only thousands of image-text pairs and everything worked fine.
So I was just wondering if this slow-loading process is expected for large datasets.
OS: Ubuntu-20.04, GPU: 32GB V100 x 8, dependencies: according to the current installation instructions
Metadata
Metadata
Assignees
Labels
No labels