-
I can't figure out the settings for SD 3.5-Large. ~60 image 1024x1024 square dataset, 16GB CUDA card, Fedora linux. Here's the configuration I'm currently using:
I have to train on 512px because otherwise I'm getting an OOM while SimpleTuner is initially processing the dataset (I'm not sure if 1024px training is possible on 16GB, but would appreciate help here too). Here's the dataset config:
Unless I use max_grad_norm 0.1, images quickly turn into noise, and resulting lora is unfunctional: If I'm using max_grad_norm 0.1, images don't turn into noise (one exception being an unmprompted image) but the model seems to not be learning the concept, I've let it go to 3500 steps and the images were not really improving: For SD 3.5, I've tried several optimizers and quantization settings to no avail. To check if I'm doing something fundamentally wrong, I've used the same dataset to train a Flux-dev lora and it trained just fine, just in case here's the JSON I was using for Flux:
I'm not sure how to proceed further and would appreciate help, I'd be happy to provide any additional information if needed. Thanks in advance for any suggestions. |
Beta Was this translation helpful? Give feedback.
Replies: 1 comment
-
it was a mix of me missing that mixed_precision should be set to 'no' and a possible vae cache mixup. |
Beta Was this translation helpful? Give feedback.
it was a mix of me missing that mixed_precision should be set to 'no' and a possible vae cache mixup.