-
Notifications
You must be signed in to change notification settings - Fork 71
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Training Experiments and Insights #14
Comments
Hello, thanks for your feedback! Yes, the first arXiv paper was a bit rushed and lacks some refinement. Since this project is still under development, it might not fully align with the paper at this stage. We will continue to refine the paper until the final version is ready. We would greatly appreciate any help you can offer in correcting it! |
Thanks for your reply. I have a few questions: 1.The text mentions "plug-and-play," but the base model was trained. Isn't this contradictory? 2.Is there no comparison with methods like ControlNet? 3.Formulas (8) and (9) seem incorrect. 4.Figure 5 does not explain what the three results represent. 5.Figure 8 is unclear. 6.Figure 9 does not explain the conditions of the experiment on the left side. |
Hello, thanks for your question, and I think that they are all good questions! I will share more details.
|
Since the paper has limited space, I would like to share additional experiences. Can I change the title to 'Training Experiments and Insights'? @BJQ123456 SVD-related
We spent a lot of time to find these tips, now share all with all of you. May there will help you! |
2.We have compared the efficiency and training convergence. More detailed results will be added later. |
Thank you for your response; it was very helpful. |
There's one point I still don't understand. If it's plug-and-play, does it mean that after training, I can directly use it on odels? But during training, a part of the base model was trained, so if I insert it into a new model, that part hasn't been trained and could lead to a performance drop, right? |
Yes, as demonstrated in our experiments with SD1.5 and SDXL, we trained on a single backbone and then conducted experiments across various backbones. The results show that our method effectively performs control on different backbones. We also considered this approach and initially attempted to store the weight increments, same as LoRA only without low-rank compression. However, we eventually found that this step was unnecessary. |
I get it, thanks |
hi could the authors also share light on the combining this with IP-ADAPTERS? will it cause any issues making it work with IP-adapters? does it work well with pretrained IP-Adapters? |
感觉论文里面错误好多啊
The text was updated successfully, but these errors were encountered: