Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Add Orthogonal Finetuning (OFT) #1133

Closed
lukaskuhn-lku opened this issue Nov 14, 2023 · 0 comments
Closed

Add Orthogonal Finetuning (OFT) #1133

lukaskuhn-lku opened this issue Nov 14, 2023 · 0 comments

Comments

@lukaskuhn-lku
Copy link
Contributor

lukaskuhn-lku commented Nov 14, 2023

Feature request

The paper Controlling Text-to-Image Diffusion by Orthogonal Finetuning proposes a new method of fine-tuning text-to-image diffusion models by using multiple learned orthogonal transformations on the layers of the pretrained model.

This preserves hyperspherical energy of the pretrained model (the sum of hyperspherical similarity (e.g., cosine similarity) between all pairwise neurons in the same layer), which leads to better generalization, stable training and faster convergence. It is basically a rotation of the neurons

In theory OFT can be applied to any layer and has some interesting interpretations for convolutional layers. For comparison reasons they only trained OFT on the same layers as LoRa in the original paper.

The corresponding repository gained 200 stars in 5 months, so there is definitely interest in the method. I think what is lacking is an easy to use implementation.

Motivation

I stumbled upon this paper and thought that it is a really unique idea of fine-tuning diffusion models and definitely works well in practice. I think an easy to use implementation is all that is lacking for this method to really benefit a lot of users.

Your contribution

I would love to fully contribute this method to peft, if there is no reason not to include it in the main branch.

@lukaskuhn-lku lukaskuhn-lku closed this as not planned Won't fix, can't repro, duplicate, stale Nov 14, 2023
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

1 participant