-
Notifications
You must be signed in to change notification settings - Fork 28k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Flex attention + refactor #34809
Comments
@ArthurZucker I'll try to open a pr for some architecture maybe llama and gemma |
I will do the Clip and Llama |
@ArthurZucker I would like to try out |
Picking this up for gpt2 and moshi! |
@mayankagarwals hey there since moshi actually copies some code form Gemma is it ok if I handle it |
I've already started some efforts but happy to work on it together if you are interested too? @dame-cell . Shot you a dm on kaggle. |
@mayankagarwals |
Opening this to add support for all models following #34282
Lets bring support for flex attention to more models! 🤗
It would be great to add the support for more architectures such as
... and many more
For anyone who wants to contribute just open a PR and link it to this issue, and ping me for a review!! 🤗
The text was updated successfully, but these errors were encountered: