Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Flex attention + refactor #34809

Open
1 of 7 tasks
ArthurZucker opened this issue Nov 19, 2024 · 7 comments
Open
1 of 7 tasks

Flex attention + refactor #34809

ArthurZucker opened this issue Nov 19, 2024 · 7 comments
Labels
Feature request Request for a new feature Good Difficult Issue PyTorch Anything PyTorch

Comments

@ArthurZucker
Copy link
Collaborator

Opening this to add support for all models following #34282

Lets bring support for flex attention to more models! 🤗

  • Gemma2

It would be great to add the support for more architectures such as

  • Qwen2
  • Llama
  • Gemma
  • QwenVl
  • Mistral
  • Clip

... and many more

For anyone who wants to contribute just open a PR and link it to this issue, and ping me for a review!! 🤗

@ArthurZucker ArthurZucker added Good Difficult Issue Feature request Request for a new feature PyTorch Anything PyTorch labels Nov 19, 2024
@dame-cell
Copy link

@ArthurZucker I'll try to open a pr for some architecture maybe llama and gemma

@farrosalferro
Copy link
Contributor

farrosalferro commented Nov 19, 2024

I will do the Clip and Llama

@OmarManzoor
Copy link
Contributor

@ArthurZucker I would like to try out Mistral

@mayankagarwals
Copy link
Contributor

Picking this up for gpt2 and moshi!

@dame-cell
Copy link

dame-cell commented Nov 22, 2024

Picking this up for gpt2 and moshi

@mayankagarwals hey there since moshi actually copies some code form Gemma is it ok if I handle it

@mayankagarwals
Copy link
Contributor

I've already started some efforts but happy to work on it together if you are interested too? @dame-cell . Shot you a dm on kaggle.

@dame-cell
Copy link

dame-cell commented Nov 24, 2024

I've already started some efforts but happy to work on it together if you are interested too? @dame-cell . Shot you a dm on kaggle.

@mayankagarwals
Thanks for reaching out
You are right i think it's best if you do it :)

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
Feature request Request for a new feature Good Difficult Issue PyTorch Anything PyTorch
Projects
None yet
Development

No branches or pull requests

5 participants