-
Notifications
You must be signed in to change notification settings - Fork 19.5k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Add MultiQueryAttention
& GroupedQueryAttention
#18402
Comments
Probably easiest to just write This is also some discussion in #18423 for more context. I definitely think adding support here makes sense. And probably clearer to have this standalone from Thanks for filing! |
I was thinking the same thing. |
Should I open a PR for this?? |
Sounds good! Thank you! |
MultiQueryAttention (MQA) [Used in Falcon LLM] and GroupedQueryAttention (GQA) [Used in Llama 2 LLM] are alternatives to MultiHeadAttention (MHA) but they are a lot faster. Here's the speed comparison in my naive implementation,
I think it would be nice to have these layers in
keras-core
.Reference Papers:
The text was updated successfully, but these errors were encountered: