-
Notifications
You must be signed in to change notification settings - Fork 531
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Flash Attention vs Triton Flash Attention #180
Comments
Without delving into the implementation details, using ALiBi as the networks position embedding is beneficial regardless of how exactly it is supported. Relative position embeddingThe default positional embedding is learned positional embeddings. The issue with learned positional embeddings is that inference max seq len is limited to the training max seq len. ConvergenceIn my experience, ALiBi also has faster convergence to other position embedding schemes we tried. Implementation details ie TritonThis is the flash attn implementation we use if
Given that triton has faster forward + backward, I'd advocate for using the triton version. The fact that it supports ALiBi is a bonus (a very welcome bonus); I wouldn't necessarily call it a tradeoff. |
great stuff, if it's faster it's cool, but here they told about slower computation, I am not wrong? |
by the way can you tell me about setpu.py please, why you use torch 1.13.1 there, if we want to use torch 2? |
why you turn off this from setup.py but in main branch i guess its ok and its triton pre mlir |
upd: everything is fine with current main version here is versions (just to information when branch will be updates):
installed everything from setup.py(gpu setup) with this docker |
Hi, i want to know about choice in your MPT model
Yes, Triton version supports alibi and fast forward, but it's have some disadvantages:
Do you think alibi choice is so important in this case?
It's looks like some trade off
The text was updated successfully, but these errors were encountered: