Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Fused attention op backward #35935

Closed

Conversation

limin2021
Copy link
Contributor

@limin2021 limin2021 commented Sep 23, 2021

PR types

New features

PR changes

OPs

Describe

Add fused_attention_op backward impl:
1.fused_attention_op.cc and fused_attention_op.cu
The C++ impl of backward for fused_attention_op.
Related preceding RRs:
#34883, #35308, #35350 #35621 , #35903, #35905

2.common/fused_multhead_attention():
Add static graph construction method.

3.test_fused_attention_op.py
Add code to test the correctness of backward of fused_attention_op.

4.fused_transformer.py/FusedMultiHeadAttention layer:
Add FusedMultiHeadAttention layer.

5.test_fused_attention_op_api.py
Test the correctness of fused_attention_op python API, both dynamic and static graph.

Unittest results
b6c4dbda073ab3442b30cd770a84ce24
:

@paddle-bot-old
Copy link

Thanks for your contribution!
Please wait for the result of CI firstly. See Paddle CI Manual for details.

@paddle-bot-old
Copy link

Sorry to inform you that 6fb079c's CIs have passed for more than 7 days. To prevent PR conflicts, you need to re-run all CIs manually.

@limin2021 limin2021 closed this Oct 18, 2021
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

Successfully merging this pull request may close these issues.

1 participant