Skip to content

Conversation

@limin2021
Copy link
Contributor

@limin2021 limin2021 commented Sep 23, 2021

PR types

New features

PR changes

OPs

Describe

Add fused_attention_op backward impl:
1.fused_attention_op.cc and fused_attention_op.cu
The C++ impl of backward for fused_attention_op.
Related preceding RRs:
#34883, #35308, #35350 #35621 , #35903, #35905

2.common/fused_multhead_attention():
Add static graph construction method.

3.test_fused_attention_op.py
Add code to test the correctness of backward of fused_attention_op.

4.fused_transformer.py/FusedMultiHeadAttention layer:
Add FusedMultiHeadAttention layer.

5.test_fused_attention_op_api.py
Test the correctness of fused_attention_op python API, both dynamic and static graph.

Unittest results
b6c4dbda073ab3442b30cd770a84ce24
:

@paddle-bot-old
Copy link

Thanks for your contribution!
Please wait for the result of CI firstly. See Paddle CI Manual for details.

@paddle-bot-old
Copy link

Sorry to inform you that 6fb079c's CIs have passed for more than 7 days. To prevent PR conflicts, you need to re-run all CIs manually.

@limin2021 limin2021 closed this Oct 18, 2021
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

None yet

Projects

None yet

Development

Successfully merging this pull request may close these issues.

1 participant