Skip to content

[hybrid] Fused attention support cache & tensor model parallel#40361

Closed
wangxicoding wants to merge 2 commits intoPaddlePaddle:developfrom
wangxicoding:fused_attention_support_cache
Closed

[hybrid] Fused attention support cache & tensor model parallel#40361
wangxicoding wants to merge 2 commits intoPaddlePaddle:developfrom
wangxicoding:fused_attention_support_cache

Conversation

@wangxicoding
Copy link
Copy Markdown
Contributor

PR types

New features

PR changes

OPs

Describe

Fused attention support cache & tensor model parallel

@paddle-bot-old
Copy link
Copy Markdown

paddle-bot-old Bot commented Mar 9, 2022

Thanks for your contribution!
Please wait for the result of CI firstly. See Paddle CI Manual for details.

@wangxicoding wangxicoding changed the title [WIP] Fused attention support cache & tensor model parallel [hybrid] Fused attention support cache & tensor model parallel Mar 9, 2022
@wangxicoding
Copy link
Copy Markdown
Contributor Author

close with same PR:#40101

@wangxicoding wangxicoding deleted the fused_attention_support_cache branch March 10, 2022 08:29
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

None yet

Projects

None yet

Development

Successfully merging this pull request may close these issues.

1 participant