Skip to content

Conversation

@yangjianfengo1
Copy link
Contributor

在集中式场景下,若export FD_ATTENTION_BACKEND="FLASH_ATTN",开启FLASH_ATTN,那么encoder 的时候在A卡上会使用fa2,在H卡上会使用fa3,decoder的时候会使用append attn。

@paddle-bot
Copy link

paddle-bot bot commented Jul 31, 2025

Thanks for your contribution!

@gongshaotian gongshaotian merged commit 64d7a31 into PaddlePaddle:develop Aug 1, 2025
11 of 14 checks passed
@yangjianfengo1 yangjianfengo1 deleted the develop123 branch August 4, 2025 06:56
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

None yet

Projects

None yet

Development

Successfully merging this pull request may close these issues.

3 participants