Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Add flash attention to speedup fused_gate_attention. #52731

Merged
merged 58 commits into from
May 19, 2023

Conversation

JamesLim-sy
Copy link
Contributor

@JamesLim-sy JamesLim-sy commented Apr 10, 2023

PR types

Performance optimization

PR changes

OPs

Description

Pcard-70461

@paddle-bot
Copy link

paddle-bot bot commented Apr 10, 2023

你的PR提交成功,感谢你对开源项目的贡献!
请关注后续CI自动化测试结果,详情请参考Paddle-CI手册
Your PR has been submitted. Thanks for your contribution!
Please wait for the result of CI firstly. See Paddle CI Manual for details.

@JamesLim-sy JamesLim-sy force-pushed the add_flash_attn_for_af2 branch from e58d315 to 5bee3a6 Compare April 20, 2023 10:27
@JamesLim-sy JamesLim-sy force-pushed the add_flash_attn_for_af2 branch from 0cba876 to 9f76b5f Compare April 23, 2023 12:36
@Xreki Xreki force-pushed the add_flash_attn_for_af2 branch from d73c078 to c98186c Compare April 25, 2023 01:50
@Xreki Xreki force-pushed the add_flash_attn_for_af2 branch from c98186c to ade7a07 Compare April 25, 2023 01:51
@Xreki Xreki force-pushed the add_flash_attn_for_af2 branch from ade7a07 to 1ddf939 Compare April 25, 2023 07:41
wanghuancoder
wanghuancoder previously approved these changes Apr 25, 2023
Copy link
Contributor

@wanghuancoder wanghuancoder left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

LGTM

@Xreki Xreki force-pushed the add_flash_attn_for_af2 branch from c740bdc to 3b97303 Compare April 25, 2023 14:16
Shixiaowei02
Shixiaowei02 previously approved these changes May 9, 2023
luotao1
luotao1 previously approved these changes May 9, 2023
jiweibo
jiweibo previously approved these changes May 9, 2023
Xreki
Xreki previously approved these changes May 9, 2023
Copy link
Contributor

@Xreki Xreki left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

LGTM

@Xreki Xreki dismissed stale reviews from jiweibo, luotao1, Shixiaowei02, and themself via 7ff9f5e May 11, 2023 03:42
@@ -19,8 +19,8 @@ add_definitions(-DPADDLE_WITH_FLASHATTN)
set(FLASHATTN_PREFIX_DIR ${THIRD_PARTY_PATH}/flashattn)
set(FLASHATTN_SOURCE_SUBDIR csrc/flash_attn)
set(FLASHATTN_INSTALL_DIR ${THIRD_PARTY_PATH}/install/flashattn)
set(FLASHATTN_REPOSITORY ${GIT_URL}/PaddlePaddle/flash-attention.git)
set(FLASHATTN_TAG 5ff4bbf56ad066750407c4aef16ac740ebda0717)
set(FLASHATTN_REPOSITORY ${GIT_URL}/Xreki/flash-attention.git)
Copy link
Member

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Why use personal repo ?

cc @sneaxiy

Copy link
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

目前因包体积增大的比较多,在调试中,这是中间状态。

@Xreki Xreki force-pushed the add_flash_attn_for_af2 branch from b9debab to bee8537 Compare May 18, 2023 09:31
@Xreki Xreki changed the title Add flash attention for AF2 Add flash attention to speedup fused_gate_attention. May 19, 2023
@Xreki Xreki merged commit d29c1f8 into PaddlePaddle:develop May 19, 2023
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

Successfully merging this pull request may close these issues.

8 participants