-
Notifications
You must be signed in to change notification settings - Fork 5.6k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
FusedMultiTransformer optimization #59385
FusedMultiTransformer optimization #59385
Conversation
你的PR提交成功,感谢你对开源项目的贡献! |
wufeisheng seems not to be a GitHub user. You need a GitHub account to be able to sign the CLA. If you have already a GitHub account, please add the email address used for this commit to your account. You have signed the CLA already but the status is still pending? Let us recheck it. |
446cfd2
to
52ba6ac
Compare
@@ -246,6 +246,7 @@ PD_REGISTER_KERNEL(flash_attn_unpadded, | |||
GPU, | |||
ALL_LAYOUT, | |||
phi::FlashAttnUnpaddedKernel, | |||
float, |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
为什么注册这个类型?应该不支持?
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
因为FuseMT算子注册了float类型,如果这里不注册会报链接错误;实际测试确实也没使用过float,我测试一下,如果确认不支持的话可能需要加Dispatch逻辑
…ddle into infer_merge_to_train
PR types
New features
PR changes
OPs
Description