-
Notifications
You must be signed in to change notification settings - Fork 5.7k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
[DCU] fix bugs and surpport some fused ops #63217
Conversation
你的PR提交成功,感谢你对开源项目的贡献! |
❌ The PR is not created using PR's template. You can refer to this Demo. |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
LGTM
@@ -35,7 +35,11 @@ struct GeluFunctor { | |||
template <typename T> | |||
struct FastGeluFunctor { | |||
inline __device__ T operator()(const T x) const { | |||
#ifdef PADDLE_WITH_HIP | |||
PADDLE_ENFORCE(0, "FastGelu not surpport for rocm"); |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
CI这里提示错误,建议修改为
PADDLE_THROW(phi::errors::Unimplemented("ROCM does not support FastGelu"));
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
LGTM
PR Category
Custom Device
PR Types
Bug fixes
Description
重新支持dcu complex64/128
dcu支持fused_bias_residual_layernorm op
dcu支持fused_bias_dropout_residual_layer_norm前反向op
dcu支持rms_norm前反向op
解决depthwise conv grad op bug【https://github.com/PaddlePaddle/Paddle/issues/60500】
解决 hip graph test bugs ,已知问题:需要 export HIP_GRAPH_LARGE_LAUNCH=0
相关test均通过