diff --git a/Paddle b/Paddle index 2588f489910..3b674aa2f0d 160000 --- a/Paddle +++ b/Paddle @@ -1 +1 @@ -Subproject commit 2588f4899106cd27bdfcc84ba4c2f5f7aac570ab +Subproject commit 3b674aa2f0db8b9cb41c6d0738e3ebfebab993cd diff --git a/backends/metax_gpu/CMakeLists.txt b/backends/metax_gpu/CMakeLists.txt index c80982449ae..de678c4ded2 100755 --- a/backends/metax_gpu/CMakeLists.txt +++ b/backends/metax_gpu/CMakeLists.txt @@ -72,6 +72,9 @@ include(zlib) include(protobuf) include(generate_pb) +# Add fusion headers search path for phi::fusion::FusedLayerNormKernel +include_directories("${PADDLE_SOURCE_DIR}/paddle/phi/kernels/fusion") + set(PROTO_FILE "${PADDLE_SOURCE_DIR}/paddle/phi/core/external_error.proto") get_filename_component(PROTO_WE "${PROTO_FILE}" NAME_WE) diff --git a/backends/metax_gpu/cmake/paddle.cmake b/backends/metax_gpu/cmake/paddle.cmake index 899ffd2dd30..71a0b9a2c90 100755 --- a/backends/metax_gpu/cmake/paddle.cmake +++ b/backends/metax_gpu/cmake/paddle.cmake @@ -83,10 +83,14 @@ else() ABSOLUTE) message( STATUS "Run 'git submodule update --init Paddle' in ${REPO_SOURCE_DIR}") - # execute_process( COMMAND git submodule update --init Paddle - # WORKING_DIRECTORY ${REPO_SOURCE_DIR} RESULT_VARIABLE result_var) if(NOT - # result_var EQUAL 0) message( FATAL_ERROR "Failed to get submodule Paddle', - # please check your network !" ) endif() + execute_process( + COMMAND git submodule update --init Paddle + WORKING_DIRECTORY ${REPO_SOURCE_DIR} + RESULT_VARIABLE result_var) + if(NOT result_var EQUAL 0) + message(FATAL_ERROR "Failed to get submodule Paddle', + please check your network !") + endif() get_filename_component(PADDLE_SOURCE_DIR "${REPO_SOURCE_DIR}/Paddle" ABSOLUTE) message(STATUS "PADDLE_SOURCE_DIR=${PADDLE_SOURCE_DIR}") diff --git a/backends/metax_gpu/kernels/fusion/fused_layernorm_kernel_register.cu b/backends/metax_gpu/kernels/fusion/fused_layernorm_kernel_register.cu index bdb809a2149..ff6a97ba691 100644 --- a/backends/metax_gpu/kernels/fusion/fused_layernorm_kernel_register.cu +++ b/backends/metax_gpu/kernels/fusion/fused_layernorm_kernel_register.cu @@ -14,11 +14,7 @@ #include "paddle/phi/core/kernel_registry.h" #include "paddle/phi/kernels/fused_layernorm_kernel.h" -#include "paddle/phi/kernels/fusion/gpu/attention_layer.norm.h" -#include "paddle/phi/kernels/fusion/gpu/fused_dropout_helper.h" -#ifndef PADDLE_WITH_HIP -#if CUDNN_VERSION_MIN(8, 1, 0) PD_CUSTOM_KERNEL_REGISTER(fused_bias_residual_layernorm, metax_gpu, ALL_LAYOUT, @@ -32,32 +28,3 @@ PD_CUSTOM_KERNEL_REGISTER(fused_bias_residual_layernorm, kernel->OutputAt(2).SetDataType(phi::DataType::FLOAT32); kernel->OutputAt(3).SetDataType(phi::DataType::FLOAT32); } -#else -PD_CUSTOM_KERNEL_REGISTER(fused_bias_residual_layernorm, - metax_gpu, - ALL_LAYOUT, - phi::fusion::FusedLayerNormKernel, - float, - phi::dtype::float16) { - kernel->InputAt(3).SetDataType(phi::DataType::FLOAT32); - kernel->InputAt(4).SetDataType(phi::DataType::FLOAT32); - kernel->OutputAt(0).SetDataType(phi::DataType::UNDEFINED); - kernel->OutputAt(2).SetDataType(phi::DataType::FLOAT32); - kernel->OutputAt(3).SetDataType(phi::DataType::FLOAT32); -} -#endif // CUDNN_VERSION_MIN -#else -PD_CUSTOM_KERNEL_REGISTER(fused_bias_residual_layernorm, - metax_gpu, - ALL_LAYOUT, - phi::fusion::FusedLayerNormKernel, - float, - phi::dtype::float16, - phi::dtype::bfloat16) { - kernel->InputAt(3).SetDataType(phi::DataType::FLOAT32); - kernel->InputAt(4).SetDataType(phi::DataType::FLOAT32); - kernel->OutputAt(0).SetDataType(phi::DataType::UNDEFINED); - kernel->OutputAt(2).SetDataType(phi::DataType::FLOAT32); - kernel->OutputAt(3).SetDataType(phi::DataType::FLOAT32); -} -#endif // PADDLE_WITH_HIP