From dda502ebaa79e2113c739f85c54d70dedaa1c4fd Mon Sep 17 00:00:00 2001 From: co63oc Date: Sat, 9 Aug 2025 06:50:14 +0800 Subject: [PATCH] test/cpp rename use_mkldnn --- .../cpp/eager/performance_tests/benchmark_utils.cc | 2 +- .../elementwise/test_elementwise_add_grad_grad.cc | 2 +- .../elementwise/test_elementwise_div_grad_grad.cc | 2 +- test/cpp/fluid/mkldnn/test_conv_onednn_nhwc.cc | 10 +++++----- test/cpp/fluid/mkldnn/test_onednn_caching.cc | 4 ++-- .../mkldnn/test_onednn_conv2d_transpose_bias.cc | 2 +- test/cpp/fluid/mkldnn/test_onednn_op_inplace.cc | 8 ++++---- test/cpp/fluid/mkldnn/test_onednn_op_nhwc.cc | 14 +++++++------- .../fluid/mkldnn/test_onednn_pool_adaptive_op.cc | 2 +- test/cpp/fluid/mkldnn/test_onednn_squeeze.cc | 4 ++-- test/cpp/fluid/op_debug_string_test.cc | 2 +- test/cpp/imperative/test_hooks.cc | 4 ++-- test/cpp/imperative/test_prepare_op.cc | 2 +- test/cpp/imperative/test_tracer.cc | 14 +++++++------- 14 files changed, 36 insertions(+), 36 deletions(-) diff --git a/test/cpp/eager/performance_tests/benchmark_utils.cc b/test/cpp/eager/performance_tests/benchmark_utils.cc index 7b95d911bc5345..23218075517c4c 100644 --- a/test/cpp/eager/performance_tests/benchmark_utils.cc +++ b/test/cpp/eager/performance_tests/benchmark_utils.cc @@ -228,7 +228,7 @@ void benchmark_fluid_scale(const std::shared_ptr& X, imperative::Tracer tracer; framework::AttributeMap attrs; - attrs["use_mkldnn"] = false; + attrs["use_onednn"] = false; attrs["scale"] = 2; attrs["bias"] = 3; attrs["bias_after_scale"] = true; diff --git a/test/cpp/fluid/elementwise/test_elementwise_add_grad_grad.cc b/test/cpp/fluid/elementwise/test_elementwise_add_grad_grad.cc index 6186cfa2c9756f..ec00557d6a0dd5 100644 --- a/test/cpp/fluid/elementwise/test_elementwise_add_grad_grad.cc +++ b/test/cpp/fluid/elementwise/test_elementwise_add_grad_grad.cc @@ -59,7 +59,7 @@ class TestElementwiseAddGradGradWithoutDDX this->op_type_, {{"Y", {"Y"}}, {"DOut", {"DOut"}}, {"DDY", {"DDY"}}}, {{"DDOut", {"DDOut"}}}, - {{"use_mkldnn", false}, {"axis", 0}}); + {{"use_onednn", false}, {"axis", 0}}); return op; } }; diff --git a/test/cpp/fluid/elementwise/test_elementwise_div_grad_grad.cc b/test/cpp/fluid/elementwise/test_elementwise_div_grad_grad.cc index 8f1ed87888ba44..f4ecb943a8dd9c 100644 --- a/test/cpp/fluid/elementwise/test_elementwise_div_grad_grad.cc +++ b/test/cpp/fluid/elementwise/test_elementwise_div_grad_grad.cc @@ -87,7 +87,7 @@ class TestElementwiseDivGradGradWithDout : public TestElementwiseOpGradGrad { {"DDY", {"DDY"}}, {"DX", {"DX"}}}, {{"Y@GRAD", {"Y@GRAD"}}, {"DDOut", {"DDOut"}}, {"DOut", {"DOut"}}}, - {{"use_mkldnn", false}, {"axis", 0}}); + {{"use_onednn", false}, {"axis", 0}}); return op; } }; diff --git a/test/cpp/fluid/mkldnn/test_conv_onednn_nhwc.cc b/test/cpp/fluid/mkldnn/test_conv_onednn_nhwc.cc index 28028858c3bac0..49071d5938a744 100644 --- a/test/cpp/fluid/mkldnn/test_conv_onednn_nhwc.cc +++ b/test/cpp/fluid/mkldnn/test_conv_onednn_nhwc.cc @@ -66,7 +66,7 @@ TEST(test_conv2d_output, fp32) { conv2d_op.SetAttr("paddings", paddings); conv2d_op.SetAttr("dilations", dilations); conv2d_op.SetAttr("groups", groups); - conv2d_op.SetAttr("use_mkldnn", true); + conv2d_op.SetAttr("use_onednn", true); auto op = paddle::framework::OpRegistry::CreateOp(conv2d_op); @@ -95,7 +95,7 @@ TEST(test_conv2d_output, int8) { conv2d_op.SetAttr("paddings", paddings); conv2d_op.SetAttr("dilations", dilations); conv2d_op.SetAttr("groups", groups); - conv2d_op.SetAttr("use_mkldnn", true); + conv2d_op.SetAttr("use_onednn", true); conv2d_op.SetAttr("mkldnn_data_type", std::string("int8")); conv2d_op.SetAttr("force_fp32_output", false); @@ -126,7 +126,7 @@ TEST(test_conv2d_output, ic1) { conv2d_op.SetAttr("paddings", paddings); conv2d_op.SetAttr("dilations", dilations); conv2d_op.SetAttr("groups", groups); - conv2d_op.SetAttr("use_mkldnn", true); + conv2d_op.SetAttr("use_onednn", true); auto op = paddle::framework::OpRegistry::CreateOp(conv2d_op); @@ -156,7 +156,7 @@ TEST(test_conv2d_output, ic2) { conv2d_op.SetAttr("paddings", paddings); conv2d_op.SetAttr("dilations", dilations); conv2d_op.SetAttr("groups", groups); - conv2d_op.SetAttr("use_mkldnn", true); + conv2d_op.SetAttr("use_onednn", true); auto op = paddle::framework::OpRegistry::CreateOp(conv2d_op); @@ -186,7 +186,7 @@ TEST(test_conv2d_output, ic4) { conv2d_op.SetAttr("paddings", paddings); conv2d_op.SetAttr("dilations", dilations); conv2d_op.SetAttr("groups", groups); - conv2d_op.SetAttr("use_mkldnn", true); + conv2d_op.SetAttr("use_onednn", true); auto op = paddle::framework::OpRegistry::CreateOp(conv2d_op); diff --git a/test/cpp/fluid/mkldnn/test_onednn_caching.cc b/test/cpp/fluid/mkldnn/test_onednn_caching.cc index 694d9aeb6e3bc7..d87e1c4145f5b2 100644 --- a/test/cpp/fluid/mkldnn/test_onednn_caching.cc +++ b/test/cpp/fluid/mkldnn/test_onednn_caching.cc @@ -115,12 +115,12 @@ void RunOperator(const phi::Place &place, {{first_input_var_name, {first_input}}, {second_input_var_name, {"x1"}}}, {{output_var_name, {output_name}}}, - {{"use_mkldnn", {true}}}) + {{"use_onednn", {true}}}) : framework::OpRegistry::CreateOp( op_type, {{first_input_var_name, {first_input}}}, {{output_var_name, {output_name}}}, - {{"use_mkldnn", {true}}}); + {{"use_onednn", {true}}}); op->Run(scope, place); pool.Get(place)->Wait(); diff --git a/test/cpp/fluid/mkldnn/test_onednn_conv2d_transpose_bias.cc b/test/cpp/fluid/mkldnn/test_onednn_conv2d_transpose_bias.cc index 87311b8e9a2acd..6e5218c157f41e 100644 --- a/test/cpp/fluid/mkldnn/test_onednn_conv2d_transpose_bias.cc +++ b/test/cpp/fluid/mkldnn/test_onednn_conv2d_transpose_bias.cc @@ -61,7 +61,7 @@ void test_conv2d_transpose_bias() { AddVarToScope("convtranspose-Bias", &scope, {256}); AddVarToScope("convtranspose-Out", &scope, {1, 256, 27, 23}); - desc.SetAttr("use_mkldnn", true); + desc.SetAttr("use_onednn", true); desc.SetAttr("is_test", true); auto op = paddle::framework::OpRegistry::CreateOp(desc); diff --git a/test/cpp/fluid/mkldnn/test_onednn_op_inplace.cc b/test/cpp/fluid/mkldnn/test_onednn_op_inplace.cc index 54ff2aa51bb8e4..90e296790107e2 100644 --- a/test/cpp/fluid/mkldnn/test_onednn_op_inplace.cc +++ b/test/cpp/fluid/mkldnn/test_onednn_op_inplace.cc @@ -87,11 +87,11 @@ bool TestMain(const phi::Place &place, ? framework::OpRegistry::CreateOp(op_type, {{"X", {"x"}}, {"Y", {"x1"}}}, {{"Out", {"y"}}}, - {{"use_mkldnn", {true}}}) + {{"use_onednn", {true}}}) : framework::OpRegistry::CreateOp(op_type, {{"X", {"x"}}}, {{"Out", {"y"}}}, - {{"use_mkldnn", {true}}}); + {{"use_onednn", {true}}}); op_ref->Run(scope, place); pool.Get(place)->Wait(); @@ -104,11 +104,11 @@ bool TestMain(const phi::Place &place, ? framework::OpRegistry::CreateOp(op_type, {{"X", {"x"}}, {"Y", {"x1"}}}, {{"Out", {"x"}}}, - {{"use_mkldnn", {true}}}) + {{"use_onednn", {true}}}) : framework::OpRegistry::CreateOp(op_type, {{"X", {"x"}}}, {{"Out", {"x"}}}, - {{"use_mkldnn", {true}}}); + {{"use_onednn", {true}}}); op->Run(scope, place); phi::DeviceContextPool::Instance().Get(place)->Wait(); diff --git a/test/cpp/fluid/mkldnn/test_onednn_op_nhwc.cc b/test/cpp/fluid/mkldnn/test_onednn_op_nhwc.cc index f946a0aee1f49c..fc3073f1440759 100644 --- a/test/cpp/fluid/mkldnn/test_onednn_op_nhwc.cc +++ b/test/cpp/fluid/mkldnn/test_onednn_op_nhwc.cc @@ -67,7 +67,7 @@ void Test_Pool2d_Transpose_NHWC(const std::string &transpose_type) { {{"pooling_type", {std::string("max")}}, {"ksize", {ksize}}, {"data_format", {std::string("NHWC")}}, - {"use_mkldnn", {true}}}); + {"use_onednn", {true}}}); auto axis = std::vector(4, 0); axis[1] = 2; @@ -77,7 +77,7 @@ void Test_Pool2d_Transpose_NHWC(const std::string &transpose_type) { transpose_type, {{"X", {"y"}}}, {{"Out", {"z"}}}, - {{"axis", {axis}}, {"use_mkldnn", {true}}}); + {{"axis", {axis}}, {"use_onednn", {true}}}); op_pool->Run(scope, p); op_transpose->Run(scope, p); @@ -130,7 +130,7 @@ TEST(test_pool2d_relu_relu_nhwc, cpu_place) { {{"pooling_type", {std::string("max")}}, {"ksize", {ksize}}, {"data_format", {std::string("NHWC")}}, - {"use_mkldnn", {true}}}); + {"use_onednn", {true}}}); auto axis = std::vector(4, 0); axis[1] = 2; @@ -140,10 +140,10 @@ TEST(test_pool2d_relu_relu_nhwc, cpu_place) { "relu", {{"X", {"y"}}}, {{"Out", {"u"}}}, - {{"axis", {axis}}, {"use_mkldnn", {false}}}); + {{"axis", {axis}}, {"use_onednn", {false}}}); auto op_relu2 = framework::OpRegistry::CreateOp( - "relu", {{"X", {"u"}}}, {{"Out", {"z"}}}, {{"use_mkldnn", {true}}}); + "relu", {{"X", {"u"}}}, {{"Out", {"z"}}}, {{"use_onednn", {true}}}); op_pool->Run(scope, p); op_relu1->Run(scope, p); @@ -192,10 +192,10 @@ TEST(test_pool2d_shape_nhwc, cpu_place) { {{"pooling_type", {std::string("max")}}, {"ksize", {ksize}}, {"data_format", {std::string("NHWC")}}, - {"use_mkldnn", {true}}}); + {"use_onednn", {true}}}); auto op_shape = framework::OpRegistry::CreateOp( - "shape", {{"Input", {"y"}}}, {{"Out", {"z"}}}, {{"use_mkldnn", {true}}}); + "shape", {{"Input", {"y"}}}, {{"Out", {"z"}}}, {{"use_onednn", {true}}}); op_pool->Run(scope, p); op_shape->Run(scope, p); diff --git a/test/cpp/fluid/mkldnn/test_onednn_pool_adaptive_op.cc b/test/cpp/fluid/mkldnn/test_onednn_pool_adaptive_op.cc index 684ad2f1cc3775..1e45aad938ca8d 100644 --- a/test/cpp/fluid/mkldnn/test_onednn_pool_adaptive_op.cc +++ b/test/cpp/fluid/mkldnn/test_onednn_pool_adaptive_op.cc @@ -67,7 +67,7 @@ void test_pool2d(bool adaptive, bool ceil_mode, std::string pool_type = "max") { desc.SetAttr("paddings", paddings); desc.SetAttr("adaptive", adaptive); desc.SetAttr("ceil_mode", ceil_mode); - desc.SetAttr("use_mkldnn", true); + desc.SetAttr("use_onednn", true); auto op = paddle::framework::OpRegistry::CreateOp(desc); diff --git a/test/cpp/fluid/mkldnn/test_onednn_squeeze.cc b/test/cpp/fluid/mkldnn/test_onednn_squeeze.cc index 0a5b253e05bcab..b1dfd5ab5d1b79 100644 --- a/test/cpp/fluid/mkldnn/test_onednn_squeeze.cc +++ b/test/cpp/fluid/mkldnn/test_onednn_squeeze.cc @@ -62,7 +62,7 @@ void test_squeeze() { std::vector axes({-2}); desc.SetAttr("axes", axes); - desc.SetAttr("use_mkldnn", true); + desc.SetAttr("use_onednn", true); auto op = paddle::framework::OpRegistry::CreateOp(desc); @@ -86,7 +86,7 @@ void test_squeeze2() { std::vector axes({-1}); desc.SetAttr("axes", axes); - desc.SetAttr("use_mkldnn", true); + desc.SetAttr("use_onednn", true); auto op = paddle::framework::OpRegistry::CreateOp(desc); diff --git a/test/cpp/fluid/op_debug_string_test.cc b/test/cpp/fluid/op_debug_string_test.cc index 5195a53f5826cf..8d797f97e02f47 100644 --- a/test/cpp/fluid/op_debug_string_test.cc +++ b/test/cpp/fluid/op_debug_string_test.cc @@ -37,7 +37,7 @@ TEST(op_debug_str, test_unknown_dtype) { desc.SetOutput(framework::GradVarName("X"), {framework::GradVarName("X")}); desc.SetOutput(framework::GradVarName("Y"), {framework::GradVarName("Y")}); desc.SetAttr("axis", -1); - desc.SetAttr("use_mkldnn", false); + desc.SetAttr("use_onednn", false); auto x_tensor = scope.Var("X")->GetMutable(); x_tensor->Resize(dim); diff --git a/test/cpp/imperative/test_hooks.cc b/test/cpp/imperative/test_hooks.cc index 521e505b98b894..1350bd31539fd9 100644 --- a/test/cpp/imperative/test_hooks.cc +++ b/test/cpp/imperative/test_hooks.cc @@ -104,7 +104,7 @@ TEST(TestHooks, TestGradVarLeafBackwardHook) { NameVarBaseMap ins = {x_pair, y_pair}; NameVarBaseMap outs = {out_pair}; framework::AttributeMap mul_attr_map; - mul_attr_map["use_mkldnn"] = false; + mul_attr_map["use_onednn"] = false; // add VariableWrapper hook x->GradVarBase()->AddVariableWrapperHook( @@ -211,7 +211,7 @@ void GradVarLeafBackwardHookWithGradAccumulatedTest() { NameVarBaseMap ins = {x_pair, y_pair}; NameVarBaseMap outs = {out_xy_pair}; framework::AttributeMap mul_attr_map; - mul_attr_map["use_mkldnn"] = false; + mul_attr_map["use_onednn"] = false; tracer.TraceOp("mul", ins, outs, mul_attr_map, place, true); var_pair z_pair = var_pair("Y", vb_vector(1, z)); diff --git a/test/cpp/imperative/test_prepare_op.cc b/test/cpp/imperative/test_prepare_op.cc index 1393683e44100e..ae1030abac0e27 100644 --- a/test/cpp/imperative/test_prepare_op.cc +++ b/test/cpp/imperative/test_prepare_op.cc @@ -247,7 +247,7 @@ TEST(test_prepare_op, test_complex_eager) { #ifdef PADDLE_WITH_DNNL TEST(test_prepare_op, test_prepare_data_cpu_onednn) { - TestPrepareDataSamePlace({{"use_mkldnn", true}}); + TestPrepareDataSamePlace({{"use_onednn", true}}); } #endif } // namespace imperative diff --git a/test/cpp/imperative/test_tracer.cc b/test/cpp/imperative/test_tracer.cc index 305334c6a92bb7..ecca7eb41eb441 100644 --- a/test/cpp/imperative/test_tracer.cc +++ b/test/cpp/imperative/test_tracer.cc @@ -89,7 +89,7 @@ TEST(test_tracer, test_trace_op) { imperative::NameVarBaseMap ins = {x_pair, y_pair}; imperative::NameVarBaseMap outs = {out_pair}; framework::AttributeMap mul_attr_map; - mul_attr_map["use_mkldnn"] = false; + mul_attr_map["use_onednn"] = false; tracer.TraceOp("mul", ins, outs, mul_attr_map, place, true); #ifndef PADDLE_WITH_XPU @@ -141,7 +141,7 @@ TEST(test_tracer, test_trace_op_with_backward) { imperative::NameVarBaseMap ins = {x_pair, y_pair}; imperative::NameVarBaseMap outs = {out_pair}; framework::AttributeMap mul_attr_map; - mul_attr_map["use_mkldnn"] = false; + mul_attr_map["use_onednn"] = false; tracer.TraceOp("mul", ins, outs, mul_attr_map, place, true); const auto& out_tensor = vout->Var().Get(); for (int i = 0; i < vout->Var().Get().numel(); i++) { @@ -187,7 +187,7 @@ TEST(test_tracer, test_track_backward_output) { imperative::NameVarBaseMap ins = {x_pair, y_pair}; imperative::NameVarBaseMap outs = {out_pair}; framework::AttributeMap mul_attr_map; - mul_attr_map["use_mkldnn"] = false; + mul_attr_map["use_onednn"] = false; tracer.TraceOp("mul", ins, outs, mul_attr_map, place, true); ASSERT_EQ(x_in->GradVarBase()->GradOpNum(), 0UL); ASSERT_EQ(y_in->GradVarBase()->GradOpNum(), 0UL); @@ -232,7 +232,7 @@ TEST(test_tracer, test_track_backward_input) { imperative::NameVarBaseMap ins = {x_pair, y_pair}; imperative::NameVarBaseMap outs = {out_pair}; framework::AttributeMap mul_attr_map; - mul_attr_map["use_mkldnn"] = false; + mul_attr_map["use_onednn"] = false; tracer.TraceOp("mul", ins, outs, mul_attr_map, place, true); ASSERT_EQ(x_in->GradVarBase()->GradOpNum(), 0UL); @@ -280,7 +280,7 @@ TEST(test_tracer, test_trace_op_with_multi_device_inputs) { imperative::NameVarBaseMap ins = {x_pair, y_pair}; imperative::NameVarBaseMap outs = {out_pair}; framework::AttributeMap mul_attr_map; - mul_attr_map["use_mkldnn"] = false; + mul_attr_map["use_onednn"] = false; tracer.TraceOp( "elementwise_add", ins, outs, mul_attr_map, gpu_place, true); @@ -417,7 +417,7 @@ TEST(test_tracer, test_var_without_grad_var) { imperative::NameVarBaseMap ins = {x_pair, y_pair}; imperative::NameVarBaseMap outs = {out_pair}; framework::AttributeMap mul_attr_map; - mul_attr_map["use_mkldnn"] = false; + mul_attr_map["use_onednn"] = false; tracer.TraceOp("mul", ins, outs, mul_attr_map, place, true); const auto& out_tensor = vout->Var().Get(); @@ -636,7 +636,7 @@ TEST(test_tracer, eager_tracer) { imperative::NameTensorMap ins = {x_pair, y_pair}; imperative::NameTensorMap outs = {out_pair}; framework::AttributeMap mul_attr_map; - mul_attr_map["use_mkldnn"] = false; + mul_attr_map["use_onednn"] = false; tracer.TraceOp( "mul", ins, outs, mul_attr_map, place, true);