Skip to content

Commit 7d8f639

Browse files
Merge pull request #15902 from colourful-tree/new_develop
remove mkldnn & fix commit
2 parents effec86 + 08c96d1 commit 7d8f639

File tree

4 files changed

+8
-13
lines changed

4 files changed

+8
-13
lines changed

paddle/fluid/API.spec

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -92,7 +92,7 @@ paddle.fluid.layers.pool3d ArgSpec(args=['input', 'pool_size', 'pool_type', 'poo
9292
paddle.fluid.layers.adaptive_pool2d ArgSpec(args=['input', 'pool_size', 'pool_type', 'require_index', 'name'], varargs=None, keywords=None, defaults=('max', False, None))
9393
paddle.fluid.layers.adaptive_pool3d ArgSpec(args=['input', 'pool_size', 'pool_type', 'require_index', 'name'], varargs=None, keywords=None, defaults=('max', False, None))
9494
paddle.fluid.layers.batch_norm ArgSpec(args=['input', 'act', 'is_test', 'momentum', 'epsilon', 'param_attr', 'bias_attr', 'data_layout', 'in_place', 'name', 'moving_mean_name', 'moving_variance_name', 'do_model_average_for_mean_and_var', 'fuse_with_relu', 'use_global_stats'], varargs=None, keywords=None, defaults=(None, False, 0.9, 1e-05, None, None, 'NCHW', False, None, None, None, False, False, False))
95-
paddle.fluid.layers.data_norm ArgSpec(args=['input', 'act', 'epsilon', 'param_attr', 'data_layout', 'in_place', 'use_mkldnn', 'name', 'moving_mean_name', 'moving_variance_name', 'do_model_average_for_mean_and_var'], varargs=None, keywords=None, defaults=(None, 1e-05, None, 'NCHW', False, False, None, None, None, False))
95+
paddle.fluid.layers.data_norm ArgSpec(args=['input', 'act', 'epsilon', 'param_attr', 'data_layout', 'in_place', 'name', 'moving_mean_name', 'moving_variance_name', 'do_model_average_for_mean_and_var'], varargs=None, keywords=None, defaults=(None, 1e-05, None, 'NCHW', False, None, None, None, False))
9696
paddle.fluid.layers.beam_search_decode ArgSpec(args=['ids', 'scores', 'beam_size', 'end_id', 'name'], varargs=None, keywords=None, defaults=(None,))
9797
paddle.fluid.layers.conv2d_transpose ArgSpec(args=['input', 'num_filters', 'output_size', 'filter_size', 'padding', 'stride', 'dilation', 'groups', 'param_attr', 'bias_attr', 'use_cudnn', 'act', 'name'], varargs=None, keywords=None, defaults=(None, None, 0, 1, 1, None, None, None, True, None, None))
9898
paddle.fluid.layers.conv3d_transpose ArgSpec(args=['input', 'num_filters', 'output_size', 'filter_size', 'padding', 'stride', 'dilation', 'groups', 'param_attr', 'bias_attr', 'use_cudnn', 'act', 'name'], varargs=None, keywords=None, defaults=(None, None, 0, 1, 1, None, None, None, True, None, None))

paddle/fluid/operators/data_norm_op.cc

Lines changed: 0 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -140,9 +140,6 @@ class DataNormOpMaker : public framework::OpProtoAndCheckerMaker {
140140
"Scales of the history data batch, "
141141
"will apply to output when training")
142142
.AsIntermediate();
143-
AddAttr<bool>("use_mkldnn",
144-
"(bool, default false) Only used in mkldnn kernel")
145-
.SetDefault(false);
146143
AddComment(R"DOC(
147144
Data Normalization.
148145

paddle/fluid/operators/teacher_student_sigmoid_loss_op.cc

Lines changed: 5 additions & 5 deletions
Original file line numberDiff line numberDiff line change
@@ -117,11 +117,11 @@ class TeacherStudentSigmoidLossOpMaker
117117
"[N x 1]. The teacher student sigmoid loss.");
118118
AddAttr<float>(
119119
"soft_max_up_bound",
120-
"fp32, if input > soft_max_up_bound, will be bound, default 15.0")
120+
"fp32, if input > soft_max_up_bound, input will be bound, default 15.0")
121121
.SetDefault(15.0);
122-
AddAttr<float>(
123-
"soft_max_lower_bound",
124-
"fp32, if input < soft_max_lower_bound, will be bound, default -15.0")
122+
AddAttr<float>("soft_max_lower_bound",
123+
"fp32, if input < soft_max_lower_bound, input will be "
124+
"bound, default -15.0")
125125
.SetDefault(-15.0);
126126
AddComment(R"DOC(
127127
TeacherStudentSigmoidLoss Operator.
@@ -134,7 +134,7 @@ we add another label(z') to original.
134134
label = {-2, -1, [0, 2]}
135135
when z' is not exist, clk = 0 : label = -2;
136136
when z' is not exist, clk = 1 : label = -1;
137-
when z' is exist , clk = 0 : label = 0 + z';
137+
when z' is exist , clk = 0 : label = 0 + z';
138138
when z' is exist , clk = 1 : label = 1 + z';
139139
140140
)DOC");

python/paddle/fluid/layers/nn.py

Lines changed: 2 additions & 4 deletions
Original file line numberDiff line numberDiff line change
@@ -3041,7 +3041,6 @@ def data_norm(input,
30413041
param_attr=None,
30423042
data_layout='NCHW',
30433043
in_place=False,
3044-
use_mkldnn=False,
30453044
name=None,
30463045
moving_mean_name=None,
30473046
moving_variance_name=None,
@@ -3075,7 +3074,6 @@ def data_norm(input,
30753074
param_attr(ParamAttr): The parameter attribute for Parameter `scale`.
30763075
data_layout(string, default NCHW): NCHW|NHWC
30773076
in_place(bool, Default False): Make the input and output of batch norm reuse memory.
3078-
use_mkldnn(bool, Default false): ${use_mkldnn_comment}
30793077
name(string, Default None): A name for this layer(optional). If set None, the layer
30803078
will be named automatically.
30813079
moving_mean_name(string, Default None): The name of moving_mean which store the global Mean.
@@ -3156,8 +3154,7 @@ def data_norm(input,
31563154
outputs={"Y": data_norm_out,
31573155
"Means": means,
31583156
"Scales": scales},
3159-
attrs={"epsilon": epsilon,
3160-
"use_mkldnn": use_mkldnn})
3157+
attrs={"epsilon": epsilon})
31613158

31623159
return helper.append_activation(data_norm_out)
31633160

@@ -9945,6 +9942,7 @@ def teacher_student_sigmoid_loss(input,
99459942
99469943
Examples:
99479944
.. code-block:: python
9945+
99489946
cost = fluid.layers.teacher_student_sigmoid_loss(input=similarity, label=label)
99499947
"""
99509948
helper = LayerHelper('teacher_student_sigmoid_loss', **locals())

0 commit comments

Comments
 (0)