Skip to content

Commit b57a2b9

Browse files
committed
remove test_seq_concat_op
1 parent 2845195 commit b57a2b9

File tree

3 files changed

+35
-162
lines changed

3 files changed

+35
-162
lines changed

paddle/fluid/API.spec

Lines changed: 6 additions & 5 deletions
Original file line numberDiff line numberDiff line change
@@ -61,11 +61,11 @@ paddle.fluid.DistributeTranspiler.get_pserver_programs ArgSpec(args=['self', 'en
6161
paddle.fluid.DistributeTranspiler.get_startup_program ArgSpec(args=['self', 'endpoint', 'pserver_program', 'startup_program'], varargs=None, keywords=None, defaults=(None, None))
6262
paddle.fluid.DistributeTranspiler.get_trainer_program ArgSpec(args=['self', 'wait_port'], varargs=None, keywords=None, defaults=(True,))
6363
paddle.fluid.DistributeTranspiler.transpile ArgSpec(args=['self', 'trainer_id', 'program', 'pservers', 'trainers', 'sync_mode', 'startup_program'], varargs=None, keywords=None, defaults=(None, '127.0.0.1:6174', 1, True, None))
64-
paddle.fluid.InferenceTranspiler.__init__
64+
paddle.fluid.InferenceTranspiler.__init__
6565
paddle.fluid.InferenceTranspiler.transpile ArgSpec(args=['self', 'program', 'place', 'scope'], varargs=None, keywords=None, defaults=(None,))
6666
paddle.fluid.memory_optimize ArgSpec(args=['input_program', 'skip_opt_set', 'print_log', 'level'], varargs=None, keywords=None, defaults=(None, False, 0))
6767
paddle.fluid.release_memory ArgSpec(args=['input_program', 'skip_opt_set'], varargs=None, keywords=None, defaults=(None,))
68-
paddle.fluid.DistributeTranspilerConfig.__init__
68+
paddle.fluid.DistributeTranspilerConfig.__init__
6969
paddle.fluid.ParallelExecutor.__init__ ArgSpec(args=['self', 'use_cuda', 'loss_name', 'main_program', 'share_vars_from', 'exec_strategy', 'build_strategy', 'num_trainers', 'trainer_id', 'scope'], varargs=None, keywords='kwargs', defaults=(None, None, None, None, None, 1, 0, None))
7070
paddle.fluid.ParallelExecutor.run ArgSpec(args=['self', 'fetch_list', 'feed', 'feed_dict', 'return_numpy'], varargs=None, keywords=None, defaults=(None, None, True))
7171
paddle.fluid.ExecutionStrategy.__init__ __init__(self: paddle.fluid.core.ExecutionStrategy) -> None
@@ -174,6 +174,7 @@ paddle.fluid.layers.stack ArgSpec(args=['x', 'axis'], varargs=None, keywords=Non
174174
paddle.fluid.layers.pad2d ArgSpec(args=['input', 'paddings', 'mode', 'pad_value', 'data_format', 'name'], varargs=None, keywords=None, defaults=([0, 0, 0, 0], 'constant', 0.0, 'NCHW', None))
175175
paddle.fluid.layers.unstack ArgSpec(args=['x', 'axis', 'num'], varargs=None, keywords=None, defaults=(0, None))
176176
paddle.fluid.layers.sequence_enumerate ArgSpec(args=['input', 'win_size', 'pad_value', 'name'], varargs=None, keywords=None, defaults=(0, None))
177+
paddle.fluid.layers.sequence_concat ArgSpec(args=['input', 'name'], varargs=None, keywords=None, defaults=(None,))
177178
paddle.fluid.layers.data ArgSpec(args=['name', 'shape', 'append_batch_size', 'dtype', 'lod_level', 'type', 'stop_gradient'], varargs=None, keywords=None, defaults=(True, 'float32', 0, VarType.LOD_TENSOR, True))
178179
paddle.fluid.layers.open_recordio_file ArgSpec(args=['filename', 'shapes', 'lod_levels', 'dtypes', 'pass_num', 'for_parallel'], varargs=None, keywords=None, defaults=(1, True))
179180
paddle.fluid.layers.open_files ArgSpec(args=['filenames', 'shapes', 'lod_levels', 'dtypes', 'thread_num', 'buffer_size', 'pass_num', 'is_test'], varargs=None, keywords=None, defaults=(None, None, 1, None))
@@ -348,7 +349,7 @@ paddle.fluid.transpiler.DistributeTranspiler.get_pserver_programs ArgSpec(args=[
348349
paddle.fluid.transpiler.DistributeTranspiler.get_startup_program ArgSpec(args=['self', 'endpoint', 'pserver_program', 'startup_program'], varargs=None, keywords=None, defaults=(None, None))
349350
paddle.fluid.transpiler.DistributeTranspiler.get_trainer_program ArgSpec(args=['self', 'wait_port'], varargs=None, keywords=None, defaults=(True,))
350351
paddle.fluid.transpiler.DistributeTranspiler.transpile ArgSpec(args=['self', 'trainer_id', 'program', 'pservers', 'trainers', 'sync_mode', 'startup_program'], varargs=None, keywords=None, defaults=(None, '127.0.0.1:6174', 1, True, None))
351-
paddle.fluid.transpiler.InferenceTranspiler.__init__
352+
paddle.fluid.transpiler.InferenceTranspiler.__init__
352353
paddle.fluid.transpiler.InferenceTranspiler.transpile ArgSpec(args=['self', 'program', 'place', 'scope'], varargs=None, keywords=None, defaults=(None,))
353354
paddle.fluid.transpiler.memory_optimize ArgSpec(args=['input_program', 'skip_opt_set', 'print_log', 'level'], varargs=None, keywords=None, defaults=(None, False, 0))
354355
paddle.fluid.transpiler.release_memory ArgSpec(args=['input_program', 'skip_opt_set'], varargs=None, keywords=None, defaults=(None,))
@@ -358,7 +359,7 @@ paddle.fluid.transpiler.HashName.reset ArgSpec(args=['self'], varargs=None, keyw
358359
paddle.fluid.transpiler.RoundRobin.__init__ ArgSpec(args=['self', 'pserver_endpoints'], varargs=None, keywords=None, defaults=None)
359360
paddle.fluid.transpiler.RoundRobin.dispatch ArgSpec(args=['self', 'varlist'], varargs=None, keywords=None, defaults=None)
360361
paddle.fluid.transpiler.RoundRobin.reset ArgSpec(args=['self'], varargs=None, keywords=None, defaults=None)
361-
paddle.fluid.transpiler.DistributeTranspilerConfig.__init__
362+
paddle.fluid.transpiler.DistributeTranspilerConfig.__init__
362363
paddle.fluid.nets.simple_img_conv_pool ArgSpec(args=['input', 'num_filters', 'filter_size', 'pool_size', 'pool_stride', 'pool_padding', 'pool_type', 'global_pooling', 'conv_stride', 'conv_padding', 'conv_dilation', 'conv_groups', 'param_attr', 'bias_attr', 'act', 'use_cudnn', 'use_mkldnn'], varargs=None, keywords=None, defaults=(0, 'max', False, 1, 0, 1, 1, None, None, None, True, False))
363364
paddle.fluid.nets.sequence_conv_pool ArgSpec(args=['input', 'num_filters', 'filter_size', 'param_attr', 'act', 'pool_type'], varargs=None, keywords=None, defaults=(None, 'sigmoid', 'max'))
364365
paddle.fluid.nets.glu ArgSpec(args=['input', 'dim'], varargs=None, keywords=None, defaults=(-1,))
@@ -425,4 +426,4 @@ paddle.fluid.Scope.__init__ __init__(self: paddle.fluid.core.Scope) -> None
425426
paddle.fluid.Scope.drop_kids drop_kids(self: paddle.fluid.core.Scope) -> None
426427
paddle.fluid.Scope.find_var find_var(self: paddle.fluid.core.Scope, arg0: unicode) -> paddle.fluid.core.Variable
427428
paddle.fluid.Scope.new_scope new_scope(self: paddle.fluid.core.Scope) -> paddle.fluid.core.Scope
428-
paddle.fluid.Scope.var var(self: paddle.fluid.core.Scope, arg0: unicode) -> paddle.fluid.core.Variable
429+
paddle.fluid.Scope.var var(self: paddle.fluid.core.Scope, arg0: unicode) -> paddle.fluid.core.Variable

paddle/fluid/operators/sequence_concat_op.cc

Lines changed: 29 additions & 33 deletions
Original file line numberDiff line numberDiff line change
@@ -37,42 +37,38 @@ class SeqConcatOpMaker : public framework::OpProtoAndCheckerMaker {
3737
class SeqConcatShapeInferer : public framework::InferShapeBase {
3838
public:
3939
void operator()(framework::InferShapeContext *context) const override {
40-
try {
41-
PADDLE_ENFORCE(context->HasInputs("X"),
42-
"Input(X) of Sequence Concat Op should not be null.");
43-
PADDLE_ENFORCE(context->HasOutput("Out"),
44-
"Output(Out) of Sequence Concat Op should not be null.");
40+
PADDLE_ENFORCE(context->HasInputs("X"),
41+
"Input(X) of Sequence Concat Op should not be null.");
42+
PADDLE_ENFORCE(context->HasOutput("Out"),
43+
"Output(Out) of Sequence Concat Op should not be null.");
4544

46-
PADDLE_ENFORCE_GT(context->HasInputs("X"), 1,
47-
"The number of input sequences is at least two.");
48-
auto x_dims = context->GetInputsDim("X");
49-
int64_t batch_size = 0;
50-
int64_t feature_size = 0;
51-
std::vector<int64_t> out_dims;
52-
for (auto &x_dim : x_dims) {
53-
if (out_dims.empty()) {
54-
out_dims = framework::vectorize(x_dim);
55-
}
56-
batch_size += x_dim[0];
57-
if (feature_size == 0) {
58-
feature_size = framework::product(x_dim) / x_dim[0];
59-
} else {
60-
PADDLE_ENFORCE_EQ(
61-
feature_size, framework::product(x_dim) / x_dim[0],
62-
"Inputs of sequence concat must have same feature size");
63-
}
45+
PADDLE_ENFORCE_GT(context->Inputs("X").size(), 1,
46+
"The number of input sequences is at least two.");
47+
auto x_dims = context->GetInputsDim("X");
48+
int64_t batch_size = 0;
49+
int64_t feature_size = 0;
50+
std::vector<int64_t> out_dims;
51+
for (auto &x_dim : x_dims) {
52+
if (out_dims.empty()) {
53+
out_dims = framework::vectorize(x_dim);
6454
}
65-
if (batch_size < 0) {
66-
batch_size = -1; // Normalize batch size for compile time.
55+
batch_size += x_dim[0];
56+
if (feature_size == 0) {
57+
feature_size = framework::product(x_dim) / x_dim[0];
58+
} else {
59+
PADDLE_ENFORCE_EQ(
60+
feature_size, framework::product(x_dim) / x_dim[0],
61+
"Inputs of sequence concat must have same feature size");
6762
}
68-
out_dims[0] = batch_size;
69-
context->SetOutputDim("Out", framework::make_ddim(out_dims));
70-
if (!context->IsRuntime()) { // Runtime LoD infershape will be computed
71-
// in Kernel.
72-
context->ShareLoD("X", "Out");
73-
}
74-
} catch (...) {
75-
PADDLE_THROW("Unknown error");
63+
}
64+
if (batch_size < 0) {
65+
batch_size = -1; // Normalize batch size for compile time.
66+
}
67+
out_dims[0] = batch_size;
68+
context->SetOutputDim("Out", framework::make_ddim(out_dims));
69+
if (!context->IsRuntime()) { // Runtime LoD infershape will be computed
70+
// in Kernel.
71+
context->ShareLoD("X", "Out");
7672
}
7773
}
7874
};

python/paddle/fluid/tests/unittests/test_seq_concat_op.py

Lines changed: 0 additions & 124 deletions
This file was deleted.

0 commit comments

Comments
 (0)