Skip to content

Commit 696f645

Browse files
committed
test=release/1.0.0
2 parents 876e77a + 8d16de7 commit 696f645

File tree

15 files changed

+89
-84
lines changed

15 files changed

+89
-84
lines changed

paddle/contrib/float16/float16_transpiler.py

Lines changed: 3 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -102,8 +102,8 @@ def _adjust_input(self, skip=False):
102102
continue
103103
for input_arg in current_op.input_arg_names:
104104
if input_arg in self.input_map:
105-
current_op.rename_input(input_arg,
106-
self.input_map[input_arg])
105+
current_op._rename_input(input_arg,
106+
self.input_map[input_arg])
107107

108108
def _remove_unused_var(self):
109109
'''
@@ -187,7 +187,7 @@ def find_op(var):
187187
shape=var.shape,
188188
persistable=var.persistable)
189189
find_op(var)
190-
var.op.rename_output(var_name, tmp_var_name)
190+
var.op._rename_output(var_name, tmp_var_name)
191191
self.block._insert_op(
192192
i,
193193
type="cast",

paddle/fluid/API.spec

Lines changed: 0 additions & 17 deletions
Original file line numberDiff line numberDiff line change
@@ -6,26 +6,9 @@ paddle.fluid.Program.global_block ArgSpec(args=['self'], varargs=None, keywords=
66
paddle.fluid.Program.list_vars ArgSpec(args=['self'], varargs=None, keywords=None, defaults=None)
77
paddle.fluid.Program.parse_from_string ArgSpec(args=['binary_str'], varargs=None, keywords=None, defaults=None)
88
paddle.fluid.Program.to_string ArgSpec(args=['self', 'throw_on_error', 'with_details'], varargs=None, keywords=None, defaults=(False,))
9-
paddle.fluid.Operator.__init__ ArgSpec(args=['self', 'block', 'desc', 'type', 'inputs', 'outputs', 'attrs'], varargs=None, keywords=None, defaults=(None, None, None, None))
10-
paddle.fluid.Operator.all_attrs ArgSpec(args=['self'], varargs=None, keywords=None, defaults=None)
11-
paddle.fluid.Operator.attr ArgSpec(args=['self', 'name'], varargs=None, keywords=None, defaults=None)
12-
paddle.fluid.Operator.attr_type ArgSpec(args=['self', 'name'], varargs=None, keywords=None, defaults=None)
13-
paddle.fluid.Operator.block_attr ArgSpec(args=['self', 'name'], varargs=None, keywords=None, defaults=None)
14-
paddle.fluid.Operator.block_attr_id ArgSpec(args=['self', 'name'], varargs=None, keywords=None, defaults=None)
15-
paddle.fluid.Operator.blocks_attr ArgSpec(args=['self', 'name'], varargs=None, keywords=None, defaults=None)
16-
paddle.fluid.Operator.blocks_attr_ids ArgSpec(args=['self', 'name'], varargs=None, keywords=None, defaults=None)
17-
paddle.fluid.Operator.has_attr ArgSpec(args=['self', 'name'], varargs=None, keywords=None, defaults=None)
18-
paddle.fluid.Operator.has_kernel ArgSpec(args=['self', 'op_type'], varargs=None, keywords=None, defaults=None)
19-
paddle.fluid.Operator.input ArgSpec(args=['self', 'name'], varargs=None, keywords=None, defaults=None)
20-
paddle.fluid.Operator.output ArgSpec(args=['self', 'name'], varargs=None, keywords=None, defaults=None)
21-
paddle.fluid.Operator.rename_input ArgSpec(args=['self', 'old_name', 'new_name'], varargs=None, keywords=None, defaults=None)
22-
paddle.fluid.Operator.rename_output ArgSpec(args=['self', 'old_name', 'new_name'], varargs=None, keywords=None, defaults=None)
23-
paddle.fluid.Operator.set_attr ArgSpec(args=['self', 'name', 'val'], varargs=None, keywords=None, defaults=None)
24-
paddle.fluid.Operator.to_string ArgSpec(args=['self', 'throw_on_error'], varargs=None, keywords=None, defaults=None)
259
paddle.fluid.default_startup_program ArgSpec(args=[], varargs=None, keywords=None, defaults=None)
2610
paddle.fluid.default_main_program ArgSpec(args=[], varargs=None, keywords=None, defaults=None)
2711
paddle.fluid.program_guard ArgSpec(args=[], varargs='args', keywords='kwds', defaults=None)
28-
paddle.fluid.get_var ArgSpec(args=['name', 'program'], varargs=None, keywords=None, defaults=(None,))
2912
paddle.fluid.name_scope ArgSpec(args=[], varargs='args', keywords='kwds', defaults=None)
3013
paddle.fluid.Executor.__init__ ArgSpec(args=['self', 'place'], varargs=None, keywords=None, defaults=None)
3114
paddle.fluid.Executor.close ArgSpec(args=['self'], varargs=None, keywords=None, defaults=None)

paddle/fluid/framework/ir/graph_traits.cc

Lines changed: 2 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -14,6 +14,8 @@
1414

1515
#include "paddle/fluid/framework/ir/graph_traits.h"
1616

17+
#include <vector>
18+
1719
namespace paddle {
1820
namespace framework {
1921
namespace ir {

paddle/fluid/operators/sequence_slice_op.h

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -75,11 +75,11 @@ class SequenceSliceOpKernel : public framework::OpKernel<T> {
7575
}
7676

7777
for (size_t i = 0; i < n; ++i) {
78-
PADDLE_ENFORCE_LT(0, offset_data[i],
78+
PADDLE_ENFORCE_LE(0, offset_data[i],
7979
"The offset[%d] must greater than zero.", i);
8080
PADDLE_ENFORCE_LT(0, length_data[i],
8181
"The length[%d] must greater than zero.", i);
82-
PADDLE_ENFORCE_LT(lod[0][i] + offset_data[i] + length_data[i],
82+
PADDLE_ENFORCE_LE(lod[0][i] + offset_data[i] + length_data[i],
8383
lod[0][i + 1], "The target tensor's length overflow.");
8484
}
8585

paddle/fluid/pybind/protobuf.cc

Lines changed: 5 additions & 5 deletions
Original file line numberDiff line numberDiff line change
@@ -285,12 +285,12 @@ void BindOpDesc(pybind11::module *m) {
285285
.def("set_output", &pd::OpDesc::SetOutput)
286286
.def("input_arg_names", &pd::OpDesc::InputArgumentNames)
287287
.def("output_arg_names", &pd::OpDesc::OutputArgumentNames)
288-
.def("rename_input", &pd::OpDesc::RenameInput)
289-
.def("rename_output", &pd::OpDesc::RenameOutput)
288+
.def("_rename_input", &pd::OpDesc::RenameInput)
289+
.def("_rename_output", &pd::OpDesc::RenameOutput)
290290
.def("has_attr", &pd::OpDesc::HasAttr)
291291
.def("attr_type", &pd::OpDesc::GetAttrType)
292292
.def("attr_names", &pd::OpDesc::AttrNames)
293-
.def("set_attr", &pd::OpDesc::SetAttr)
293+
.def("_set_attr", &pd::OpDesc::SetAttr)
294294
.def("attr", &pd::OpDesc::GetAttr)
295295
.def("set_block_attr", &pd::OpDesc::SetBlockAttr)
296296
.def("set_blocks_attr", &pd::OpDesc::SetBlocksAttr)
@@ -300,8 +300,8 @@ void BindOpDesc(pybind11::module *m) {
300300
std::string ser(seriralized);
301301
self.SetAttr(name, ser);
302302
})
303-
.def("block_attr_id", &pd::OpDesc::GetBlockAttrId)
304-
.def("blocks_attr_ids", &pd::OpDesc::GetBlocksAttrIds)
303+
.def("_block_attr_id", &pd::OpDesc::GetBlockAttrId)
304+
.def("_blocks_attr_ids", &pd::OpDesc::GetBlocksAttrIds)
305305
.def("check_attrs", &pd::OpDesc::CheckAttrs)
306306
.def("infer_shape", &pd::OpDesc::InferShape)
307307
.def("infer_var_type", &pd::OpDesc::InferVarType)

python/paddle/fluid/backward.py

Lines changed: 12 additions & 12 deletions
Original file line numberDiff line numberDiff line change
@@ -38,8 +38,8 @@ def _rename_arg_(op_descs, old_name, new_name, begin_idx=None, end_idx=None):
3838
op_desc = op_descs[i]
3939
if isinstance(op_desc, tuple):
4040
op_desc = op_desc[0]
41-
op_desc.rename_input(old_name, new_name)
42-
op_desc.rename_output(old_name, new_name)
41+
op_desc._rename_input(old_name, new_name)
42+
op_desc._rename_output(old_name, new_name)
4343

4444

4545
def _create_op_desc_(op_type, inputs, outputs, attrs):
@@ -70,7 +70,7 @@ def _create_op_desc_(op_type, inputs, outputs, attrs):
7070
if isinstance(val, framework.Block):
7171
op_desc.set_block_attr(name, val.desc)
7272
else:
73-
op_desc.set_attr(name, val)
73+
op_desc._set_attr(name, val)
7474
return op_desc
7575

7676

@@ -346,7 +346,7 @@ def _append_backward_ops_(block,
346346
grad_sub_block_list = []
347347
# If the op has its own sub-block, deal with the sub-block first
348348
if op.has_attr("sub_block"):
349-
sub_block = program.block(op.block_attr_id("sub_block"))
349+
sub_block = program.block(op._block_attr_id("sub_block"))
350350
grad_sub_block = program._create_block()
351351
grad_sub_block._set_forward_block_idx(sub_block.idx)
352352
cb = _callback_lookup_(op)
@@ -382,7 +382,7 @@ def _append_backward_ops_(block,
382382
for op_desc in grad_op_descs:
383383
new_op_desc = target_block.desc.append_op()
384384
new_op_desc.copy_from(op_desc)
385-
new_op_desc.set_attr(op_role_attr_name, backward)
385+
new_op_desc._set_attr(op_role_attr_name, backward)
386386
grad_to_var["__current_op_desc__"] = new_op_desc
387387
if callbacks is not None:
388388
assert (isinstance(callbacks, list))
@@ -408,7 +408,7 @@ def _append_backward_vars_(block, start_op_idx, grad_to_var, grad_info_map):
408408
for op_idx in range(start_op_idx, block.desc.op_size()):
409409
op_desc = block.desc.op(op_idx)
410410
if op_desc.has_attr("sub_block"):
411-
sub_block = block.program.block(op_desc.block_attr_id("sub_block"))
411+
sub_block = block.program.block(op_desc._block_attr_id("sub_block"))
412412
_append_backward_vars_(sub_block, 0, grad_to_var, grad_info_map)
413413
new_vars = set()
414414
# create new gradient variables
@@ -438,12 +438,12 @@ def _rename_grad_(block, start_op_idx, grad_to_var, target_grad_map):
438438
op_desc = block.desc.op(op_idx)
439439
for name in op_desc.input_arg_names():
440440
if name in var_map:
441-
op_desc.rename_input(name, var_map[name])
441+
op_desc._rename_input(name, var_map[name])
442442

443443
for name in op_desc.output_arg_names():
444444
if block.desc.find_var(name.encode("ascii")):
445445
new_name = unique_name.generate(name)
446-
op_desc.rename_output(name, new_name)
446+
op_desc._rename_output(name, new_name)
447447
var_map[name] = new_name
448448

449449
for g, ng in six.iteritems(var_map):
@@ -542,9 +542,9 @@ def append_backward(loss, parameter_list=None, no_grad_set=None,
542542
if loss.op is None:
543543
raise ValueError("loss.op is None. Should not happend")
544544

545-
loss.op.set_attr(core.op_proto_and_checker_maker.kOpRoleAttrName(),
546-
int(core.op_proto_and_checker_maker.OpRole.Forward) |
547-
int(core.op_proto_and_checker_maker.OpRole.Loss))
545+
loss.op._set_attr(core.op_proto_and_checker_maker.kOpRoleAttrName(),
546+
int(core.op_proto_and_checker_maker.OpRole.Forward) |
547+
int(core.op_proto_and_checker_maker.OpRole.Loss))
548548

549549
if callbacks is not None:
550550
isinstance(callbacks, list)
@@ -631,7 +631,7 @@ def append_backward(loss, parameter_list=None, no_grad_set=None,
631631
attr_val = [p.name, g.name]
632632
if g.op.has_attr(op_role_var_attr_name):
633633
attr_val.extend(g.op.attr(op_role_var_attr_name))
634-
g.op.set_attr(op_role_var_attr_name, attr_val)
634+
g.op._set_attr(op_role_var_attr_name, attr_val)
635635

636636
return params_and_grads
637637

python/paddle/fluid/clip.py

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -75,8 +75,8 @@ def _append_clip_op(self, block, grad_name):
7575
clip_op_desc.set_type("clip")
7676
clip_op_desc.set_input("X", [grad_name])
7777
clip_op_desc.set_output("Out", [grad_name])
78-
clip_op_desc.set_attr("min", self.min)
79-
clip_op_desc.set_attr("max", self.max)
78+
clip_op_desc._set_attr("min", self.min)
79+
clip_op_desc._set_attr("max", self.max)
8080

8181

8282
def error_clip_callback(block, context):

python/paddle/fluid/framework.py

Lines changed: 20 additions & 22 deletions
Original file line numberDiff line numberDiff line change
@@ -37,11 +37,9 @@
3737

3838
__all__ = [
3939
'Program',
40-
'Operator',
4140
'default_startup_program',
4241
'default_main_program',
4342
'program_guard',
44-
'get_var',
4543
'name_scope',
4644
]
4745

@@ -654,11 +652,11 @@ def find_name(var_list, name):
654652
self._update_desc_attr(attr_name, attr_val)
655653

656654
self.desc.check_attrs()
657-
if self.has_kernel(type):
655+
if self._has_kernel(type):
658656
self.desc.infer_var_type(self.block.desc)
659657
self.desc.infer_shape(self.block.desc)
660658

661-
def has_kernel(self, op_type):
659+
def _has_kernel(self, op_type):
662660
return op_type not in self.OP_WITHOUT_KERNEL_SET
663661

664662
def to_string(self, throw_on_error):
@@ -699,7 +697,7 @@ def input(self, name):
699697
"""
700698
return self.desc.input(name)
701699

702-
def rename_input(self, old_name, new_name):
700+
def _rename_input(self, old_name, new_name):
703701
"""
704702
Rename the `old_name` to `new_name`.
705703
@@ -710,9 +708,9 @@ def rename_input(self, old_name, new_name):
710708
Returns:
711709
None
712710
"""
713-
self.desc.rename_input(old_name, new_name)
711+
self.desc._rename_input(old_name, new_name)
714712

715-
def rename_output(self, old_name, new_name):
713+
def _rename_output(self, old_name, new_name):
716714
"""
717715
Rename the `old_name` to `new_name`.
718716
@@ -723,7 +721,7 @@ def rename_output(self, old_name, new_name):
723721
Returns:
724722
None
725723
"""
726-
self.desc.rename_output(old_name, new_name)
724+
self.desc._rename_output(old_name, new_name)
727725

728726
@property
729727
def input_names(self):
@@ -787,7 +785,7 @@ def attr_type(self, name):
787785
"""
788786
return self.desc.attr_type(name)
789787

790-
def set_attr(self, name, val):
788+
def _set_attr(self, name, val):
791789
"""
792790
Set the value of attribute by attribute's name.
793791
@@ -820,7 +818,7 @@ def _update_desc_attr(self, name, val):
820818
isinstance(val, core.ProgramDesc):
821819
self.desc.set_serialized_attr(name, val.serialize_to_string())
822820
else:
823-
self.desc.set_attr(name, val)
821+
self.desc._set_attr(name, val)
824822

825823
@property
826824
def attr_names(self):
@@ -839,7 +837,7 @@ def attr(self, name):
839837
"""
840838
return self.desc.attr(name)
841839

842-
def block_attr_id(self, name):
840+
def _block_attr_id(self, name):
843841
"""
844842
Get the block attribute's id by name.
845843
@@ -849,9 +847,9 @@ def block_attr_id(self, name):
849847
Returns:
850848
int: the block index.
851849
"""
852-
return self.desc.block_attr_id(name)
850+
return self.desc._block_attr_id(name)
853851

854-
def block_attr(self, name):
852+
def _block_attr(self, name):
855853
"""
856854
Get the block attribute by name.
857855
@@ -862,11 +860,11 @@ def block_attr(self, name):
862860
block: the block attribute.
863861
"""
864862

865-
id = self.block_attr_id(name)
863+
id = self._block_attr_id(name)
866864
assert (id >= 0 and id < len(self.block.program.blocks))
867865
return self.block.program.blocks[id]
868866

869-
def blocks_attr(self, name):
867+
def _blocks_attr(self, name):
870868
"""
871869
Get the blocks attribute by name.
872870
@@ -877,13 +875,13 @@ def blocks_attr(self, name):
877875
list: list of the blocks attribute.
878876
"""
879877
attrs = []
880-
for i in self.blocks_attr_ids(name):
878+
for i in self._blocks_attr_ids(name):
881879
assert (i >= 0 and i < len(self.block.program.blocks))
882880
attrs.append(self.block.program.blocks[i])
883881

884882
return attrs
885883

886-
def blocks_attr_ids(self, name):
884+
def _blocks_attr_ids(self, name):
887885
"""
888886
Get the blocks attribute's ids by name.
889887
@@ -894,7 +892,7 @@ def blocks_attr_ids(self, name):
894892
list: list of the blocks ids.
895893
"""
896894

897-
return self.desc.blocks_attr_ids(name)
895+
return self.desc._blocks_attr_ids(name)
898896

899897
def all_attrs(self):
900898
"""
@@ -908,11 +906,11 @@ def all_attrs(self):
908906
for n in attr_names:
909907
attr_type = self.desc.attr_type(n)
910908
if attr_type == core.AttrType.BLOCK:
911-
attr_map[n] = self.block_attr(n)
909+
attr_map[n] = self._block_attr(n)
912910
continue
913911

914912
if attr_type == core.AttrType.BLOCKS:
915-
attr_map[n] = self.blocks_attr(n)
913+
attr_map[n] = self._blocks_attr(n)
916914
continue
917915

918916
attr_map[n] = self.attr(n)
@@ -1786,7 +1784,7 @@ def _inference_optimize(self, prune_read_op=True):
17861784
for j in six.moves.range(block.op_size()):
17871785
op = block.op(j)
17881786
if op.has_attr('is_test'):
1789-
op.set_attr('is_test', True)
1787+
op._set_attr('is_test', True)
17901788
res.blocks = [
17911789
Block(res, i) for i in six.moves.range(res.desc.num_blocks())
17921790
]
@@ -2160,7 +2158,7 @@ def program_guard(main_program, startup_program=None):
21602158
switch_startup_program(startup_program)
21612159

21622160

2163-
def get_var(name, program=None):
2161+
def _get_var(name, program=None):
21642162
"""
21652163
Get a variable by name from the global block of a program.
21662164

python/paddle/fluid/tests/unittests/dist_transformer.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -1488,7 +1488,7 @@ def wrap_decoder(trg_vocab_size,
14881488
if weight_sharing:
14891489
predict = layers.matmul(
14901490
x=dec_output,
1491-
y=fluid.get_var(word_emb_param_names[0]),
1491+
y=fluid.framework._get_var(word_emb_param_names[0]),
14921492
transpose_y=True)
14931493
else:
14941494
predict = layers.fc(input=dec_output,

python/paddle/fluid/tests/unittests/test_dist_transpiler.py

Lines changed: 19 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -264,6 +264,25 @@ def transpiler_test_impl(self):
264264
])
265265

266266

267+
class TestDecayedAdagrad(TranspilerTest):
268+
def net_conf(self):
269+
x = fluid.layers.data(name='x', shape=[1000], dtype='float32')
270+
y_predict = fluid.layers.fc(input=x,
271+
size=1000,
272+
act=None,
273+
param_attr=fluid.ParamAttr(name='fc_w'),
274+
bias_attr=fluid.ParamAttr(name='fc_b'))
275+
y = fluid.layers.data(name='y', shape=[1], dtype='float32')
276+
cost = fluid.layers.square_error_cost(input=y_predict, label=y)
277+
avg_cost = fluid.layers.mean(cost)
278+
opt = fluid.optimizer.DecayedAdagrad(learning_rate=0.1)
279+
opt.minimize(avg_cost)
280+
281+
def transpiler_test_impl(self):
282+
pserver, startup = self.get_pserver(self.pserver1_ep)
283+
trainer, _ = self.get_trainer()
284+
285+
267286
class TestLRDecayConditional(TranspilerTest):
268287
def net_conf(self):
269288
x = fluid.layers.data(name='x', shape=[1000], dtype='float32')

0 commit comments

Comments
 (0)