Skip to content

Commit 6a31509

Browse files
committed
Port release 0.15.0 code to python3
1 parent 8f2ce7c commit 6a31509

File tree

2 files changed

+25
-19
lines changed

2 files changed

+25
-19
lines changed

python/paddle/fluid/tests/unittests/dist_transformer.py

Lines changed: 20 additions & 14 deletions
Original file line numberDiff line numberDiff line change
@@ -36,6 +36,7 @@
3636
import paddle.fluid.layers as layers
3737
from paddle.fluid import core
3838
from test_dist_base import TestDistRunnerBase, runtime_main
39+
import paddle.compat as cpt
3940
from paddle.compat import long_type
4041

4142
import hashlib
@@ -315,8 +316,9 @@ def pad_batch_data(insts,
315316
"""
316317
return_list = []
317318
max_len = max(len(inst) for inst in insts)
318-
num_token = reduce(lambda x, y: x + y,
319-
[len(inst) for inst in insts]) if return_num_token else 0
319+
num_token = six.moves.reduce(
320+
lambda x, y: x + y,
321+
[len(inst) for inst in insts]) if return_num_token else 0
320322
# Any token included in dict can be used to pad, since the paddings' loss
321323
# will be masked out by weights and make no effect on parameter gradients.
322324
inst_data = np.array(
@@ -328,7 +330,7 @@ def pad_batch_data(insts,
328330
return_list += [inst_weight.astype("float32").reshape([-1, 1])]
329331
else: # position data
330332
inst_pos = np.array([
331-
range(1, len(inst) + 1) + [0] * (max_len - len(inst))
333+
list(range(1, len(inst) + 1)) + [0] * (max_len - len(inst))
332334
for inst in insts
333335
])
334336
return_list += [inst_pos.astype("int64").reshape([-1, 1])]
@@ -385,10 +387,11 @@ def prepare_batch_input(insts, data_input_names, src_pad_idx, trg_pad_idx,
385387
return_num_token=True)
386388

387389
data_input_dict = dict(
388-
zip(data_input_names, [
389-
src_word, src_pos, src_slf_attn_bias, trg_word, trg_pos,
390-
trg_slf_attn_bias, trg_src_attn_bias, lbl_word, lbl_weight
391-
]))
390+
list(
391+
zip(data_input_names, [
392+
src_word, src_pos, src_slf_attn_bias, trg_word, trg_pos,
393+
trg_slf_attn_bias, trg_src_attn_bias, lbl_word, lbl_weight
394+
])))
392395
return data_input_dict, np.asarray([num_token], dtype="float32")
393396

394397

@@ -561,7 +564,7 @@ def train_loop(exe, train_progm, dev_count, sum_cost, avg_cost, lr_scheduler,
561564
np.log(TrainTaskConfig.label_smooth_eps / (
562565
ModelHyperParams.trg_vocab_size - 1) + 1e-20))
563566
init = False
564-
for pass_id in xrange(TrainTaskConfig.pass_num):
567+
for pass_id in six.moves.xrange(TrainTaskConfig.pass_num):
565568
pass_start_time = time.time()
566569
for batch_id, data in enumerate(train_data()):
567570
if batch_id >= 5:
@@ -587,11 +590,11 @@ def train_loop(exe, train_progm, dev_count, sum_cost, avg_cost, lr_scheduler,
587590
ModelHyperParams.eos_idx, ModelHyperParams.n_head,
588591
ModelHyperParams.d_model)
589592
total_num_token += num_token
590-
feed_kv_pairs = data_input_dict.items()
593+
feed_kv_pairs = list(data_input_dict.items())
591594
if TrainTaskConfig.local:
592-
feed_kv_pairs += {
595+
feed_kv_pairs += list({
593596
lr_scheduler.learning_rate.name: lr_rate
594-
}.items()
597+
}.items())
595598
feed_list.append(dict(feed_kv_pairs))
596599

597600
if not init:
@@ -873,6 +876,7 @@ def _load_lines(self, fpattern, tar_fname):
873876

874877
f = tarfile.open(fpaths[0], "r")
875878
for line in f.extractfile(tar_fname):
879+
line = cpt.to_text(line)
876880
fields = line.strip("\n").split(self._field_delimiter)
877881
if (not self._only_src and len(fields) == 2) or (
878882
self._only_src and len(fields) == 1):
@@ -882,8 +886,9 @@ def _load_lines(self, fpattern, tar_fname):
882886
if not os.path.isfile(fpath):
883887
raise IOError("Invalid file: %s" % fpath)
884888

885-
with open(fpath, "r") as f:
889+
with open(fpath, "rb") as f:
886890
for line in f:
891+
line = cpt.to_text(line)
887892
fields = line.strip("\n").split(self._field_delimiter)
888893
if (not self._only_src and len(fields) == 2) or (
889894
self._only_src and len(fields) == 1):
@@ -892,8 +897,9 @@ def _load_lines(self, fpattern, tar_fname):
892897
@staticmethod
893898
def load_dict(dict_path, reverse=False):
894899
word_dict = {}
895-
with open(dict_path, "r") as fdict:
900+
with open(dict_path, "rb") as fdict:
896901
for idx, line in enumerate(fdict):
902+
line = cpt.to_text(line)
897903
if reverse:
898904
word_dict[idx] = line.strip("\n")
899905
else:
@@ -1034,7 +1040,7 @@ def __combine_heads(x):
10341040
# size of the input as the output dimension size.
10351041
return layers.reshape(
10361042
x=trans_x,
1037-
shape=map(int, [0, 0, trans_x.shape[2] * trans_x.shape[3]]))
1043+
shape=list(map(int, [0, 0, trans_x.shape[2] * trans_x.shape[3]])))
10381044

10391045
def scaled_dot_product_attention(q, k, v, attn_bias, d_model, dropout_rate):
10401046
"""

python/paddle/fluid/transpiler/distribute_transpiler.py

Lines changed: 5 additions & 5 deletions
Original file line numberDiff line numberDiff line change
@@ -293,7 +293,7 @@ def transpile(self,
293293
input_deps = grad_name_to_send_dummy_out.values()
294294
program.global_block().append_op(
295295
type="send_barrier",
296-
inputs={"X": input_deps},
296+
inputs={"X": list(input_deps)},
297297
outputs={"Out": send_barrier_out},
298298
attrs={
299299
"endpoints": pserver_endpoints,
@@ -394,7 +394,7 @@ def _get_trainer_startup_program(self, recv_vars, eplist):
394394
395395
Args:
396396
recv_vars (list): Variable list to recv for current trainer_id
397-
eplist (list): A list of strings indicating
397+
eplist (list): A list of strings indicating
398398
399399
Returns:
400400
Program: trainer side startup program.
@@ -448,7 +448,7 @@ def _get_trainer_startup_program(self, recv_vars, eplist):
448448
if len(splited_var) <= 1:
449449
continue
450450
# NOTE: if enable memory optimization, origin vars maybe removed.
451-
if startup_program.global_block().vars.has_key(varname):
451+
if varname in startup_program.global_block().vars:
452452
orig_param = startup_program.global_block().vars[varname]
453453
else:
454454
origin_param_var = self.origin_program.global_block().vars[
@@ -677,7 +677,7 @@ def get_pserver_programs(self, endpoint):
677677
678678
Args:
679679
endpoint (str): current pserver endpoint.
680-
680+
681681
Returns:
682682
tuple: (main_program, startup_program), of type "Program"
683683
"""
@@ -700,7 +700,7 @@ def get_startup_program(self,
700700
endpoint (str): current pserver endpoint.
701701
pserver_program (Program): deprecated, call get_pserver_program first.
702702
startup_program (Program): deprecated, should pass startup_program
703-
when initalizing
703+
when initalizing
704704
705705
Returns:
706706
Program: parameter server side startup program.

0 commit comments

Comments
 (0)