Skip to content

Commit 1d6cfd9

Browse files
committed
Merge branch 'fix_collective_files_partition' of https://github.com/vslyu/PaddleRec into fix_collective_files_partition
2 parents 7643789 + 641e55e commit 1d6cfd9

File tree

4 files changed

+34
-16
lines changed

4 files changed

+34
-16
lines changed

core/trainers/framework/runner.py

Lines changed: 22 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -18,11 +18,15 @@
1818
import time
1919
import warnings
2020
import numpy as np
21+
import logging
2122
import paddle.fluid as fluid
2223

2324
from paddlerec.core.utils import envs
2425
from paddlerec.core.metric import Metric
2526

27+
logging.basicConfig(
28+
format='%(asctime)s - %(levelname)s: %(message)s', level=logging.INFO)
29+
2630
__all__ = [
2731
"RunnerBase", "SingleRunner", "PSRunner", "CollectiveRunner", "PslibRunner"
2832
]
@@ -140,8 +144,16 @@ def _executor_dataloader_train(self, model_dict, context):
140144

141145
metrics_varnames = []
142146
metrics_format = []
147+
148+
if context["is_infer"]:
149+
metrics_format.append("\t[Infer]\t{}: {{}}".format("batch"))
150+
else:
151+
metrics_format.append("\t[Train]\t{}: {{}}".format("batch"))
152+
153+
metrics_format.append("{}: {{:.2f}}s".format("time_each_interval"))
154+
143155
metrics_names = ["total_batch"]
144-
metrics_format.append("{}: {{}}".format("batch"))
156+
145157
for name, var in metrics.items():
146158
metrics_names.append(name)
147159
metrics_varnames.append(var.name)
@@ -151,6 +163,7 @@ def _executor_dataloader_train(self, model_dict, context):
151163
reader = context["model"][model_dict["name"]]["model"]._data_loader
152164
reader.start()
153165
batch_id = 0
166+
begin_time = time.time()
154167
scope = context["model"][model_name]["scope"]
155168
result = None
156169
with fluid.scope_guard(scope):
@@ -160,16 +173,22 @@ def _executor_dataloader_train(self, model_dict, context):
160173
program=program,
161174
fetch_list=metrics_varnames,
162175
return_numpy=False)
163-
metrics = [batch_id]
164176

177+
metrics = [batch_id]
165178
metrics_rets = [
166179
as_numpy(metrics_tensor)
167180
for metrics_tensor in metrics_tensors
168181
]
169182
metrics.extend(metrics_rets)
170183

171184
if batch_id % fetch_period == 0 and batch_id != 0:
172-
print(metrics_format.format(*metrics))
185+
end_time = time.time()
186+
seconds = end_time - begin_time
187+
metrics_logging = metrics[:]
188+
metrics_logging = metrics.insert(1, seconds)
189+
begin_time = end_time
190+
191+
logging.info(metrics_format.format(*metrics))
173192
batch_id += 1
174193
except fluid.core.EOFException:
175194
reader.reset()

models/contentunderstanding/classification/readme.md

Lines changed: 6 additions & 6 deletions
Original file line numberDiff line numberDiff line change
@@ -4,11 +4,11 @@
44

55
```
66
├── data #样例数据
7-
├── train
8-
├── train.txt #训练数据样例
9-
├── test
10-
├── test.txt #测试数据样例
11-
├── preprocess.py #数据处理程序
7+
├── train
8+
├── train.txt #训练数据样例
9+
├── test
10+
├── test.txt #测试数据样例
11+
├── preprocess.py #数据处理程序
1212
├── __init__.py
1313
├── README.md #文档
1414
├── model.py #模型文件
@@ -44,7 +44,7 @@ Yoon Kim在论文[EMNLP 2014][Convolutional neural networks for sentence classic
4444

4545
| 模型 | dev | test |
4646
| :------| :------ | :------
47-
| TextCNN | 90.75% | 92.19% |
47+
| TextCNN | 90.75% | 91.27% |
4848

4949

5050
您可以直接执行以下命令下载我们分词完毕后的数据集,文件解压之后,senta_data目录下会存在训练数据(train.tsv)、开发集数据(dev.tsv)、测试集数据(test.tsv)以及对应的词典(word_dict.txt):

models/multitask/mmoe/config.yaml

Lines changed: 4 additions & 5 deletions
Original file line numberDiff line numberDiff line change
@@ -17,12 +17,12 @@ workspace: "models/multitask/mmoe"
1717
dataset:
1818
- name: dataset_train
1919
batch_size: 5
20-
type: QueueDataset
20+
type: DataLoader # or QueueDataset
2121
data_path: "{workspace}/data/train"
2222
data_converter: "{workspace}/census_reader.py"
2323
- name: dataset_infer
2424
batch_size: 5
25-
type: QueueDataset
25+
type: DataLoader # or QueueDataset
2626
data_path: "{workspace}/data/train"
2727
data_converter: "{workspace}/census_reader.py"
2828

@@ -37,7 +37,6 @@ hyper_parameters:
3737
learning_rate: 0.001
3838
strategy: async
3939

40-
#use infer_runner mode and modify 'phase' below if infer
4140
mode: [train_runner, infer_runner]
4241

4342
runner:
@@ -49,10 +48,10 @@ runner:
4948
save_inference_interval: 4
5049
save_checkpoint_path: "increment"
5150
save_inference_path: "inference"
52-
print_interval: 10
51+
print_interval: 1
5352
- name: infer_runner
5453
class: infer
55-
init_model_path: "increment/0"
54+
init_model_path: "increment/1"
5655
device: cpu
5756

5857
phase:

models/rank/fibinet/config.yaml

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -102,9 +102,9 @@ phase:
102102
- name: phase1
103103
model: "{workspace}/model.py" # user-defined model
104104
dataset_name: dataloader_train # select dataset by name
105-
thread_num: 8
105+
thread_num: 1
106106

107107
- name: phase2
108108
model: "{workspace}/model.py" # user-defined model
109109
dataset_name: dataset_infer # select dataset by name
110-
thread_num: 8
110+
thread_num: 1

0 commit comments

Comments
 (0)