Skip to content

Commit 717ac03

Browse files
committed
pre-commit fix
1 parent 2016921 commit 717ac03

File tree

6 files changed

+34
-27
lines changed

6 files changed

+34
-27
lines changed

mmocr/datasets/preparers/config_generators/__init__.py

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -1,10 +1,10 @@
11
# Copyright (c) OpenMMLab. All rights reserved.
22
from .base import BaseDatasetConfigGenerator
3+
from .re_config_generator import REConfigGenerator
4+
from .ser_config_generator import SERConfigGenerator
35
from .textdet_config_generator import TextDetConfigGenerator
46
from .textrecog_config_generator import TextRecogConfigGenerator
57
from .textspotting_config_generator import TextSpottingConfigGenerator
6-
from .ser_config_generator import SERConfigGenerator
7-
from .re_config_generator import REConfigGenerator
88

99
__all__ = [
1010
'BaseDatasetConfigGenerator', 'TextDetConfigGenerator',

mmocr/datasets/preparers/obtainers/naive_data_obtainer.py

Lines changed: 1 addition & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -186,8 +186,7 @@ def move(self, mapping: List[Tuple[str, str]]) -> None:
186186
if '*' in src:
187187
mkdir_or_exist(dst)
188188
for f in glob.glob(src):
189-
if not osp.exists(
190-
osp.join(dst, osp.basename(f))):
189+
if not osp.exists(osp.join(dst, osp.basename(f))):
191190
shutil.move(f, dst)
192191

193192
elif osp.exists(src) and not osp.exists(dst):

mmocr/datasets/preparers/packers/__init__.py

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -1,11 +1,11 @@
11
# Copyright (c) OpenMMLab. All rights reserved.
22
from .base import BasePacker
3+
from .re_packer import REPacker
4+
from .ser_packer import SERPacker
35
from .textdet_packer import TextDetPacker
46
from .textrecog_packer import TextRecogCropPacker, TextRecogPacker
57
from .textspotting_packer import TextSpottingPacker
68
from .wildreceipt_packer import WildReceiptPacker
7-
from .ser_packer import SERPacker
8-
from .re_packer import REPacker
99

1010
__all__ = [
1111
'BasePacker', 'TextDetPacker', 'TextRecogPacker', 'TextRecogCropPacker',

mmocr/datasets/preparers/packers/re_packer.py

Lines changed: 8 additions & 5 deletions
Original file line numberDiff line numberDiff line change
@@ -10,8 +10,8 @@
1010

1111
@DATA_PACKERS.register_module()
1212
class REPacker(BasePacker):
13-
"""Relation Extraction packer.
14-
It is used to pack the parsed annotation info to.
13+
"""Relation Extraction packer. It is used to pack the parsed annotation
14+
info to.
1515
1616
.. code-block:: python
1717
@@ -43,7 +43,8 @@ class REPacker(BasePacker):
4343
"instances":
4444
{
4545
"texts": ["绩效目标申报表(一级项目)", "项目名称", ...],
46-
"bboxes": [[906,195,1478,259], [357,325,467,357], ...],
46+
"bboxes": [[906,195,1478,259],
47+
[357,325,467,357], ...],
4748
"labels": ["header", "question", ...],
4849
"linkings": [[0, 1], [2, 3], ...],
4950
"ids": [0, 1, ...],
@@ -151,8 +152,10 @@ def add_meta(self, sample: List) -> Dict:
151152
'dataset_type': 'REDataset',
152153
'task_name': 're',
153154
're_labels': label_list,
154-
'id2label': {k: v for k, v in enumerate(label_list)},
155-
'label2id': {v: k for k, v in enumerate(label_list)}
155+
'id2label': {k: v
156+
for k, v in enumerate(label_list)},
157+
'label2id': {v: k
158+
for k, v in enumerate(label_list)}
156159
},
157160
'data_list': sample
158161
}

mmocr/datasets/preparers/packers/ser_packer.py

Lines changed: 8 additions & 5 deletions
Original file line numberDiff line numberDiff line change
@@ -10,8 +10,8 @@
1010

1111
@DATA_PACKERS.register_module()
1212
class SERPacker(BasePacker):
13-
"""Semantic Entity Recognition packer.
14-
It is used to pack the parsed annotation info to.
13+
"""Semantic Entity Recognition packer. It is used to pack the parsed
14+
annotation info to.
1515
1616
.. code-block:: python
1717
@@ -43,7 +43,8 @@ class SERPacker(BasePacker):
4343
"instances":
4444
{
4545
"texts": ["绩效目标申报表(一级项目)", "项目名称", ...],
46-
"bboxes": [[906,195,1478,259], [357,325,467,357], ...],
46+
"bboxes": [[906,195,1478,259],
47+
[357,325,467,357], ...],
4748
"labels": ["header", "question", ...],
4849
"words": [[{
4950
"box": [
@@ -139,8 +140,10 @@ def add_meta(self, sample: List) -> Dict:
139140
'dataset_type': 'SERDataset',
140141
'task_name': 'ser',
141142
'ser_labels': label_list,
142-
'id2label': {k: v for k, v in enumerate(label_list)},
143-
'label2id': {v: k for k, v in enumerate(label_list)}
143+
'id2label': {k: v
144+
for k, v in enumerate(label_list)},
145+
'label2id': {v: k
146+
for k, v in enumerate(label_list)}
144147
},
145148
'data_list': sample
146149
}

mmocr/datasets/preparers/parsers/xfund_parser.py

Lines changed: 13 additions & 11 deletions
Original file line numberDiff line numberDiff line change
@@ -24,7 +24,7 @@ def parse_files(self, img_dir: str, ann_path: str) -> List:
2424
for img_fname, instance in self.loader(ann_path):
2525
samples.append((osp.join(img_dir, img_fname), instance))
2626
return samples
27-
27+
2828
def loader(self, file_path: str):
2929
with open(file_path, 'r', encoding='utf-8') as f:
3030
data = json.load(f)
@@ -33,10 +33,11 @@ def loader(self, file_path: str):
3333
instances = list()
3434
for j in range(len(data['documents'][i]['document'])):
3535
cur_item = data['documents'][i]['document'][j]
36-
instance = dict(text=cur_item['text'],
37-
box=cur_item['box'],
38-
label=cur_item['label'],
39-
words=cur_item['words'])
36+
instance = dict(
37+
text=cur_item['text'],
38+
box=cur_item['box'],
39+
label=cur_item['label'],
40+
words=cur_item['words'])
4041
instances.append(instance)
4142
yield img_fname, instances
4243

@@ -59,11 +60,12 @@ def loader(self, file_path: str):
5960
instances = list()
6061
for j in range(len(data['documents'][i]['document'])):
6162
cur_item = data['documents'][i]['document'][j]
62-
instance = dict(text=cur_item['text'],
63-
box=cur_item['box'],
64-
label=cur_item['label'],
65-
words=cur_item['words'],
66-
linking=cur_item['linking'],
67-
id=cur_item['id'])
63+
instance = dict(
64+
text=cur_item['text'],
65+
box=cur_item['box'],
66+
label=cur_item['label'],
67+
words=cur_item['words'],
68+
linking=cur_item['linking'],
69+
id=cur_item['id'])
6870
instances.append(instance)
6971
yield img_fname, instances

0 commit comments

Comments
 (0)