Skip to content

Commit 88a0ced

Browse files
committed
revise the introduction of OCR
1 parent b1363e3 commit 88a0ced

File tree

2 files changed

+25
-10
lines changed

2 files changed

+25
-10
lines changed

README.md

Lines changed: 25 additions & 10 deletions
Original file line numberDiff line numberDiff line change
@@ -5,7 +5,7 @@
55
- The PyTroch 0.4.1 version is available [here](https://github.com/HRNet/HRNet-Semantic-Segmentation/tree/master).
66

77
## News
8-
- [2020/07] Our paper is accepted by ECCV 2020: [Object-Contextual Representations for Semantic Segmentation](https://arxiv.org/pdf/1909.11065.pdf)
8+
- [2020/07] Our paper is accepted by ECCV 2020: [Object-Contextual Representations for Semantic Segmentation](https://arxiv.org/pdf/1909.11065.pdf). Notably, the reseachers from Nvidia set a new state-of-the-art performance on Cityscapes leaderboard: [85.4%](https://www.cityscapes-dataset.com/method-details/?submissionID=7836) via combining our HRNet + OCR with a new [hierarchical mult-scale attention scheme](https://arxiv.org/abs/2005.10821).
99
- [2020/03/13] Our paper is accepted by TPAMI: [Deep High-Resolution Representation Learning for Visual Recognition](https://arxiv.org/pdf/1908.07919.pdf).
1010
- HRNet + OCR + SegFix: Rank \#1 (84.5) in [Cityscapes leaderboard](https://www.cityscapes-dataset.com/benchmarks/). OCR: object contextual represenations [pdf](https://arxiv.org/pdf/1909.11065.pdf). ***HRNet + OCR is reproduced [here](https://github.com/HRNet/HRNet-Semantic-Segmentation/tree/HRNet-OCR)***.
1111
- Thanks Google and UIUC researchers. A modified HRNet combined with semantic and instance multi-scale context achieves SOTA panoptic segmentation result on the Mapillary Vista challenge. See [the paper](https://arxiv.org/pdf/1910.04751.pdf).
@@ -18,9 +18,27 @@
1818
This is the official code of [high-resolution representations for Semantic Segmentation](https://arxiv.org/abs/1904.04514).
1919
We augment the HRNet with a very simple segmentation head shown in the figure below. We aggregate the output representations at four different resolutions, and then use a 1x1 convolutions to fuse these representations. The output representations is fed into the classifier. We evaluate our methods on three datasets, Cityscapes, PASCAL-Context and LIP.
2020

21-
![](figures/seg-hrnet.png)
22-
23-
Besides, we further combine HRNet with [Object Contextual Representation](https://arxiv.org/pdf/1909.11065.pdf) and achieve higher performance on the three datasets. The code of HRNet+OCR is contained in this branch.
21+
<!-- ![](figures/seg-hrnet.png) -->
22+
<figure>
23+
<text-align: center;>
24+
<img src="./figures/seg-hrnet.png" alt="hrnet" title="Framework of Object Contextual Representation" width="900" height="200" />
25+
<figcaption>Fig.1 - An example of a high-resolution network. Only the main body is illustrated, and the stem (two stride-2 3 × 3 convolutions) is not included.
26+
There are four stages. The 1st stage consists of high-resolution convolutions. The 2nd (3rd, 4th) stage repeats two-resolution (three-resolution,
27+
four-resolution) blocks.
28+
</figcaption>
29+
</figure>
30+
31+
Besides, we further combine HRNet with [Object Contextual Representation](https://arxiv.org/pdf/1909.11065.pdf) and achieve higher performance on the three datasets. The code of HRNet+OCR is contained in this branch. We illustrate the overall framework of OCR in the Figure as shown below:
32+
33+
<figure>
34+
<text-align: center;>
35+
<img src="./figures/OCR.PNG" alt="OCR" title="Framework of Object Contextual Representation" width="900" height="200" />
36+
<figcaption>Fig.2 - Illustrating the pipeline of OCR. (i) form the soft object regions in the
37+
pink dashed box. (ii) estimate the object region representations in the purple dashed box.
38+
(iii) compute the object contextual representations and the augmented representations
39+
in the orange dashed box.
40+
</figcaption>
41+
</figure>
2442

2543
## Segmentation models
2644
The models are initialized by the weights pretrained on the ImageNet. You can download the pretrained models from https://github.com/HRNet/HRNet-Image-Classification. *Slightly different, we use align_corners = True for upsampling in HRNet*.
@@ -255,13 +273,10 @@ If you find this work or code is helpful in your research, please cite:
255273
256274
@article{YuanCW19,
257275
title={Object-Contextual Representations for Semantic Segmentation},
258-
author={Yuhui Yuan and Xilin Chen and Jingdong Wang},
259-
journal = {CoRR},
260-
volume = {abs/1909.11065},
261-
year={2019}
276+
author={Yuan, Yuhui and Chen, Xilin and Wang, Jingdong},
277+
booktitle = {ECCV},
278+
year={2020}
262279
}
263-
264-
265280
````
266281

267282
## Reference

figures/OCR.PNG

64.1 KB
Loading

0 commit comments

Comments
 (0)