|
5 | 5 | - The PyTroch 0.4.1 version is available [here](https://github.com/HRNet/HRNet-Semantic-Segmentation/tree/master). |
6 | 6 |
|
7 | 7 | ## News |
8 | | -- [2020/07] Our paper is accepted by ECCV 2020: [Object-Contextual Representations for Semantic Segmentation](https://arxiv.org/pdf/1909.11065.pdf) |
| 8 | +- [2020/07] Our paper is accepted by ECCV 2020: [Object-Contextual Representations for Semantic Segmentation](https://arxiv.org/pdf/1909.11065.pdf). Notably, the reseachers from Nvidia set a new state-of-the-art performance on Cityscapes leaderboard: [85.4%](https://www.cityscapes-dataset.com/method-details/?submissionID=7836) via combining our HRNet + OCR with a new [hierarchical mult-scale attention scheme](https://arxiv.org/abs/2005.10821). |
9 | 9 | - [2020/03/13] Our paper is accepted by TPAMI: [Deep High-Resolution Representation Learning for Visual Recognition](https://arxiv.org/pdf/1908.07919.pdf). |
10 | 10 | - HRNet + OCR + SegFix: Rank \#1 (84.5) in [Cityscapes leaderboard](https://www.cityscapes-dataset.com/benchmarks/). OCR: object contextual represenations [pdf](https://arxiv.org/pdf/1909.11065.pdf). ***HRNet + OCR is reproduced [here](https://github.com/HRNet/HRNet-Semantic-Segmentation/tree/HRNet-OCR)***. |
11 | 11 | - Thanks Google and UIUC researchers. A modified HRNet combined with semantic and instance multi-scale context achieves SOTA panoptic segmentation result on the Mapillary Vista challenge. See [the paper](https://arxiv.org/pdf/1910.04751.pdf). |
|
18 | 18 | This is the official code of [high-resolution representations for Semantic Segmentation](https://arxiv.org/abs/1904.04514). |
19 | 19 | We augment the HRNet with a very simple segmentation head shown in the figure below. We aggregate the output representations at four different resolutions, and then use a 1x1 convolutions to fuse these representations. The output representations is fed into the classifier. We evaluate our methods on three datasets, Cityscapes, PASCAL-Context and LIP. |
20 | 20 |
|
21 | | - |
22 | | - |
23 | | -Besides, we further combine HRNet with [Object Contextual Representation](https://arxiv.org/pdf/1909.11065.pdf) and achieve higher performance on the three datasets. The code of HRNet+OCR is contained in this branch. |
| 21 | +<!--  --> |
| 22 | +<figure> |
| 23 | + <text-align: center;> |
| 24 | + <img src="./figures/seg-hrnet.png" alt="hrnet" title="Framework of Object Contextual Representation" width="900" height="200" /> |
| 25 | + <figcaption>Fig.1 - An example of a high-resolution network. Only the main body is illustrated, and the stem (two stride-2 3 × 3 convolutions) is not included. |
| 26 | + There are four stages. The 1st stage consists of high-resolution convolutions. The 2nd (3rd, 4th) stage repeats two-resolution (three-resolution, |
| 27 | + four-resolution) blocks. |
| 28 | +</figcaption> |
| 29 | +</figure> |
| 30 | + |
| 31 | +Besides, we further combine HRNet with [Object Contextual Representation](https://arxiv.org/pdf/1909.11065.pdf) and achieve higher performance on the three datasets. The code of HRNet+OCR is contained in this branch. We illustrate the overall framework of OCR in the Figure as shown below: |
| 32 | + |
| 33 | +<figure> |
| 34 | + <text-align: center;> |
| 35 | + <img src="./figures/OCR.PNG" alt="OCR" title="Framework of Object Contextual Representation" width="900" height="200" /> |
| 36 | + <figcaption>Fig.2 - Illustrating the pipeline of OCR. (i) form the soft object regions in the |
| 37 | + pink dashed box. (ii) estimate the object region representations in the purple dashed box. |
| 38 | + (iii) compute the object contextual representations and the augmented representations |
| 39 | + in the orange dashed box. |
| 40 | +</figcaption> |
| 41 | +</figure> |
24 | 42 |
|
25 | 43 | ## Segmentation models |
26 | 44 | The models are initialized by the weights pretrained on the ImageNet. You can download the pretrained models from https://github.com/HRNet/HRNet-Image-Classification. *Slightly different, we use align_corners = True for upsampling in HRNet*. |
@@ -255,13 +273,10 @@ If you find this work or code is helpful in your research, please cite: |
255 | 273 |
|
256 | 274 | @article{YuanCW19, |
257 | 275 | title={Object-Contextual Representations for Semantic Segmentation}, |
258 | | - author={Yuhui Yuan and Xilin Chen and Jingdong Wang}, |
259 | | - journal = {CoRR}, |
260 | | - volume = {abs/1909.11065}, |
261 | | - year={2019} |
| 276 | + author={Yuan, Yuhui and Chen, Xilin and Wang, Jingdong}, |
| 277 | + booktitle = {ECCV}, |
| 278 | + year={2020} |
262 | 279 | } |
263 | | -
|
264 | | -
|
265 | 280 | ```` |
266 | 281 |
|
267 | 282 | ## Reference |
|
0 commit comments