You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
@@ -259,37 +124,78 @@ A PaddlePaddle version image model zoo.
259
124
journal = {arXiv preprint arXiv:2007.00992},
260
125
year = {2020},
261
126
}
262
-
```
263
-
```
127
+
264
128
@InProceedings{Li_2021_CVPR,
265
129
title = {Involution: Inverting the Inherence of Convolution for Visual Recognition},
266
130
author = {Li, Duo and Hu, Jie and Wang, Changhu and Li, Xiangtai and She, Qi and Zhu, Lei and Zhang, Tong and Chen, Qifeng},
267
131
booktitle = {IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR)},
268
132
month = {June},
269
133
year = {2021}
270
134
}
271
-
```
272
-
```
135
+
273
136
@article{ding2021repvgg,
274
137
title={RepVGG: Making VGG-style ConvNets Great Again},
275
138
author={Ding, Xiaohan and Zhang, Xiangyu and Ma, Ningning and Han, Jungong and Ding, Guiguang and Sun, Jian},
276
139
journal={arXiv preprint arXiv:2101.03697},
277
140
year={2021}
278
141
}
279
-
```
280
-
```
142
+
281
143
@article{heo2021pit,
282
144
title={Rethinking Spatial Dimensions of Vision Transformers},
283
145
author={Byeongho Heo and Sangdoo Yun and Dongyoon Han and Sanghyuk Chun and Junsuk Choe and Seong Joon Oh},
284
146
journal={arXiv: 2103.16302},
285
147
year={2021},
286
148
}
287
-
```
288
-
```
149
+
289
150
@article{touvron2020deit,
290
151
title = {Training data-efficient image transformers & distillation through attention},
291
152
author = {Hugo Touvron and Matthieu Cord and Matthijs Douze and Francisco Massa and Alexandre Sablayrolles and Herv'e J'egou},
292
153
journal = {arXiv preprint arXiv:2012.12877},
293
154
year = {2020}
294
155
}
295
-
```
156
+
157
+
@misc{han2021transformer,
158
+
title={Transformer in Transformer},
159
+
author={Kai Han and An Xiao and Enhua Wu and Jianyuan Guo and Chunjing Xu and Yunhe Wang},
160
+
year={2021},
161
+
eprint={2103.00112},
162
+
archivePrefix={arXiv},
163
+
primaryClass={cs.CV}
164
+
}
165
+
166
+
@misc{chao2019hardnet,
167
+
title={HarDNet: A Low Memory Traffic Network},
168
+
author={Ping Chao and Chao-Yang Kao and Yu-Shan Ruan and Chien-Hsiang Huang and Youn-Long Lin},
169
+
year={2019},
170
+
eprint={1909.00948},
171
+
archivePrefix={arXiv},
172
+
primaryClass={cs.CV}
173
+
}
174
+
175
+
@misc{yu2019deep,
176
+
title={Deep Layer Aggregation},
177
+
author={Fisher Yu and Dequan Wang and Evan Shelhamer and Trevor Darrell},
178
+
year={2019},
179
+
eprint={1707.06484},
180
+
archivePrefix={arXiv},
181
+
primaryClass={cs.CV}
182
+
}
183
+
184
+
@misc{dosovitskiy2020image,
185
+
title={An Image is Worth 16x16 Words: Transformers for Image Recognition at Scale},
186
+
author={Alexey Dosovitskiy and Lucas Beyer and Alexander Kolesnikov and Dirk Weissenborn and Xiaohua Zhai and Thomas Unterthiner and Mostafa Dehghani and Matthias Minderer and Georg Heigold and Sylvain Gelly and Jakob Uszkoreit and Neil Houlsby},
187
+
year={2020},
188
+
eprint={2010.11929},
189
+
archivePrefix={arXiv},
190
+
primaryClass={cs.CV}
191
+
}
192
+
193
+
@misc{wang2021pyramid,
194
+
title={Pyramid Vision Transformer: A Versatile Backbone for Dense Prediction without Convolutions},
195
+
author={Wenhai Wang and Enze Xie and Xiang Li and Deng-Ping Fan and Kaitao Song and Ding Liang and Tong Lu and Ping Luo and Ling Shao},
0 commit comments