Skip to content

Commit 46f1544

Browse files
committed
Update README.md and docs in prep for 0.3.0 PyPi release.
1 parent af3299b commit 46f1544

File tree

6 files changed

+76
-151
lines changed

6 files changed

+76
-151
lines changed

README.md

Lines changed: 8 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -2,6 +2,13 @@
22

33
## What's New
44

5+
### Oct 30, 2020
6+
* Test with PyTorch 1.7 and fix a small top-n metric view vs reshape issue.
7+
* Convert newly added 224x224 Vision Transformer weights from official JAX repo. 81.8 top-1 for B/16, 83.1 L/16.
8+
* Support PyTorch 1.7 optimized, native SiLU (aka Swish) activation. Add mapping to 'silu' name, custom swish will eventually be deprecated.
9+
* Fix regression for loading pretrained classifier via direct model entrypoint functions. Didn't impact create_model() factory usage.
10+
* PyPi release @ 0.3.0 version!
11+
512
### Oct 26, 2020
613
* Update Vision Transformer models to be compatible with official code release at https://github.com/google-research/vision_transformer
714
* Add Vision Transformer weights (ImageNet-21k pretrain) for 384x384 base and large models converted from official jax impl
@@ -141,7 +148,7 @@ A full version of the list below with source links can be found in the [document
141148
* SelecSLS - https://arxiv.org/abs/1907.00837
142149
* Selective Kernel Networks - https://arxiv.org/abs/1903.06586
143150
* TResNet - https://arxiv.org/abs/2003.13630
144-
* Vision Transformer - https://openreview.net/forum?id=YicbFdNTTy
151+
* Vision Transformer - https://arxiv.org/abs/2010.11929
145152
* VovNet V2 and V1 - https://arxiv.org/abs/1911.06667
146153
* Xception - https://arxiv.org/abs/1610.02357
147154
* Xception (Modified Aligned, Gluon) - https://arxiv.org/abs/1802.02611

docs/archived_changes.md

Lines changed: 11 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -1,5 +1,16 @@
11
# Archived Changes
22

3+
### April 5, 2020
4+
* Add some newly trained MobileNet-V2 models trained with latest h-params, rand augment. They compare quite favourably to EfficientNet-Lite
5+
* 3.5M param MobileNet-V2 100 @ 73%
6+
* 4.5M param MobileNet-V2 110d @ 75%
7+
* 6.1M param MobileNet-V2 140 @ 76.5%
8+
* 5.8M param MobileNet-V2 120d @ 77.3%
9+
10+
### March 18, 2020
11+
* Add EfficientNet-Lite models w/ weights ported from [Tensorflow TPU](https://github.com/tensorflow/tpu/tree/master/models/official/efficientnet/lite)
12+
* Add RandAugment trained ResNeXt-50 32x4d weights with 79.8 top-1. Trained by [Andrew Lavin](https://github.com/andravin) (see Training section for hparams)
13+
314
### Feb 29, 2020
415
* New MobileNet-V3 Large weights trained from stratch with this code to 75.77% top-1
516
* IMPORTANT CHANGE - default weight init changed for all MobilenetV3 / EfficientNet / related models

docs/changes.md

Lines changed: 35 additions & 11 deletions
Original file line numberDiff line numberDiff line change
@@ -1,5 +1,40 @@
11
# Recent Changes
22

3+
### Oct 30, 2020
4+
* Test with PyTorch 1.7 and fix a small top-n metric view vs reshape issue.
5+
* Convert newly added 224x224 Vision Transformer weights from official JAX repo. 81.8 top-1 for B/16, 83.1 L/16.
6+
* Support PyTorch 1.7 optimized, native SiLU (aka Swish) activation. Add mapping to 'silu' name, custom swish will eventually be deprecated.
7+
* Fix regression for loading pretrained classifier via direct model entrypoint functions. Didn't impact create_model() factory usage.
8+
* PyPi release @ 0.3.0 version!
9+
10+
### Oct 26, 2020
11+
* Update Vision Transformer models to be compatible with official code release at https://github.com/google-research/vision_transformer
12+
* Add Vision Transformer weights (ImageNet-21k pretrain) for 384x384 base and large models converted from official jax impl
13+
* ViT-B/16 - 84.2
14+
* ViT-B/32 - 81.7
15+
* ViT-L/16 - 85.2
16+
* ViT-L/32 - 81.5
17+
18+
### Oct 21, 2020
19+
* Weights added for Vision Transformer (ViT) models. 77.86 top-1 for 'small' and 79.35 for 'base'. Thanks to [Christof](https://www.kaggle.com/christofhenkel) for training the base model w/ lots of GPUs.
20+
21+
### Oct 13, 2020
22+
* Initial impl of Vision Transformer models. Both patch and hybrid (CNN backbone) variants. Currently trying to train...
23+
* Adafactor and AdaHessian (FP32 only, no AMP) optimizers
24+
* EdgeTPU-M (`efficientnet_em`) model trained in PyTorch, 79.3 top-1
25+
* Pip release, doc updates pending a few more changes...
26+
27+
### Sept 18, 2020
28+
* New ResNet 'D' weights. 72.7 (top-1) ResNet-18-D, 77.1 ResNet-34-D, 80.5 ResNet-50-D
29+
* Added a few untrained defs for other ResNet models (66D, 101D, 152D, 200/200D)
30+
31+
### Sept 3, 2020
32+
* New weights
33+
* Wide-ResNet50 - 81.5 top-1 (vs 78.5 torchvision)
34+
* SEResNeXt50-32x4d - 81.3 top-1 (vs 79.1 cadene)
35+
* Support for native Torch AMP and channels_last memory format added to train/validate scripts (`--channels-last`, `--native-amp` vs `--apex-amp`)
36+
* Models tested with channels_last on latest NGC 20.08 container. AdaptiveAvgPool in attn layers changed to mean((2,3)) to work around bug with NHWC kernel.
37+
338
### Aug 12, 2020
439
* New/updated weights from training experiments
540
* EfficientNet-B3 - 82.1 top-1 (vs 81.6 for official with AA and 81.9 for AdvProp)
@@ -54,14 +89,3 @@ Bunch of changes:
5489
* TResNet models and SpaceToDepth, AntiAliasDownsampleLayer layers by [mrT23](https://github.com/mrT23)
5590
* ecaresnet (50d, 101d, light) models and two pruned variants using pruning as per (https://arxiv.org/abs/2002.08258) by [Yonathan Aflalo](https://github.com/yoniaflalo)
5691
* 200 pretrained models in total now with updated results csv in results folder
57-
58-
### April 5, 2020
59-
* Add some newly trained MobileNet-V2 models trained with latest h-params, rand augment. They compare quite favourably to EfficientNet-Lite
60-
* 3.5M param MobileNet-V2 100 @ 73%
61-
* 4.5M param MobileNet-V2 110d @ 75%
62-
* 6.1M param MobileNet-V2 140 @ 76.5%
63-
* 5.8M param MobileNet-V2 120d @ 77.3%
64-
65-
### March 18, 2020
66-
* Add EfficientNet-Lite models w/ weights ported from [Tensorflow TPU](https://github.com/tensorflow/tpu/tree/master/models/official/efficientnet/lite)
67-
* Add RandAugment trained ResNeXt-50 32x4d weights with 79.8 top-1. Trained by [Andrew Lavin](https://github.com/andravin) (see Training section for hparams)

docs/index.md

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -13,13 +13,13 @@ pip install timm
1313

1414
Little to no care has been taken to be Python 2.x friendly and will not support it. If you run into any challenges running on Windows, or other OS, I'm definitely open to looking into those issues so long as it's in a reproducible (read Conda) environment.
1515

16-
PyTorch versions 1.4, 1.5.x, and 1.6 have been tested with this code.
16+
PyTorch versions 1.4, 1.5.x, 1.6, and 1.7 have been tested with this code.
1717

1818
I've tried to keep the dependencies minimal, the setup is as per the PyTorch default install instructions for Conda:
1919
```
2020
conda create -n torch-env
2121
conda activate torch-env
22-
conda install -c pytorch pytorch torchvision cudatoolkit=10.2
22+
conda install -c pytorch pytorch torchvision cudatoolkit=11
2323
conda install pyyaml
2424
```
2525

docs/models.md

Lines changed: 4 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -132,6 +132,10 @@ NOTE: I am deprecating this version of the networks, the new ones are part of `r
132132
* Paper: `TResNet: High Performance GPU-Dedicated Architecture` - https://arxiv.org/abs/2003.13630
133133
* Code: https://github.com/mrT23/TResNet
134134

135+
## Vision Transformer [[vision_transformer.py](https://github.com/rwightman/pytorch-image-models/blob/master/timm/models/vision_transformer.py)]
136+
* Paper: `An Image is Worth 16x16 Words: Transformers for Image Recognition at Scale` - https://arxiv.org/abs/2010.11929
137+
* Reference code and pretrained weights: https://github.com/google-research/vision_transformer
138+
135139
## VovNet V2 and V1 [[vovnet.py](https://github.com/rwightman/pytorch-image-models/blob/master/timm/models/vovnet.py)]
136140
* Paper: `CenterMask : Real-Time Anchor-Free Instance Segmentation` - https://arxiv.org/abs/1911.06667
137141
* Reference code: https://github.com/youngwanLEE/vovnet-detectron2

0 commit comments

Comments
 (0)