You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Copy file name to clipboardExpand all lines: README.md
+21-5Lines changed: 21 additions & 5 deletions
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -6,10 +6,11 @@ This repository contains the architectures, Models, logs, etc pertaining to the
6
6
(Lets keep it simple: Using simple architectures to outperform deeper architectures ) : https://arxiv.org/abs/1608.06037
7
7
8
8
SimpleNet-V1 outperforms deeper and heavier architectures such as AlexNet, VGGNet,ResNet,GoogleNet,etc in a series of benchmark datasets, such as CIFAR10/100, MNIST, SVHN.
9
-
It also achievs a higher accuracy (currently [71.50/90.05 and 78.88/93.43*](https://github.com/Coderx7/SimpleNet_Pytorch#imagenet-result)) in imagenet, more than VGGNet, ResNet, MobileNet, AlexNet, NIN, Squeezenet, etc with only 5.7M parameters.
9
+
It also achievs a higher accuracy (currently [71.94/90.30 and 79.12/93.68*](https://github.com/Coderx7/SimpleNet_Pytorch#imagenet-result)) in imagenet, more than VGGNet, ResNet, MobileNet, AlexNet, NIN, Squeezenet, etc with only 5.7M parameters. It also achieves **74.17/91.614** and **81.24/94.63*** with 9m version.
10
10
Slimer versions of the architecture work very decently against more complex architectures such as ResNet, WRN and MobileNet as well.
11
11
12
-
*78.88/93.43 was achieved using real-imagenet-labels
12
+
*79.12/93.68 was achieved using real-imagenet-labels
13
+
*81.24/94.63 was achieved using real-imagenet-labels
13
14
14
15
## Citation
15
16
If you find SimpleNet useful in your research, please consider citing:
@@ -32,15 +33,29 @@ For using Pytorch implemnetation click [Pytorch implementation](https://github.c
32
33
33
34
34
35
## Results Overview :
35
-
ImageNet result was achieved using simple SGD without hyper parameter tuning for 100 epochs(single crop). no multicrop techniques were used. no dense evaluation or combinations of such techniques were used unlike all other architectures. the models will be uploaded when the training is finished.
36
+
ImageNet result below was achieved using the [Pytorch implementation](https://github.com/Coderx7/SimpleNet_Pytorch)
0 commit comments