Movatterモバイル変換


[0]ホーム

URL:


Skip to content

Navigation Menu

Sign in
Appearance settings

Search code, repositories, users, issues, pull requests...

Provide feedback

We read every piece of feedback, and take your input very seriously.

Saved searches

Use saved searches to filter your results more quickly

Sign up
Appearance settings

This repository contains the architectures, Models, logs, etc pertaining to the SimpleNet Paper (Lets keep it simple: Using simple architectures to outperform deeper architectures )

License

NotificationsYou must be signed in to change notification settings

Coderx7/SimpleNet

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

81 Commits
 
 
 
 
 
 

Repository files navigation

بسم الله الرحمن الرحیم
پیاده سازی رسمی سیمپل نت در کفی 2016

GitHub Logo

This repository contains the architectures, Models, logs, etc pertaining to the SimpleNet Paper(Lets keep it simple: Using simple architectures to outperform deeper architectures ) :https://arxiv.org/abs/1608.06037

SimpleNet-V1 outperforms deeper and heavier architectures such as AlexNet, VGGNet,ResNet,GoogleNet,etc in a series of benchmark datasets, such as CIFAR10/100, MNIST, SVHN.It also achievs a higher accuracy (currently72.03/90.32) in imagenet, more than VGGNet, ResNet, MobileNet, AlexNet, NIN, Squeezenet, etc with only 5.7M parameters. It also achieves74.23/91.748) with 9m version.
Slimer versions of the architecture work very decently against more complex architectures such as ResNet, WRN and MobileNet as well.

Citation

If you find SimpleNet useful in your research, please consider citing:

@article{hasanpour2016lets,  title={Lets keep it simple, Using simple architectures to outperform deeper and more complex architectures},  author={Hasanpour, Seyyed Hossein and Rouhani, Mohammad and Fayyaz, Mohsen and Sabokrou, Mohammad},  journal={arXiv preprint arXiv:1608.06037},  year={2016}}

(Check the successor of this architecture atTowards Principled Design of Deep Convolutional Networks: Introducing SimpNet)
 

Other Implementations :

OfficialPytorch implementation
 
 

Results Overview :

ImageNet result below was achieved using thePytorch implementation

DatasetAccuracy
ImageNet-top1 (9m)74.23
ImageNet-top1 (5m)72.03
Cifar1095.51
CIFAR100*78.37
MNIST99.75
SVHN98.21
  • Achieved using Pytorch implementation

ImageNet Result:

SimpleNet outperforms much deeper and larger architectures on the ImageNet dataset:

ModelParamsTop1Top5
AlexNet60M57.280.3
SqeezeNet1.2M58.1880.62
VGGNet16138M71.5990.38
VGGNet16_BN138M73.3691.52
VGGNet19143M72.3890.88
VGGNet19_BN143M74.2291.84
GoogleNet6.6M69.7889.53
WResNet1811.7M69.6089.07
ResNet1811.7M69.7689.08
ResNet3421.8M73.3191.42
SimpleNet_small_0501.5M61.6783.49
SimpleNet_small_0753.2M68.5188.15
SimpleNet_5m5.7M72.0390.32
SimpleNet_9m9.5M74.2391.75

Extended ImageNet Result:

Model#ParamsImageNetImageNet-Real-Labels
simplenetv1_9m_m2(36.3 MB)9.5m74.23 / 91.74881.22 / 94.756
simplenetv1_5m_m2(22 MB)5.7m72.03 / 90.32479.328/ 93.714
simplenetv1_small_m2_075(12.6 MB)3m68.506/ 88.1576.283/ 92.02
simplenetv1_small_m2_05(5.78 MB)1.5m61.67 / 83.48869.31 / 88.195

SimpleNet performs very decently, it outperforms VGGNet, variants of ResNet and MobileNets(1-3)
and is pretty fast as well! and its all using plain old CNN!.
To view the full benchmark results visit thebenchmark page.
To view more results checkout thethe Pytorch implementation page

Top CIFAR10/100 results:

Method#ParamsCIFAR10CIFAR100
VGGNet(16L) /Enhanced138m91.4 / 92.45-
ResNet-110L / 1202L *1.7/10.2m93.57 / 92.0774.84/72.18
SD-110L / 1202L1.7/10.2m94.77 / 95.0975.42 / -
WRN-(16/8)/(28/10)11/36m95.19 / 95.8377.11/79.5
Highway NetworkN/A92.4067.76
FitNet1M91.6164.96
FMP* (1 tests)12M95.5073.61
Max-out(k=2)6M90.6265.46
Network in Network1M91.1964.32
DSN1M92.0365.43
Max-out NIN-93.2571.14
LSUVN/A94.16N/A
SimpleNet-Arch 1(۞)5.48M94.75-
SimpleNet-Arch 2 (۩)5.48M95.5178.37

*Note that the Fractional max pooling[13] uses deeper architectures and also uses extreme data augmentation.۞ means No zero-padding or normalization with dropout and ۩ means Standard data-augmentation- with dropout. To our knowledge, our architecture has the state of the art result, without aforementioned data-augmentations.

MNIST results:

MethodError rate
DropConnect**0.21%
Multi-column DNN for Image Classification**0.23%
APAC**0.23%
Generalizing Pooling Functions in CNN**0.29%
Fractional Max-Pooling**0.32%
Batch-normalized Max-out NIN0.24%
Max-out network (k=2)0.45%
Network In Network0.45%
Deeply Supervised Network0.39%
RCNN-960.31%
SimpleNet *0.25%

*Note that we didn’t intend on achieving the state of the artperformance here as we are using a single optimization policy withoutfine-tuning hyper parameters or data-augmentation for a specific task,and still we nearly achieved state-of-the-art on MNIST. **Resultsachieved using an ensemble or extreme data-augmentation

Top SVHN results:

MethodError rate
Network in Network2.35
Deeply Supervised Net1.92
ResNet (reported by (2016))2.01
ResNet with Stochastic Depth1.75
Wide ResNet1.64
SimpleNet1.79

Table 6-Slimmed version Results on Different Datasets

ModelOursMaxoutDSNALLCNNdasNetResNet(32)WRNNIN
#Param310K460K6M1M1.3M6M475K600K
CIFAR1091.9892.3390.6292.0392.7590.7891.693.15
CIFAR10064.6866.8265.4665.4366.2966.2267.3769.11
Other datasetsOur result
MNIST(310K)*99.72
SVHN(310K)*97.63

*Since we presented their results in their respective sections, we avoided mentioning the results here again.

Cifar10 extended results:

MethodAccuracy#Params
VGGNet(16L)91.4138m
VGGNET(Enhanced-16L)*92.45138m
ResNet-110*93.571.7m
ResNet-120292.0710.2m
Stochastic depth-110L94.771.7m
Stochastic depth-1202L95.0910.2m
Wide Residual Net95.1911m
Wide Residual Net95.8336m
Highway Network92.40-
FitNet91.611M
SqueezNet-(tested by us)79.581.3M
ALLCNN92.75-
Fractional Max-pooling* (1 tests)95.5012M
Max-out(k=2)90.626M
Network in Network91.191M
Deeply Supervised Network92.031M
Batch normalized Max-out NIN93.25-
All you need is a good init (LSUV)94.16-
Generalizing Pooling Functions in CNN93.95-
Spatially-Sparse CNNs93.72-
93.63-
Recurrent CNN for Object Recognition92.91-
RCNN-16092.91-
SimpleNet-Arch194.755.4m
SimpleNet-Arch1 using data augmentation95.515.4m

CIFAR100 Extended results:

MethodAccuracy
GoogleNet with ELU*75.72
Spatially-sparse CNNs75.7
Fractional Max-Pooling(12M)73.61
Scalable Bayesian Optimization Using DNNs72.60
All you need is a good init72.34
Batch-normalized Max-out NIN(k=5)71.14
Network in Network64.32
Deeply Supervised Network65.43
ResNet-110L74.84
ResNet-1202L72.18
WRN77.11/79.5
Highway67.76
FitNet64.96
SimpleNet78.37

** Achieved using several data-augmentation tricks

Flops and Parameter Comparison:[tab:Flops_appndx]

ModelMACCCOMPADDDIVActivationsParamsSIZE(MB)
SimpleNet1.9G1.82M1.5M1.5M6.38M6.4M24.4
SqueezeNet861.34M9.67M226K1.51M12.58M1.25M4.7
Inception v4*12.27G21.87M53.42M15.09M72.56M42.71M163
Inception v3*5.72G16.53M25.94M8.97M41.33M23.83M91
Incep-Resv2*13.18G31.57M38.81M25.06M117.8M55.97M214
ResNet-15211.3G22.33M35.27M22.03M100.11M60.19M230
ResNet-503.87G10.89M16.21M10.59M46.72M25.56M97.70
AlexNet7.27G17.69M4.78M9.55M20.81M60.97M217.00
GoogleNet16.04G161.07M8.83M16.64M102.19M7M40
NIN11.06G28.93M380K20K38.79M7.6M29
VGG16154.7G196.85M10K10K288.03M138.36M512.2

Flops and Parameter Comparison of Models trained on ImageNet

*Inception v3, v4 did not have any Caffe model, so we reported theirsize related information from MXNet and Tensorflow respectively.Inception-ResNet-V2 would take 60 days of training with 2 Titan X toachieve the reported accuracy. Statistics are obtained usinghttp://dgschwend.github.io/netscope

1# Data-augmentation method used by stochastic depth paper:https://github.com/Pastromhaug/caffe-stochastic-depth.

2#https://github.com/dmlc/mxnet-model-gallery/blob/master/imagenet-1k-inception-v3.md

3#https://github.com/tensorflow/models/tree/master/slim

4#https://github.com/revilokeb/inception\_resnetv2\_caffe

Side Note:

This was based on my Master's thesis titled "Object classification using Deep Convolutional neural networks" back in 1394/2015.

Citation

If you find SimpleNet useful in your research, please consider citing:

@article{hasanpour2016lets,  title={Lets keep it simple, Using simple architectures to outperform deeper and more complex architectures},  author={Hasanpour, Seyyed Hossein and Rouhani, Mohammad and Fayyaz, Mohsen and Sabokrou, Mohammad},  journal={arXiv preprint arXiv:1608.06037},  year={2016}}

About

This repository contains the architectures, Models, logs, etc pertaining to the SimpleNet Paper (Lets keep it simple: Using simple architectures to outperform deeper architectures )

Topics

Resources

License

Stars

Watchers

Forks

Packages

No packages published

Languages


[8]ページ先頭

©2009-2026 Movatter.jp