Movatterモバイル変換


[0]ホーム

URL:


Jump to content
WikipediaThe Free Encyclopedia
Search

EfficientNet

From Wikipedia, the free encyclopedia
This article mayrequirecleanup to meet Wikipedia'squality standards. The specific problem is:citation formatting errors. Please helpimprove this article if you can.(September 2025) (Learn how and when to remove this message)
Family of computer vision models
EfficientNet
DeveloperGoogle AI
Initial releaseMay 2019
Repositorygithub.com/tensorflow/tpu/tree/master/models/official/efficientnet
Written inPython
LicenseApache License 2.0
WebsiteGoogle AI Blog

EfficientNet is a family ofconvolutional neural networks (CNNs) forcomputer vision published by researchers atGoogle AI in 2019.[1] Its key innovation iscompound scaling, which uniformly scales all dimensions of depth, width, and resolution using a single parameter.

EfficientNet models have been adopted in various computer vision tasks, includingimage classification,object detection, andsegmentation.

Compound scaling

[edit]

EfficientNet introducescompound scaling, which, instead of scaling one dimension of the network at a time, such as depth (number of layers), width (number of channels), or resolution (input image size), uses acompound coefficientϕ{\displaystyle \phi } to scale all three dimensions simultaneously. Specifically, given a baseline network, the depth, width, and resolution are scaled according to the following equations:[1]depth multiplier: d=αϕwidth multiplier: w=βϕresolution multiplier: r=γϕ{\displaystyle {\begin{aligned}{\text{depth multiplier: }}d&=\alpha ^{\phi }\\{\text{width multiplier: }}w&=\beta ^{\phi }\\{\text{resolution multiplier: }}r&=\gamma ^{\phi }\end{aligned}}}subject toαβ2γ22{\displaystyle \alpha \cdot \beta ^{2}\cdot \gamma ^{2}\approx 2} andα1,β1,γ1{\displaystyle \alpha \geq 1,\beta \geq 1,\gamma \geq 1}. Theαβ2γ22{\displaystyle \alpha \cdot \beta ^{2}\cdot \gamma ^{2}\approx 2} condition is such that increasingϕ{\displaystyle \phi } by a factor ofϕ0{\displaystyle \phi _{0}} would increase the total FLOPs of running the network on an image approximately2ϕ0{\displaystyle 2^{\phi _{0}}} times. Thehyperparametersα{\displaystyle \alpha },β{\displaystyle \beta }, andγ{\displaystyle \gamma } are determined by a smallgrid search. The original paper suggested 1.2, 1.1, and 1.15, respectively.

Architecturally, they optimized the choice of modules byneural architecture search (NAS), and found that the inverted bottleneck convolution (which they calledMBConv) used inMobileNet worked well.

The EfficientNet family is a stack of MBConv layers, with shapes determined by the compound scaling. The original publication consisted of 8 models, from EfficientNet-B0 to EfficientNet-B7, with increasing model size and accuracy. EfficientNet-B0 is the baseline network, and subsequent models are obtained by scaling the baseline network by increasingϕ{\displaystyle \phi }.

Variants

[edit]

EfficientNet has been adapted for fast inference onedgeTPUs[2] and centralized TPU orGPUclusters by NAS.[3]

EfficientNet V2 was published in June 2021. The architecture was improved by further NAS search with more types of convolutional layers.[4] It also introduced a training method, which progressively increases image size during training, and uses regularization techniques likedropout,RandAugment,[5] and Mixup.[6] The authors claim this approach mitigates accuracy drops often associated with progressive resizing.

See also

[edit]

References

[edit]
  1. ^abTan, Mingxing; Le, Quoc V. (2020-09-11),EfficientNet: Rethinking Model Scaling for Convolutional Neural Networks,arXiv:1905.11946
  2. ^"EfficientNet-EdgeTPU: Creating Accelerator-Optimized Neural Networks with AutoML".research.google. August 6, 2019. Retrieved2024-10-18.
  3. ^Li, Sheng; Tan, Mingxing; Pang, Ruoming; Li, Andrew; Cheng, Liqun; Le, Quoc; Jouppi, Norman P. (2021-02-10),Searching for Fast Model Families on Datacenter Accelerators,arXiv:2102.05610
  4. ^Tan, Mingxing; Le, Quoc V. (2021-06-23),EfficientNetV2: Smaller Models and Faster Training,arXiv:2104.00298
  5. ^Cubuk, Ekin D.; Zoph, Barret; Shlens, Jonathon; Le, Quoc V. (2020)."Randaugment: Practical Automated Data Augmentation With a Reduced Search Space":702–703.arXiv:1909.13719.{{cite journal}}:Cite journal requires|journal= (help)
  6. ^Zhang, Hongyi; Cisse, Moustapha; Dauphin, Yann N.; Lopez-Paz, David (2018-04-27),mixup: Beyond Empirical Risk Minimization,arXiv:1710.09412

External links

[edit]
Computer
programs
AlphaGo
Versions
Competitions
In popular culture
Other
Machine
learning
Neural networks
Other
Generative
AI
Chatbots
Models
Other
See also
Differentiable computing
General
Hardware
Software libraries
Retrieved from "https://en.wikipedia.org/w/index.php?title=EfficientNet&oldid=1309177911"
Categories:
Hidden categories:

[8]ページ先頭

©2009-2025 Movatter.jp