当前位置:网站首页>【MnasNet】《MnasNet:Platform-Aware Neural Architecture Search for Mobile》
【MnasNet】《MnasNet:Platform-Aware Neural Architecture Search for Mobile》
2022-07-02 07:48:00 【bryant_ meng】


CVPR-2019
List of articles
1 Background and Motivation
The author aims to design a new resource-constrained mobile model Let it be in resource-constrained platforms Run faster
2 Related Work
Compress the existing network : quantitative ,pruning ,NetAdapt etc. ,do not focus on learning novel compositions of CNN operations
hand-crafted Design ,usually take significant human efforts
NAS, Based on a variety of learning algorithms, for example reinforcement learning / evolutionary search / differentiable search
3 Advantages / Contributions
NAS Out MnasNet, Two main innovations
incorporate model latency into the main objective so that the search can identify a model that achieves a good trade-off between accuracy and latency( Not just ACC)
a novel factorized hierarchical search space that encourages layer diversity throughout the network.( Unlike NasNet That is cell Grade , It is block Grade )
achieve new state-of-the-art results on both ImageNet classification and COCO object detection under typical mobile inference latency constraints
4 Method
4.1 Problem Formulation
Previous methods objective function
m m m yes model, A C C ACC ACC yes accuracy, L A T LAT LAT yes inference latency, T T T yes target latency
above objective Only accuracy is considered , Speed is not considered
author more interested in finding multiple Pareto-optimal solutions in a single architecture search( Speed and accuracy trade-off)
Designed the following objective function

according to α \alpha α and β \beta β Different values , There are the following soft and hard edition

Abscissa is latency, Ordinate for objective
soft edition − 0.07 -0.07 −0.07 Its origin is as follows :
we empirically observed doubling the latency usually brings about 5% relative accuracy gain
R e w a r d ( M 2 ) = a ⋅ ( 1 + % 5 ) ⋅ ( 2 l / T ) β ≈ R e w a r d ( M 1 ) = a ⋅ ( l / T ) β Reward(M2) = a \cdot (1 + %5 ) \cdot (2l/T )^{\beta}\approx Reward(M1) = a \cdot (l/T )^{\beta} Reward(M2)=a⋅(1+%5)⋅(2l/T)β≈Reward(M1)=a⋅(l/T)β
Calculate according to the above formula β ≈ − 0.07 \beta \approx -0.07 β≈−0.07
4.2 Factorized Hierarchical Search Space

allowing different layer architectures in different blocks
The same block Medium N individual layer It's the same ,layer The operation inside is as follows

When searching using MobileNetV2 as a reference
Every layers Number {0, +1, -1} based on MobileNetV2
filter size per layer {0.75, 1.0, 1.25} to MobileNetV2
One of the finished structures


The size of the search space is as follows :
hypothesis B B B blocks,and each block has a sub search space of size S S S with average N N N layers per block
The search space is S B S^B SB
Every layer If it's all different , Then for S B ∗ N S^{B*N} SB∗N
4.3 Search Algorithm

sample-eval-update loop,maximize the expected reward:
reward value R(m) It's using objective function
5 Experiments
5.1 Datasets
directly perform our architecture search on the ImageNet training set but with fewer training steps (5 epochs)
The difference in NasNet Of Cifar10
5.2 Results
1)ImageNet Classification Performance
T = 75 ms, One search , Multiple model A1 / A2 / A3
comparison mobileNet v2, Introduced SE modular , Discuss SE The impact of modules 
2)Model Scaling Performance

there depth multiplier refer to channels, It can be seen that it is leading in all directions mobilenet v2
The author can also flexibly change NAS when T To control the size of the model , As can be seen from the table above , It is more powerful than cutting the number of channels on a large model
3)COCO Object Detection Performance

There is nothing to comment on , All vegetables and chickens peck each other , ha-ha , Just kidding , There is a certain improvement
5.3 Ablation Study and Discussion
1)Soft vs. Hard Latency Constraint


hard edition focus more on faster models to avoid the latency penalty(objective function It can also be seen that )
soft edition tries to search for models across a wider latency range
2)Disentangling Search Space and Reward

Decoupling discusses the role of the next two innovations
3)Layer Diversity

6 Conclusion(own)
stay mobilenet v2 Search based on
Pareto-optimal, Pareto is the best ( From Baidu Encyclopedia )
Pareto is the best (Pareto Optimality), Also known as Pareto efficiency (Pareto efficiency), It refers to an ideal state of resource allocation , Suppose there is an inherent group of people and distributable resources , The change from one allocation state to another , Without making anyone worse , Make at least one person better , This is Pareto improvement or Pareto optimization .
Pareto optimal state It is impossible to have more room for Pareto improvement ; let me put it another way , Pareto improvement is the path and method to achieve Pareto optimality . Pareto optimality is fair and efficient “ Ideal kingdom ”. It was proposed by Pareto .
边栏推荐
- How to clean up logs on notebook computers to improve the response speed of web pages
- 传统目标检测笔记1__ Viola Jones
- [CVPR‘22 Oral2] TAN: Temporal Alignment Networks for Long-term Video
- Drawing mechanism of view (3)
- Latex formula normal and italic
- PHP returns the corresponding key value according to the value in the two-dimensional array
- open3d学习笔记五【RGBD融合】
- [Sparse to Dense] Sparse to Dense: Depth Prediction from Sparse Depth samples and a Single Image
- 【Mixed Pooling】《Mixed Pooling for Convolutional Neural Networks》
- 《Handwritten Mathematical Expression Recognition with Bidirectionally Trained Transformer》论文翻译
猜你喜欢

【Programming】

【Random Erasing】《Random Erasing Data Augmentation》

What if the notebook computer cannot run the CMD command

常见CNN网络创新点

【双目视觉】双目立体匹配

Proof and understanding of pointnet principle

Translation of the paper "written mathematical expression recognition with bidirectionally trained transformer"

【MnasNet】《MnasNet:Platform-Aware Neural Architecture Search for Mobile》

【Mixup】《Mixup:Beyond Empirical Risk Minimization》

图片数据爬取工具Image-Downloader的安装和使用
随机推荐
【Wing Loss】《Wing Loss for Robust Facial Landmark Localisation with Convolutional Neural Networks》
【Wing Loss】《Wing Loss for Robust Facial Landmark Localisation with Convolutional Neural Networks》
【Programming】
Installation and use of image data crawling tool Image Downloader
Solve the problem of latex picture floating
【FastDepth】《FastDepth:Fast Monocular Depth Estimation on Embedded Systems》
Drawing mechanism of view (I)
【AutoAugment】《AutoAugment:Learning Augmentation Policies from Data》
label propagation 标签传播
Calculate the difference in days, months, and years between two dates in PHP
【MobileNet V3】《Searching for MobileNetV3》
Faster-ILOD、maskrcnn_ Benchmark training coco data set and problem summary
【Ranking】Pre-trained Language Model based Ranking in Baidu Search
论文写作tip2
Huawei machine test questions-20190417
【Hide-and-Seek】《Hide-and-Seek: A Data Augmentation Technique for Weakly-Supervised Localization xxx》
How do vision transformer work?【论文解读】
Use matlab to realize: chord cut method, dichotomy, CG method, find zero point and solve equation
Timeout docking video generation
【Programming】