当前位置:网站首页>Classic model alexnet
Classic model alexnet
2022-06-26 03:28:00 【On the right is my goddess】
List of articles
Introduction
To identify more images , We need a model with strong learning ability .CNN Is a good choice . As its depth and width increase , Its performance has also been greatly improved . meanwhile , Compared with feedforward neural network ,CNN With fewer connections and parameters .
With the improvement of computing power ,CNN There is a place for use , That's why we do CNN Why .
This article is based on CNN Designed a new network , contain 5 Convolutions and 3 All connection layers . meanwhile , To avoid overfitting , Using some special techniques .
The Dataset
Cut directly 256 × 256 256\times 256 256×256 Size image , As input .
Milepost : Directly in raw image Training on .
The Architecture
ReLU Nonlinearity
tanh and sigmoid There is the problem of gradient saturation ,ReLU There is no such problem , So the latter can help the model converge faster .
ReLU It's not necessarily good , It is simple and easy to use .
Local Response Normalization(LRN)
ReLU There is no need to standardize the input to avoid saturation , But still do one Normalization Words , The effect is still improved .
This is because ReLU There is no range , So we should normalize the results .
Create a competitive mechanism for the activity of local neurons , Make the value with larger response become larger , Suppress neurons with small feedback .
( This part is meaningless )
Overlapping Pooling
To the traditional Pooling Made a change . Generally, the step size is consistent with the pool size , Here let the step size be smaller than the size , Make the output of the pooling layer overlap , Enhance the richness of features .
Overall Architecture

The figure contains two AlexNet, In the third convolution, the data will be coupled with each other .

Every GPU Training half of the channel .
As the depth of the network increases , Compress the spatial information slowly , At this time, the semantic information increases slowly .
The full connection layer part will also couple the data , Finally, it is spliced as 4096 The vector of the dimension , And then compress it to 1000 dimension , Representative for 1000 Confidence level of two categories .
Just to be able to train , So it is divided into two parts for training , It's not really necessary .
Reduce Overfitting
Data Augmentation
- from 256x256 Buckle a piece in the middle 224x224 Size image ( He said there was 2048 Seed buckle method , But in fact, they are almost the same );
- hold RGB Make some changes on the channel , It was used PCA Methods .
Dropout
Randomly change some output of the hidden layer to 0, The probability is 50%.
In this article, it is considered that this is a different model obtained from each training and then fusion . But it's not .
It is pointed out that ,Dropout It approximates a linear model L2 The regularization .
In this paper Dropout On top of the two full connection layers .
current CNN Lost the full connection layer in , therefore Dropout Not many are used , But actually , Its presence RNN、Attention Such a fully connected structure is still very common 、 Very easy to use. .
Details of learning
- SGD;
- weight decay( Manually determine when the learning rate decreases epoch, Now it is generally to increase slowly and then decrease slowly );
- momentum( Avoid that the optimized surface is not so smooth and falls into a pit );
- The standard deviation is 0.01、 The mean for 0 The Gaussian distribution of is used to initialize the weight ; The offset is initialized to 1;
Results
At the end of the article, the effects of the two experiments are shown . The first image shows some pictures and their classification results . The second picture is to AlexNet Finally, take out the eigenvector , Group similar images together . The second experiment found that the result of classification was very good , This illustrates the AlexNet The acquired features are well learned .
summary
AlexNet The contribution of :
- Put forward AlexNet;
- ReLU、LRN、Pooling;
- Data augmentation (2 Kind of );
- Training details (4 individual )
边栏推荐
- 工业机器人之“慧眼”——机器视觉
- Partition, column, list
- On virtual memory and oom in project development
- Cliquez sur le bouton action de la liste pour passer à une autre page de menu et activer le menu correspondant
- GStreamer allocator and pool
- Qixia fire department carries out fire safety training on construction site
- Class diagram
- 网络PXE启动WinPE,支持UEFI和LEGACY引导
- QT compilation error: unknown module (s) in qt: script
- Modifying table names, deleting tables, obtaining table information, and deleting table records on specified dates for common MySQL statements
猜你喜欢

Notes on the 3rd harmonyos training in the studio

培育项目式Steam教育理念下的儿童创造力

Business process diagram design

多媒体元素,音频、视频

Learn from Taiji makers - mqtt (V) publish, subscribe and unsubscribe

Analysis of technological changes in social robots

Analysis of the multiple evaluation system of children's programming

【论文笔记】Supersizing Self-supervision: Learning to Grasp from 50K Tries and 700 Robot Hours

点击事件

图扑软件数字孪生海上风电 | 向海图强,奋楫争先
随机推荐
HL7Exception: Can‘t XML-encode a GenericMessage. Message must have a recognized struct
[reading papers] fbnetv3: joint architecture recipe search using predictor training network structure and super parameters are all trained by training parameters
stm32Cubemx:看门狗------独立看门狗和窗口看门狗
Route jump: click the operation button of the list to jump to another menu page and activate the corresponding menu
MySQL开发环境
计组笔记 数据表示与运算 校验码部分
Scratch returns 400
MySQL development environment
P2483-[模板]k短路/[SDOI2010]魔法猪学院【主席树,堆】
Class diagram
Analysis of technological changes in social robots
计组笔记——CPU的指令流水
Classic quotations from "human nature you must not know"
js array数组json去重
Analysis on the diversification of maker space mechanism construction
项目部署遇到的问题-生产环境
golang正則regexp包使用-06-其他用法(特殊字符轉換、查找正則共同前綴、切換貪婪模式、查詢正則分組個數、查詢正則分組名稱、用正則切割、查詢正則字串)
使用IDEA画结构图
拖放
Analysis of the multiple evaluation system of children's programming