当前位置:网站首页>Cs231n notes (medium) -- applicable to 0 Foundation
Cs231n notes (medium) -- applicable to 0 Foundation
2022-07-05 15:55:00 【Small margin, rush】
Catalog
Pooling layer - Pooling layer - Convergence layer
Convolutional neural networks

Conventional neural networks : The input of neural network is a vector , Then in a series of Cryptic layer Change it in . Each hidden layer is composed of several neurons , Are connected to all neurons in the previous layer . In a hidden layer , Neurons are independent of each other and do not make any connections . The last full connection layer is called “ Output layer ”, In the classification problem , Its output value is regarded as the rating value of different categories .
Convolutional neural network aims at the situation that the input is all images , Adjust the structure more reasonably , Its Neurons are 3 Dimensional arrangement : Width 、 Height and depth , Neurons in the layer will only connect to a small area in the previous layer , Instead of a full connection .
Network structure
Convolution layer
With parameters , There are some learnable filter sets , Each filter is in space ( Width and height ) They are all relatively small , But the depth is the same as the input data .
When it comes to forward propagation , Each filter slides across the width and height of the input data ( More precisely, convolution ), Then calculate the inner product of the whole filter and any part of the input data , Generate a 2 Activation diagram of dimension (activation map), The activation diagram shows the response of the filter at each spatial location .
Feel the field : The neurons in the latter layer are in the sensory space of the neurons in the former layer It can also be defined as the area size mapped by the pixel points on the feature map of each layer in the convolutional neural network in the original image

Hyperparameters :
depth , step ( When sliding the filter , Moving pixel bits ), Zero fill ( use 0 Fill at the edge )
The spatial size of the output data body can be obtained by inputting the size of the data body (W), The size of receptive field of neurons in convolution layer (F), step (S) And the number of zero fills (P) To calculate .
(W-F +2P)/S+1
Parameters of the Shared : Used to control the number of parameters . If a feature is useful , You can share it
demonstration
- The size of the input data body is

- 4 A super parameter : The number of filters K, Space size of filter F, step S. Zero fill quantity P
- Width and height of output dimension :
, among W and H by (W-F +2P)/S+1,D2=k
Convolution is essentially a dot product between the filter and the local area of the input data .
1x1 Convolution : The depth of the filter and the input data body is the same , For input yes [32x32x3], that 1x1 Convolution is efficient 3 Dimensional point product
Expansion convolution : There are some element gaps in the filter , The receptive field increases
Pooling layer - Pooling layer - Convergence layer
A convergence layer is periodically inserted between successive convolution layers , Gradually reduce the spatial size of the data body .
- There are two super parameters : The size F, step S, The size calculation of output data is the same as that of convolution layer
- Zero padding is rarely used in the convergence layer
Maximum convergence

Fully connected layer
In the full connection layer , Neurons are fully connected to all activation data in the previous layer , Refer to the conventional neural network
The only difference between the full connection layer and the convolution layer is that the neurons in the convolution layer are only connected to a local area in the input data , And the neurons in the convolution column share parameters . The full connection layer can be transformed into a convolution layer
The arrangement of layers
INPUT -> [[CONV -> RELU]*N -> POOL?]*M -> [FC -> RELU]*K -> FC
The combination of several small filter convolution layers is better than one large filter convolution layer
A structure in which multiple convolution layers alternate with nonlinear activation layers , It can extract deeper and better features than the structure of a single convolution layer . The disadvantage is , In back propagation , The convolution layer in the middle may lead to more memory consumption .
Q&A:
- Why use zero padding ? Using zero padding can keep the output data of the convolution layer and the input data unchanged in the spatial dimension in addition to the above mentioned
- Why use in convolution layer 1 Step size of ? Smaller step size is better , In steps of 1 It can let the spatial dimension downsampling be all in the charge of the convergence layer , The convolution layer is only responsible for transforming the depth of the input data volume .
common CNN
LeNet: The first convolutional neural network
AlexNet:ImageNet2012 champion , Network structure and LeNet similar , Use the convolution layer to obtain features
ZF Net:2013 Champion can modify the super parameters in the structure to realize AlexNet Improvement of , Specifically, the size of the intermediate convolution layer is increased , Make the step size and filter size of the first layer smaller
GoogLeNet: Significantly reduce the number of parameters in the network ,Inception.
VGGNet:2014 runner-up , It shows that the depth of the network is the key part of the excellent performance of the algorithm . Consume more computing resources , And more parameters are used , Cause more memory usage
ResNet:2015 champion , Used a lot Batch normalization
边栏推荐
- 异常com.alibaba.fastjson.JSONException: not match : - =
- Bugku's eyes are not real
- lv_font_conv离线转换
- 写单元测试的时候犯的错
- Value series solution report
- SQL injection sqllabs (basic challenges) 11-20
- Bugku cyberpunk
- Noi / 1.4 07: collect bottle caps to win awards
- 20.[STM32]利用超声波模块和舵机实现智能垃圾桶功能
- 【 note 】 résoudre l'erreur de code IDE golang
猜你喜欢

通过的英特尔Evo 3.0整机认证到底有多难?忆联科技告诉你

OSI seven layer model

Transfer the idea of "Zhongtai" to the code
![17. [stm32] use only three wires to drive LCD1602 LCD](/img/c6/b56c54da2553a451b526179f8b5867.png)
17. [stm32] use only three wires to drive LCD1602 LCD
![16. [stm32] starting from the principle, I will show you the DS18B20 temperature sensor - four digit digital tube displays the temperature](/img/9f/c91904b6b1d3a1e85c0b50e43972e5.jpg)
16. [stm32] starting from the principle, I will show you the DS18B20 temperature sensor - four digit digital tube displays the temperature
![21. [STM32] I don't understand the I2C protocol. Dig deep into the sequence diagram to help you write the underlying driver](/img/f4/2c935dd9933f5cd4324c29c41ab221.png)
21. [STM32] I don't understand the I2C protocol. Dig deep into the sequence diagram to help you write the underlying driver

Six common transaction solutions, you sing, I come on stage (no best, only better)

研发效能度量指标构成及效能度量方法论

六种常用事务解决方案,你方唱罢,我登场(没有最好只有更好)

First PR notes
随机推荐
复现Thinkphp 2.x 任意代码执行漏洞
Arduino控制微小的六足3D打印机器人
Maximum common subsequence
RepLKNet:不是大卷积不好,而是卷积不够大,31x31卷积了解一下 | CVPR 2022
项目sql中批量update的时候参数类型设置错误
Data communication foundation - routing communication between VLANs
I spring and autumn blasting-1
18.[stm32] read the ROM of DS18B20 temperature sensor and realize multi-point temperature measurement
助力数字经济发展,夯实数字人才底座—数字人才大赛在昆成功举办
Six common transaction solutions, you sing, I come on stage (no best, only better)
lv_ font_ Conv offline conversion
OceanBase社区版之OBD方式部署方式本地安装
Example project: simple hexapod Walker
DataArts Studio数据架构——数据标准介绍
19.[STM32]HC_SR04超声波测距_定时器方式(OLED显示)
episodic和batch的定义
Interval DP (gravel consolidation)
I spring web upload
Severlet learning foundation
Quick completion guide for manipulator (IX): forward kinematics analysis

, among W and H by