当前位置:网站首页>Complete deep neural network CNN training with tensorflow to complete picture recognition case 2
Complete deep neural network CNN training with tensorflow to complete picture recognition case 2
2022-07-03 13:27:00 【Haibao 7】
To be continued . Previous link :https://blog.csdn.net/dongbao520/article/details/125456950
Convolutional neural networks
• Convolutional neural networks
• Visual cortex 、 Feel the field , Some neurons see the line , Some neurons see the line
Direction , Some neurons have larger receptive fields , Combine the patterns on the bottom
• 1998 year Yann LeCun Et al LeNet-5 framework , Widely used in hands
Written digit recognition , Including full connection layer and sigmoid Activation function , There are also volumes
Accumulation layer and pool layer
Convolutional neural networks (Convolutional Neural Networks, CNN) It is a kind of feedforward neural network with convolution calculation and depth structure (Feedforward Neural Networks), It's deep learning (deep learning) One of the representative algorithms of [1-2] . Convolutional neural network has the characteristics of representation learning (representation learning) Ability , The input information can be classified according to its hierarchical structure (shift-invariant classification), So it's also called “ Translation invariant artificial neural networks (Shift-Invariant Artificial Neural Networks, SIANN
Convolution neural network imitates biological visual perception (visual perception) Mechanism construction , Supervised learning and unsupervised learning , The sharing of convolution kernel parameters in the hidden layer and the sparsity of inter layer connections make the convolution neural network lattice with less computation (grid-like topology) features , For example, pixels and audio for learning 、 It has a stable effect and has no additional feature engineering on the data (feature engineering) Complete principle related requirements can ---->> Reference resources
For receptive field :
For pre trained networks
Reuse TensorFlow Model 
CNN The most important building unit is the convolution layer
• Neurons in the first convolution layer are not connected to every pixel of the input picture ,
Just connect the pixels of their receptive field , And so on , Of the second accretion layer
Each neuron is only connected to a small square God located in the first convolution layer
Jing Yuan
Convolution layer diagram 

Convolution cases :


In steps of 2, Then there are 
Filter Convolution kernel
• Convolution kernels
• Vertical line filter The middle column is 1, The surrounding areas are listed as 0
• Horizontal line filter Intermediate behavior 1, Surrounding behavior 0
• 7*7 matrix

In a feature map , All neurons share the same parameters (
weights bias), Weight sharing
• Different feature maps have different parameters

Convolution training process 
Padding Pattern
VALID
• Do not apply zero padding, It is possible to ignore the right or bottom of the picture , This depends stride Set up
• SAME
• If necessary, add zero padding, In this case , The number of output neurons is equal to the number of input neurons divided by the step size ceil(13/5)=3

Pooling Pooling Handle
The goal is downsampling subsample,shrink, Reduce the calculated load , Memory usage , The number of arguments ( It can also prevent over fitting )• Reducing the size of the input image also allows the neural network to withstand a little image translation , Not affected by location
• Just like convolutional neural networks , Each neuron in the pooling layer is connected to the neuron output in the upper layer , It only corresponds to a small area of receptive field . We have to define size , step ,padding type
• Pooled neurons have no weight value , It just aggregates the input according to the maximum or the average
• 2*2 The pooled core of , In steps of 2, There is no filling , Only the maximum value is passed down

Twice as long and twice as wide , area 4 Times smaller , lose 75% The input value of
• In general , The pooling layer works on each independent input channel , So the depth of output is the same as that of input
CNN framework
• Typical CNN The architecture heap lists some volume layers :
• Usually a convolution layer is followed by ReLU layer , Then there is a pool layer , Then there are other convolutions +ReLU layer , Then another pooling layer , The pictures transmitted through the network are getting smaller and smaller , But it's getting deeper and deeper , For example, more feature maps !
• Finally, the conventional feedforward neural network is added , By some fully connected layers +ReLU layers , Finally, the output layer prediction , For example, one softmax Class probability of layer output prediction
• A common misconception is that the convolution kernel is too large , You can use and 99 Two of the same effect of the nucleus 33 The core of , The advantage is that there will be fewer parameters , Simplify the operation .
To be continued ..
边栏推荐
- Father and basketball
- Solve system has not been booted with SYSTEMd as init system (PID 1) Can‘t operate.
- rxjs Observable filter Operator 的实现原理介绍
- untiy世界边缘的物体阴影闪动,靠近远点的物体阴影正常
- Kivy教程之 盒子布局 BoxLayout将子项排列在垂直或水平框中(教程含源码)
- 2022-02-11 practice of using freetsdb to build an influxdb cluster
- PowerPoint tutorial, how to save a presentation as a video in PowerPoint?
- 双链笔记 RemNote 综合评测:快速输入、PDF 阅读、间隔重复/记忆
- JS convert pseudo array to array
- 35道MySQL面试必问题图解,这样也太好理解了吧
猜你喜欢

This math book, which has been written by senior ml researchers for 7 years, is available in free electronic version
[email protected] chianxin: Perspective of Russian Ukrainian cyber war - Security confrontation and sanctions g"/>Start signing up CCF C ³- [email protected] chianxin: Perspective of Russian Ukrainian cyber war - Security confrontation and sanctions g

Can newly graduated European college students get an offer from a major Internet company in the United States?
![[colab] [7 methods of using external data]](/img/cf/07236c2887c781580e6f402f68608a.png)
[colab] [7 methods of using external data]

stm32和电机开发(从mcu到架构设计)

Flick SQL knows why (10): everyone uses accumulate window to calculate cumulative indicators

Kivy教程之 如何自动载入kv文件

MyCms 自媒体商城 v3.4.1 发布,使用手册更新

【电脑插入U盘或者内存卡显示无法格式化FAT32如何解决】

IDEA 全文搜索快捷键Ctr+Shift+F失效问题
随机推荐
父亲和篮球
[Database Principle and Application Tutorial (4th Edition | wechat Edition) Chen Zhibo] [Chapter V exercises]
JSON serialization case summary
71 articles on Flink practice and principle analysis (necessary for interview)
Finite State Machine FSM
The shortage of graphics cards finally came to an end: 3070ti for more than 4000 yuan, 2000 yuan cheaper than the original price, and 3090ti
My creation anniversary: the fifth anniversary
Resolved (error in viewing data information in machine learning) attributeerror: target_ names
Reptile
R语言使用data函数获取当前R环境可用的示例数据集:获取datasets包中的所有示例数据集、获取所有包的数据集、获取特定包的数据集
正则表达式
Flink SQL knows why (XIV): the way to optimize the performance of dimension table join (Part 1) with source code
Logback log framework
Open PHP error prompt under Ubuntu 14.04
STM32 and motor development (from MCU to architecture design)
JS convert pseudo array to array
已解决TypeError: Argument ‘parser‘ has incorrect type (expected lxml.etree._BaseParser, got type)
Image component in ETS development mode of openharmony application development
双链笔记 RemNote 综合评测:快速输入、PDF 阅读、间隔重复/记忆
In the promotion season, how to reduce the preparation time of defense materials by 50% and adjust the mentality (personal experience summary)