当前位置:网站首页>【Mixed Pooling】《Mixed Pooling for Convolutional Neural Networks》
【Mixed Pooling】《Mixed Pooling for Convolutional Neural Networks》
2022-07-02 07:44:00 【bryant_ meng】


RSKT-2014
International conference on rough sets and knowledge technology
List of articles
1 Background and Motivation
The effect of pooling layer ( Read all the in-depth learning in one article 9 A pooling method !)
- Increase the network receptive field
- Suppress noise , Reduce information redundancy
- Reduce the amount of model calculation , Reduce the difficulty of network optimization , Prevent over fitting of network
- It makes the model more robust to the change of feature position in the input image
Author for max and ave pooling The shortcomings of ,

Put forward mix pooling——randomly employs the local max pooling and average pooling methods when training CNNs
2 Review of Convolutional Neural Networks

- Convolutional Layer, Including convolution operation and activation function
- Non-linear Transformation Layer, That is to say normalization layer , What's more popular now is BN etc. , It used to be LCN(local contrast normalization) and AlexNet Of LRN(the local response normalization) etc. ,PS: In the paper LCN There is something wrong with the formula ,LRN There is also a gap between the details and the original paper , The form is basically the same
- Feature Pooling Layer
3 Advantages / Contributions
reference dropout, blend max and ave Pooling , Put forward mixed pooling
4 Method
1)mixed pooling The formula 
λ \lambda λ is a random value being either 0 or 1
2)mixed pooling Back propagation
Have a look first max and ave pooling Back propagation of
max pooling
( This is from the Internet , Invasion and deletion !!!)
ave pooling
( This is from the Internet , Invasion and deletion !!!)
mixed pooling
I have to record it λ \lambda λ The value of , In order to correctly back spread
the pooling history about the random value λ \lambda λ in Eq. must be recorded during forward propagation.
3)Pooling at Test Time
Statistics training time pooling use max and ave Frequency of F m a x k F_{max}^{k} Fmaxk and F a v e k F_{ave}^{k} Favek, Whose frequency is high? When testing there pooling Just use who , Start metaphysics, right , Ha ha ha ha

5 Experiments
5.1 Datasets

- CIFAR-10
- CIFAR-100
- SVHN
5.2 Experimental Results
1)CIFAR-10
train error high ,acc high
The author explains This indicates that the proposed mixed pooling outperforms max pooling and average pooling to address the over-fitting problem
Visualization results 
It can be seen that mixed pooling Contains more information
2)CIFAR-100
3)SVHN
4)Time Performance
6 Conclusion(own) / Future work
LRN

k , n , α , β k, n, \alpha, \beta k,n,α,β It's all super parameters , a , b a,b a,b Input and output characteristic diagram , x , y x,y x,y Space location , i i i Channel location
The following is from Local response normalization of deep learning LRN(Local Response Normalization) understand

import tensorflow as tf
import numpy as np
x = np.array([i for i in range(1,33)]).reshape([2,2,2,4])
y = tf.nn.lrn(input=x,depth_radius=2,bias=0,alpha=1,beta=1)
with tf.Session() as sess:
print(x)
print('#############')
print(y.eval())


LCN
《What is the best multi-stage architecture for object recognition?》
边栏推荐
- TimeCLR: A self-supervised contrastive learning framework for univariate time series representation
- A slide with two tables will help you quickly understand the target detection
- Implementation of yolov5 single image detection based on pytorch
- Common machine learning related evaluation indicators
- 论文写作tip2
- Faster-ILOD、maskrcnn_benchmark训练coco数据集及问题汇总
- The difference and understanding between generative model and discriminant model
- What if the laptop task manager is gray and unavailable
- PointNet原理证明与理解
- [mixup] mixup: Beyond Imperial Risk Minimization
猜你喜欢

Alpha Beta Pruning in Adversarial Search

【Paper Reading】

点云数据理解(PointNet实现第3步)

Implementation of purchase, sales and inventory system with ssm+mysql

【Sparse-to-Dense】《Sparse-to-Dense:Depth Prediction from Sparse Depth Samples and a Single Image》

How do vision transformer work?【论文解读】
![[paper introduction] r-drop: regulated dropout for neural networks](/img/09/4755e094b789b560c6b10323ebd5c1.png)
[paper introduction] r-drop: regulated dropout for neural networks

Record of problems in the construction process of IOD and detectron2

The difference and understanding between generative model and discriminant model

Faster-ILOD、maskrcnn_benchmark训练coco数据集及问题汇总
随机推荐
[mixup] mixup: Beyond Imperial Risk Minimization
【FastDepth】《FastDepth:Fast Monocular Depth Estimation on Embedded Systems》
Convert timestamp into milliseconds and format time in PHP
[introduction to information retrieval] Chapter 1 Boolean retrieval
Regular expressions in MySQL
Mmdetection installation problem
SSM second hand trading website
How to efficiently develop a wechat applet
Thesis writing tip2
ABM thesis translation
【Mixed Pooling】《Mixed Pooling for Convolutional Neural Networks》
Ppt skills
机器学习理论学习:感知机
论文写作tip2
[introduction to information retrieval] Chapter 3 fault tolerant retrieval
What if the laptop task manager is gray and unavailable
Win10+vs2017+denseflow compilation
Spark SQL task performance optimization (basic)
【Mixup】《Mixup:Beyond Empirical Risk Minimization》
【BiSeNet】《BiSeNet:Bilateral Segmentation Network for Real-time Semantic Segmentation》