当前位置:网站首页>Target detection for long tail distribution -- balanced group softmax
Target detection for long tail distribution -- balanced group softmax
2022-07-02 07:57:00 【MezereonXP】
Deal with long tailed target detection – Balanced Group Softmax
List of articles
This time I will introduce an article CVPR2020 The article , Titled “Overcoming Classifier Imbalance for Long-tail Object Detection with Balanced Group Softmax”, It mainly solves the problem of long tail data distribution in target detection , The solution is also very simple .
Long tailed data
First , Long tailed data exist widely , Here we use COCO and LVIS Take two data sets as examples , As shown in the figure below :

The abscissa is the index of the category , The ordinate is the number of samples in the corresponding category .
You can see , In these two data sets , There is an obvious long tail distribution .
Previous methods for dealing with long tail distribution
Here are some related works , Given by category :
- Resampling based on data (data re-sampling)
- Oversampling the tail data :Borderline-smote: a new over-sampling method in im- balanced data sets learning
- Delete the header data :class imbalance, and cost sensitivity: why under-sampling beats over sampling
- Sampling based on category balance :Exploring the limits of weakly supervised pretraining.
- Cost sensitive learning (cost- sensitive learning)
- Through to loss Adjustment , Give different weights to different categories
These methods are usually sensitive to hyperparameters , And poor performance when migrating to the detection framework ( The difference between classification task and detection task )
Balanced Group Softmax
Here is the specific framework of the algorithm :

As shown in the figure above , In the training phase , We will group the categories , Calculate separately in different groups Softmax, Then calculate the respective cross entropy error .
For grouping , The paper is given by 0,10,100,1000,+inf As a segmentation point
Here we need to add one for each group other Category , bring , When the target category is not in a group ,groundtruth Set to other.
The final error form is :
L k = − ∑ n = 0 N ∑ i ∈ G n y i n log ( p i n ) \mathcal{L}_k=-\sum_{n=0}^{N}\sum_{i\in \mathcal{G}_n}y_i^n\log (p_i^n) Lk=−n=0∑Ni∈Gn∑yinlog(pin)
among , N N N It's the number of groups , G n \mathcal{G}_n Gn It's No n n n Category collection of groups , p i n p_i^n pin Is the probability of model output , y i n y_i^n yin Is the label .
Effect evaluation
Here is a precision table for comprehensive comparison

AP The subscript of corresponds to the index of the divided group , You can see , Precision in the tail , That is to say A P 1 AP_1 AP1 and A C C 1 ACC_1 ACC1 It has reached SOTA Performance of .
边栏推荐
- Memory model of program
- ABM thesis translation
- Replace self attention with MLP
- 【Programming】
- How gensim freezes some word vectors for incremental training
- 【Cutout】《Improved Regularization of Convolutional Neural Networks with Cutout》
- 【MnasNet】《MnasNet:Platform-Aware Neural Architecture Search for Mobile》
- 【学习笔记】反向误差传播之数值微分
- Yolov3 trains its own data set (mmdetection)
- 【Mixup】《Mixup:Beyond Empirical Risk Minimization》
猜你喜欢

open3d学习笔记四【表面重建】

用全连接层替代掉卷积 -- RepMLP

Graph Pooling 简析

论文写作tip2
![[mixup] mixup: Beyond Imperial Risk Minimization](/img/14/8d6a76b79a2317fa619e6b7bf87f88.png)
[mixup] mixup: Beyond Imperial Risk Minimization
![[learning notes] numerical differentiation of back error propagation](/img/1c/e28e31d7cc5ccc38607c7839ccc5f0.png)
[learning notes] numerical differentiation of back error propagation

【AutoAugment】《AutoAugment:Learning Augmentation Policies from Data》

【双目视觉】双目矫正

Eklavya -- infer the parameters of functions in binary files using neural network

Income in the first month of naked resignation
随机推荐
conda常用命令
【Hide-and-Seek】《Hide-and-Seek: A Data Augmentation Technique for Weakly-Supervised Localization xxx》
C#与MySQL数据库连接
Thesis writing tip2
Faster-ILOD、maskrcnn_ Benchmark installation process and problems encountered
解决latex图片浮动的问题
Feature Engineering: summary of common feature transformation methods
Replace convolution with full connection layer -- repmlp
Implementation of yolov5 single image detection based on onnxruntime
[binocular vision] binocular stereo matching
【Programming】
[Sparse to Dense] Sparse to Dense: Depth Prediction from Sparse Depth samples and a Single Image
How do vision transformer work?【论文解读】
【Cutout】《Improved Regularization of Convolutional Neural Networks with Cutout》
MoCO ——Momentum Contrast for Unsupervised Visual Representation Learning
Remplacer l'auto - attention par MLP
Correction binoculaire
Open3d learning notes II [file reading and writing]
PPT的技巧
Latex formula normal and italic