Preface This paper proposes a new weakly supervised multi label classification (WSML) Method , This method rejects or corrects large loss samples , To prevent the model from remembering noisy labels . Because there are no heavy and complex components , The proposed method sets labels in several parts ( Include Pascal VOC 2012、MS COCO、NUSWIDE、CUB and OpenImages V3 Data sets ) Superior to the most advanced before WSML Method . Various analyses also show that , The practical effect of this method is very good , It is proved that it is very important to deal with the loss correctly in the weakly supervised multi label classification .
Welcome to the official account CV Technical guide , Focus on computer vision technology summary 、 The latest technology tracking 、 Interpretation of classic papers 、CV Recruitment information .
The paper :Large Loss Matters in Weakly Supervised Multi-Label Classification
The paper :http://arxiv.org/pdf/2206.03740
Code :https://github.com/snucml/LargeLossMatters
background
Weakly supervised multi label classification (WSML) The task is to use part of each image to observe labels to learn multi label classification , Because of its huge labeling cost , Becoming more and more important .
at present , There are two simple ways to train a model using partial tags . One is to use only observed tags to train the model , And ignore the unobserved labels . The other is to assume that all unobserved labels are negative , And incorporate it into your training , Because in multi label settings , Most labels are negative .
But the second method has one limitation , That is, this assumption will generate some noise in the tag , Thus hindering model learning , Therefore, most of the previous work followed the first method , And try to use various technologies ( Such as bootstrap or regularization ) Explore clues to unobserved tags . However , These methods include extensive calculations or complex optimization of pipelines .
Based on the above ideas , The author assumes that , If the label noise can be properly handled , The second approach may be a good starting point , Because it has the advantage of incorporating many real negative labels into model training . therefore , The author looks at it from the perspective of noise label learning WSML problem .
as everyone knows , When training models with noise labels , The model first adapts to clean labels , Then start remembering noise labels . Although previous studies have shown that memory effect only exists in noisy multi category classification scenes , But the author found that , The same effect also exists in noisy multi label classification scenarios . Pictured 1 Shown , During training , From the clean label ( True negative sample ) The loss value of is reduced from the beginning , And from the noise tag ( False negative sample ) The loss of is reduced from the middle .
chart 1 WSML Memory effect in
Based on this discovery , The author has developed three different schemes , By rejecting or correcting large loss samples during training , To prevent false positive labels from being memorized into the multi label classification model .
contribution
1) It is proved by experiments for the first time , Memory effect occurs in the process of multi label classification with noise .
2) A new weakly supervised multi label classification scheme is proposed , This scheme explicitly utilizes the learning technology with noise labels .
3) The proposed method is light and simple , The most advanced classification performance is achieved on various partial label datasets .
Method
In this paper , The author puts forward a new WSML Method , The motivation is based on the idea of noise multiclass learning , It ignores the huge loss in the process of model training . The weight term is further introduced into the loss function λi:
The authors propose three ways to provide weights λi Different schemes , The schematic diagram is described in Figure 2 Shown .
chart 2 The overall pipeline of the proposed method
1. Loss rejection
One way to handle large loss samples is by setting λi=0 To reject it . In multi class tasks with noise ,B.Han Et al. Proposed a method to gradually increase the rejection rate in the training process . The author also sets the function λi,
Because the model learns clean patterns at the initial stage , So in t=1 Do not reject any loss value . Use small batches instead of full batches in each iteration D′ To form a loss set . The author calls this method LL-R.
2. Loss correction ( temporary )
Another way to deal with a large loss sample is to correct it rather than reject it . In multi label settings , This can be easily achieved by switching the corresponding annotation from a negative value to a positive value .“ temporary ” The word means , It does not change the actual label , Only the loss calculated according to the modified label is used , Will function λi Defined as
The author named this method LL-Ct. The advantage of this method is , It increases the number of true positive tags in the tags that have never been observed .
3. Loss correction ( permanent )
Deal more aggressively with larger loss values by permanently correcting labels . Directly change the label from negative to positive , And use the modified tag during the next training . So , Define... For each case λi=1, And modify the label as follows :
The author named this method LL-Cp.
experiment
surface 2 Quantitative results of artificially created partial label data sets
surface 3 OpenImages V3 Quantitative results in the dataset
chart 3 Artificially generated COCO Qualitative results of some label datasets
chart 4 COCO Accuracy analysis of the proposed method on the dataset
chart 5 LL-Ct Yes COCO The hyperparametric effect of data sets
chart 6 Training with fewer images
surface 4 Pointing Game
Conclusion
In this paper , The author puts forward a loss modification scheme , This scheme rejects or corrects the large loss samples when training multi label classification models with partial label annotations . This comes from empirical observation , That is, the memory effect also occurs in noisy multi label classification scenarios .
Although it does not include heavy and complex components , But the author's scheme successfully prevents the multi label classification model from remembering false negative labels with noise , State of the art performance on a variety of partially labeled multi label datasets .
---------------------------------------------------------------------------------
CV The technical guide creates a computer vision technology exchange group and a free version of the knowledge planet , At present, the number of people on the planet has 700+, The number of topics reached 200+.
The knowledge planet will release some homework every day , It is used to guide people to learn something , You can continue to punch in and learn according to your homework .CV Every day in the technology group, the top conference papers published in recent days will be sent , You can choose the papers you are interested in to read , continued follow Latest technology , If you write an interpretation after reading it and submit it to us , You can also receive royalties . in addition , The technical group and my circle of friends will also publish various periodicals 、 Notice of solicitation of contributions for the meeting , If you need it, please scan your friends , And pay attention to .
Add groups and planets : Official account CV Technical guide , Get and edit wechat , Invite to join .
Welcome to the official account CV Technical guide , Focus on computer vision technology summary 、 The latest technology tracking 、 Interpretation of classic papers 、CV Recruitment information .

Welcome to contact me if you can write the following ( Get the contact information after following the official account ).
- Interpretation of the latest summit . For example, the most recent CVPR2022 The paper .
- Systematic overview of all directions 、 Development and evolution of main models 、 Innovative ideas, advantages and disadvantages of each model 、 Code analysis, etc . Such as the summary of target detection : A summary of target detection from traditional methods to deep learning , It mainly includes traditional method detection 、RCNN series 、YOLO series 、anchor-free series 、 Summary of small target detection methods 、 Summary of small sample target detection methods 、 Summary of object detection methods in video 、 Summary of loss function used in target detection . Support writing while learning .
- TVM Introduction to practice tutorial
- MNN Introduction to practice tutorial
- OpenVINO Introduction to practice tutorial
- libtorch Introduction to practice tutorial
- Oneflow Introduction to practice tutorial
- Detectron Introduction to practice tutorial
- caffe Source code reading
- Deep learning from introduction to mastery ( Start with convolutional neural networks )
- If you have something you want to write and it's not mentioned above , You can contact me . Statement : Paid , Please contact us for details .
Other articles of official account
Introduction to computer vision
Summary of common words in computer vision papers
YOLO Series carding ( Four ) About YOLO Deployment of
YOLO Series carding ( 3、 ... and )YOLOv5
YOLO Series carding ( Two )YOLOv4
YOLO Series carding ( One )YOLOv1-YOLOv3
CVPR2022 | Based on egocentric data OCR assessment
CVPR 2022 | Using contrast regularization method to deal with noise labels
CVPR2022 | Loss problem in weakly supervised multi label classification
CVPR2022 | iFS-RCNN: An incremental small sample instance divider
CVPR2022 | A ConvNet for the 2020s & How to design neural network Summary
CVPR2022 | PanopticDepth: A unified framework for depth aware panoramic segmentation
CVPR2022 | Reexamine pooling : Your feeling field is not ideal
CVPR2022 | Unknown target detection module STUD: Learn about unknown targets in the video
CVPR2022 | Ranking based siamese Visual tracking
Build from scratch Pytorch Model tutorial ( Four ) Write the training process -- Argument parsing
Build from scratch Pytorch Model tutorial ( 3、 ... and ) build Transformer The Internet
Build from scratch Pytorch Model tutorial ( Two ) Build network
Build from scratch Pytorch Model tutorial ( One ) data fetch
Some personal thinking habits and thought summary about learning a new technology or field quickly
CVPR2022 | More related articles on the loss problem in weakly supervised multi label classification
- CVPR2020: Multi-path region mining for point cloud weakly supervised 3D semantic segmentation
CVPR2020: Multi-path region mining for point cloud weakly supervised 3D semantic segmentation Multi-Path Region Mining for Weakly Supervised 3D Semantic Segmentation ...
- Valse2019 note —— Weakly supervised visual understanding
Cheng Mingming ( Nankai University ): Adaptive visual perception for open environment ( The picture is from valse2019 Mr. Cheng Mingming ppt) Common technology of neural network for recognition and understanding General architecture of deep neural network -- VggNet(ICLR'15).ResN ...
- Change numerous for brief , Weak supervision is a new trend in the field of target orientation SOTA - Pseudo supervised target location method (PSOL) | CVPR 2020
This paper proposes a method of pseudo supervised target location (PSOL) To solve the problem of weak supervision target location method , This method separates location and classification into two independent networks , Then use it on the training set Deep descriptor transformation(DDT ...
- scikit-learn General example 8 : Multi label classification
This example simulates a multi label document classification problem . The dataset is randomly generated based on the following processing : Number of selected tags : Poisson (n~Poisson,n_labels) n Time , Select category C: polynomial (c~Multinomial,theta) Select text ...
- Mobile front end has to understand HTML5 head Head tag ( Middle part I )
Meta label meta The label is HTML in head An auxiliary label on the head , It is located in HTML Of the document header <head> and <title> Between the marks , It provides information that is not visible to users . Although this part of the letter ...
- html( Commonly used labels , Label classification ), Page template , CSS(css Three ways to introduce ), Priority of the three methods of introduction
HTML Markup language is a non programming language responsible for completing the structure of the page form : label : By <> The package begins with a letter , You can combine legal characters ( -| Numbers ), Special symbols that can be parsed by browsers , The label has a beginning and an end Instructions : By <> package ...
- front end HTML Label classification
Three : 1. Block level label : Monopolize one line , Width can be set , Height . If you set the width and height , The current width and height . If the width and height are not set , The width is the width of the parent box , The height is filled in according to the content . 2. In line labels : Show... In one line , Can't set width , high ...
- Supervised text classification algorithm FastText
The algorithm consists of facebook stay 2016 In open source , The typical application scenario is “ Supervised text classification problem ”. Model The optimization objectives of the model are as follows : among ,$<x_n,y_n>$ It's a training sample ,$y_n$ It's a training goal ,$ ...
- How to use softmax and sigmoid To do multi classification and multi label classification
First , Let's talk about the difference between multi category classification and multi label classification Multi label classification : A sample can belong to multiple categories ( Or tags ), Different classes are related , For example, a text is divided into “ figure ” and “ Sports people ” Two labels . Obviously these two tags are not mutually exclusive , It's connected ...
- Introduction to the front html( Common labels and label classification )
day47 Reference resources :https://www.cnblogs.com/liwenzhou/p/7988087.html Any tag has three properties :ID,class.style <!DOCTYPE ht ...
Random recommendation
- ASP.NET 5 Use TestServer Unit test
If you were right ASP.NET WebAPI Unit test (HttpClient Initiate request , And it can be debugged WebAPI), It is generally used Owin The way , Specific reference :< Development Notes : use Owin Host Realization ...
- .split() Function usage
split explain split(): Split string . Slice a string by specifying a separator , And returns a list of split strings (list) split grammar str.split('type',num)[n] The whole grammar means : With ...
- [Effective JavaScript note ] The first 64 strip : Using recursion for asynchronous loops
Suppose you need to have a function like this , Receive one URL And try to download each file in turn until one file is successfully downloaded . If API It's synchronous , Using loops is easy to implement . function downloadOneSync(urls){ f ...
- const modification
const int A() //const // ====>int A(const this) { // Point of view 1:const It's embellishment a, But pass the test , We found that ,b++ It doesn't compile // This explanation :co ...
- SQL Server Aggregate functions
Aggregate functions return a single value after calculating a set of values . except count( Count the number of items ) Function , Other aggregate functions ignore null values in their expressions (null). All aggregate functions are deterministic functions . That is, at any time using the same set of input values to call the aggregate function after execution ...
- jQ Implementation of a rotating map
as everyone knows , The rotation chart is widely used . We can see the rotation chart on many websites , For example, Taobao. . Jingdong these websites are very common . Now let's start our rotation tour : Build our skeleton : <!DOCTYPE html> < ...
- [PHP] PHP Server interface SAPI Structural body in
SAPI: The same conventions are followed between the server abstraction layers , Here we call it SAPI Interface . For example, the implementation of command line program ,Apache Of mod_php Module implementation and fastcgi And so on 1. Structure : Use structure (Struc ...
- logrotate- Example log cutting
logrotate yes linux The tools that come with the system , It can automatically truncate the log ( Or rotation ). Compress and delete old log files . 1) Sample configuration file # cat /wls/wls81/bin/weblogic/wls/app ...
- sql A field stores multiple... Of another table id Values separated by commas , Based on id Go to Chinese and splice with commas
First, I will introduce the two functions used charindex( The expression to find 1, expression 2), The return value is the expression 1 In the expression 2 Subscript in , Return if not found 0.(sql The subscript of is from 1 At the beginning ), for example select charindex('s ...
- json Conversion tool class :json<===>list Or object
public class JsonTools { /** * POJO turn JSON */ public static String createJsonString(Object object) { ...








![spacy.load(“en_core_web_sm“)###OSError: [E050] Can‘t find model ‘en_core_web_sm‘.](/img/f5/e6e480f69481bef826b155fab57669.png)
