Preface   This paper proposes a new weakly supervised multi label classification (WSML) Method , This method rejects or corrects large loss samples , To prevent the model from remembering noisy labels . Because there are no heavy and complex components , The proposed method sets labels in several parts ( Include Pascal VOC 2012、MS COCO、NUSWIDE、CUB and OpenImages V3 Data sets ) Superior to the most advanced before WSML Method . Various analyses also show that , The practical effect of this method is very good , It is proved that it is very important to deal with the loss correctly in the weakly supervised multi label classification .

Welcome to the official account CV Technical guide , Focus on computer vision technology summary 、 The latest technology tracking 、 Interpretation of classic papers 、CV Recruitment information .

The paper :Large Loss Matters in Weakly Supervised Multi-Label Classification

The paper :http://arxiv.org/pdf/2206.03740

Code :https://github.com/snucml/LargeLossMatters

background

Weakly supervised multi label classification (WSML) The task is to use part of each image to observe labels to learn multi label classification , Because of its huge labeling cost , Becoming more and more important .

at present , There are two simple ways to train a model using partial tags . One is to use only observed tags to train the model , And ignore the unobserved labels . The other is to assume that all unobserved labels are negative , And incorporate it into your training , Because in multi label settings , Most labels are negative .

But the second method has one limitation , That is, this assumption will generate some noise in the tag , Thus hindering model learning , Therefore, most of the previous work followed the first method , And try to use various technologies ( Such as bootstrap or regularization ) Explore clues to unobserved tags . However , These methods include extensive calculations or complex optimization of pipelines .

Based on the above ideas , The author assumes that , If the label noise can be properly handled , The second approach may be a good starting point , Because it has the advantage of incorporating many real negative labels into model training . therefore , The author looks at it from the perspective of noise label learning WSML problem .

as everyone knows , When training models with noise labels , The model first adapts to clean labels , Then start remembering noise labels . Although previous studies have shown that memory effect only exists in noisy multi category classification scenes , But the author found that , The same effect also exists in noisy multi label classification scenarios . Pictured 1 Shown , During training , From the clean label ( True negative sample ) The loss value of is reduced from the beginning , And from the noise tag ( False negative sample ) The loss of is reduced from the middle .

chart 1 WSML Memory effect in

Based on this discovery , The author has developed three different schemes , By rejecting or correcting large loss samples during training , To prevent false positive labels from being memorized into the multi label classification model .

contribution

1) It is proved by experiments for the first time , Memory effect occurs in the process of multi label classification with noise .

2) A new weakly supervised multi label classification scheme is proposed , This scheme explicitly utilizes the learning technology with noise labels .

3) The proposed method is light and simple , The most advanced classification performance is achieved on various partial label datasets .

Method

In this paper , The author puts forward a new WSML Method , The motivation is based on the idea of noise multiclass learning , It ignores the huge loss in the process of model training . The weight term is further introduced into the loss function λi:

The authors propose three ways to provide weights λi Different schemes , The schematic diagram is described in Figure 2 Shown .

chart 2 The overall pipeline of the proposed method

1. Loss rejection

One way to handle large loss samples is by setting λi=0 To reject it . In multi class tasks with noise ,B.Han Et al. Proposed a method to gradually increase the rejection rate in the training process . The author also sets the function λi,

Because the model learns clean patterns at the initial stage , So in t=1 Do not reject any loss value . Use small batches instead of full batches in each iteration D′ To form a loss set . The author calls this method LL-R.

2. Loss correction ( temporary )

Another way to deal with a large loss sample is to correct it rather than reject it . In multi label settings , This can be easily achieved by switching the corresponding annotation from a negative value to a positive value .“ temporary ” The word means , It does not change the actual label , Only the loss calculated according to the modified label is used , Will function λi Defined as

The author named this method LL-Ct. The advantage of this method is , It increases the number of true positive tags in the tags that have never been observed .

3. Loss correction ( permanent )

Deal more aggressively with larger loss values by permanently correcting labels . Directly change the label from negative to positive , And use the modified tag during the next training . So , Define... For each case λi=1, And modify the label as follows :

The author named this method LL-Cp.

experiment

surface 2 Quantitative results of artificially created partial label data sets

surface 3 OpenImages V3 Quantitative results in the dataset

chart 3 Artificially generated COCO Qualitative results of some label datasets

chart 4 COCO Accuracy analysis of the proposed method on the dataset

chart 5 LL-Ct Yes COCO The hyperparametric effect of data sets

chart 6 Training with fewer images

surface 4 Pointing Game

Conclusion

In this paper , The author puts forward a loss modification scheme , This scheme rejects or corrects the large loss samples when training multi label classification models with partial label annotations . This comes from empirical observation , That is, the memory effect also occurs in noisy multi label classification scenarios .

Although it does not include heavy and complex components , But the author's scheme successfully prevents the multi label classification model from remembering false negative labels with noise , State of the art performance on a variety of partially labeled multi label datasets .

---------------------------------------------------------------------------------

 

CV The technical guide creates a computer vision technology exchange group and a free version of the knowledge planet , At present, the number of people on the planet has 700+, The number of topics reached 200+.

The knowledge planet will release some homework every day , It is used to guide people to learn something , You can continue to punch in and learn according to your homework .CV Every day in the technology group, the top conference papers published in recent days will be sent , You can choose the papers you are interested in to read , continued follow Latest technology , If you write an interpretation after reading it and submit it to us , You can also receive royalties . in addition , The technical group and my circle of friends will also publish various periodicals 、 Notice of solicitation of contributions for the meeting , If you need it, please scan your friends , And pay attention to .

Add groups and planets : Official account CV Technical guide , Get and edit wechat , Invite to join .

Welcome to the official account CV Technical guide , Focus on computer vision technology summary 、 The latest technology tracking 、 Interpretation of classic papers 、CV Recruitment information .

Welcome to contact me if you can write the following ( Get the contact information after following the official account ).

  1. Interpretation of the latest summit . For example, the most recent CVPR2022 The paper .
  2. Systematic overview of all directions 、 Development and evolution of main models 、 Innovative ideas, advantages and disadvantages of each model 、 Code analysis, etc . Such as the summary of target detection : A summary of target detection from traditional methods to deep learning , It mainly includes traditional method detection 、RCNN series 、YOLO series 、anchor-free series 、 Summary of small target detection methods 、 Summary of small sample target detection methods 、 Summary of object detection methods in video 、 Summary of loss function used in target detection . Support writing while learning .
  3. TVM Introduction to practice tutorial
  4. MNN Introduction to practice tutorial
  5. OpenVINO Introduction to practice tutorial
  6. libtorch Introduction to practice tutorial
  7. Oneflow Introduction to practice tutorial
  8. Detectron Introduction to practice tutorial
  9. caffe Source code reading
  10. Deep learning from introduction to mastery ( Start with convolutional neural networks )
  11. If you have something you want to write and it's not mentioned above , You can contact me . Statement : Paid , Please contact us for details .

Other articles of official account

Introduction to computer vision

Summary of common words in computer vision papers

YOLO Series carding ( Four ) About YOLO Deployment of

YOLO Series carding ( 3、 ... and )YOLOv5

YOLO Series carding ( Two )YOLOv4

YOLO Series carding ( One )YOLOv1-YOLOv3

CVPR2022 | Based on egocentric data OCR assessment

CVPR 2022 | Using contrast regularization method to deal with noise labels

CVPR2022 | Loss problem in weakly supervised multi label classification

CVPR2022 | iFS-RCNN: An incremental small sample instance divider

CVPR2022 | Time 3D: End to end joint monocular 3D object detection and tracking for autonomous driving

CVPR2022 | A ConvNet for the 2020s & How to design neural network Summary

CVPR2022 | PanopticDepth: A unified framework for depth aware panoramic segmentation

CVPR2022 | Reexamine pooling : Your feeling field is not ideal

CVPR2022 | Unknown target detection module STUD: Learn about unknown targets in the video

CVPR2022 | Ranking based siamese Visual tracking

Build from scratch Pytorch Model tutorial ( 6、 ... and ) Write the training process and reasoning process

Build from scratch Pytorch Model tutorial ( 5、 ... and ) Write the training process -- Some basic configuration

Build from scratch Pytorch Model tutorial ( Four ) Write the training process -- Argument parsing

Build from scratch Pytorch Model tutorial ( 3、 ... and ) build Transformer The Internet

Build from scratch Pytorch Model tutorial ( Two ) Build network

Build from scratch Pytorch Model tutorial ( One ) data fetch

Some personal thinking habits and thought summary about learning a new technology or field quickly

CVPR2022 | More related articles on the loss problem in weakly supervised multi label classification

  1. CVPR2020: Multi-path region mining for point cloud weakly supervised 3D semantic segmentation

    CVPR2020: Multi-path region mining for point cloud weakly supervised 3D semantic segmentation Multi-Path Region Mining for Weakly Supervised 3D Semantic Segmentation ...

  2. Valse2019 note —— Weakly supervised visual understanding

    Cheng Mingming ( Nankai University ): Adaptive visual perception for open environment ( The picture is from valse2019 Mr. Cheng Mingming ppt) Common technology of neural network for recognition and understanding General architecture of deep neural network -- VggNet(ICLR'15).ResN ...

  3. Change numerous for brief , Weak supervision is a new trend in the field of target orientation SOTA - Pseudo supervised target location method (PSOL) | CVPR 2020

    This paper proposes a method of pseudo supervised target location (PSOL) To solve the problem of weak supervision target location method , This method separates location and classification into two independent networks , Then use it on the training set Deep descriptor transformation(DDT ...

  4. scikit-learn General example 8 : Multi label classification

    This example simulates a multi label document classification problem . The dataset is randomly generated based on the following processing : Number of selected tags : Poisson (n~Poisson,n_labels) n Time , Select category C: polynomial (c~Multinomial,theta) Select text ...

  5. Mobile front end has to understand HTML5 head Head tag ( Middle part I )

    Meta label meta The label is HTML in head An auxiliary label on the head , It is located in HTML Of the document header <head> and <title> Between the marks , It provides information that is not visible to users . Although this part of the letter ...

  6. html( Commonly used labels , Label classification ), Page template , CSS(css Three ways to introduce ), Priority of the three methods of introduction

    HTML Markup language is a non programming language responsible for completing the structure of the page form : label : By <> The package begins with a letter , You can combine legal characters ( -| Numbers ), Special symbols that can be parsed by browsers , The label has a beginning and an end Instructions : By <> package ...

  7. front end HTML Label classification

    Three : 1. Block level label : Monopolize one line , Width can be set , Height . If you set the width and height , The current width and height . If the width and height are not set , The width is the width of the parent box , The height is filled in according to the content . 2. In line labels : Show... In one line , Can't set width , high ...

  8. Supervised text classification algorithm FastText

    The algorithm consists of facebook stay 2016 In open source , The typical application scenario is “ Supervised text classification problem ”.   Model The optimization objectives of the model are as follows :   among ,$<x_n,y_n>$ It's a training sample ,$y_n$ It's a training goal ,$ ...

  9. How to use softmax and sigmoid To do multi classification and multi label classification

    First , Let's talk about the difference between multi category classification and multi label classification Multi label classification : A sample can belong to multiple categories ( Or tags ), Different classes are related , For example, a text is divided into “ figure ” and “ Sports people ” Two labels . Obviously these two tags are not mutually exclusive , It's connected ...

  10. Introduction to the front html( Common labels and label classification )

    day47 Reference resources :https://www.cnblogs.com/liwenzhou/p/7988087.html Any tag has three properties :ID,class.style <!DOCTYPE ht ...

Random recommendation

  1. ASP.NET 5 Use TestServer Unit test

    If you were right ASP.NET WebAPI Unit test (HttpClient Initiate request , And it can be debugged WebAPI), It is generally used Owin The way , Specific reference :< Development Notes : use Owin Host Realization ...

  2. .split() Function usage

    split explain split(): Split string . Slice a string by specifying a separator , And returns a list of split strings (list) split grammar str.split('type',num)[n] The whole grammar means : With ...

  3. [Effective JavaScript note ] The first 64 strip : Using recursion for asynchronous loops

    Suppose you need to have a function like this , Receive one URL And try to download each file in turn until one file is successfully downloaded . If API It's synchronous , Using loops is easy to implement . function downloadOneSync(urls){ f ...

  4. const modification

    const int A() //const // ====>int A(const this) { // Point of view 1:const It's embellishment a, But pass the test , We found that ,b++ It doesn't compile // This explanation :co ...

  5. SQL Server Aggregate functions

    Aggregate functions return a single value after calculating a set of values . except count( Count the number of items ) Function , Other aggregate functions ignore null values in their expressions (null). All aggregate functions are deterministic functions . That is, at any time using the same set of input values to call the aggregate function after execution ...

  6. jQ Implementation of a rotating map

    as everyone knows , The rotation chart is widely used . We can see the rotation chart on many websites , For example, Taobao. . Jingdong these websites are very common . Now let's start our rotation tour : Build our skeleton : <!DOCTYPE html> < ...

  7. [PHP] PHP Server interface SAPI Structural body in

    SAPI: The same conventions are followed between the server abstraction layers , Here we call it SAPI Interface . For example, the implementation of command line program ,Apache Of mod_php Module implementation and fastcgi And so on 1. Structure : Use structure (Struc ...

  8. logrotate- Example log cutting

    logrotate yes linux The tools that come with the system , It can automatically truncate the log ( Or rotation ). Compress and delete old log files . 1) Sample configuration file # cat /wls/wls81/bin/weblogic/wls/app ...

  9. sql A field stores multiple... Of another table id Values separated by commas , Based on id Go to Chinese and splice with commas

    First, I will introduce the two functions used charindex( The expression to find 1, expression 2), The return value is the expression 1 In the expression 2 Subscript in , Return if not found 0.(sql The subscript of is from 1 At the beginning ), for example select charindex('s ...

  10. json Conversion tool class :json&lt;===&gt;list Or object

    public class JsonTools { /** * POJO turn JSON */ public static String createJsonString(Object object) { ...