当前位置:网站首页>Semi supervised mixpatch
Semi supervised mixpatch
2022-07-02 07:40:00 【Xiao Chen who wants money】
Self consistent regularization : There was too little tag data before , When the generalization ability of supervised learning is poor , People generally expand training data , For example, random translation of images , The zoom , rotate , Distortion , shear , Change the brightness , saturation , Add noise, etc . Data augmentation can produce countless modified new images , Expand the training data set . The idea of self consistent regularization is , Data augmentation of unlabeled data , The new data generated is input into the classifier , The prediction results should be self consistent . That is, the sample generated by the same data expansion , The prediction results of the model should be consistent . This rule is added to the loss function , There are the following forms ,

among x Is unmarked data ,Augment(x) Said to x Do random augmentation to generate new data , θ Is the model parameter ,y Is the result of the model prediction . Note that data augmentation is a random operation , Two Augment(x) The output is different . This L2 Loss item , Constrained machine learning model , All new images obtained by enlarging the same image , Make self consistent predictions .MixMatch Integrated self consistent regularization . Data augmentation uses random left-right flipping and clipping of images (Crop).
The second scheme is called Minimize entropy (Entropy Minimization)【5】. Many semi supervised learning methods are based on a consensus , That is, the classification boundary of the classifier should not pass through the high-density region of marginal distribution . The specific method is to force the classifier to make low entropy prediction for unlabeled data . The implementation method is to simply add a term to the loss function , To minimize the
Corresponding entropy .
MixMatch Use "sharpening" function , Minimize the entropy of unlabeled data . This part will be introduced later .
The third scheme is called traditional regularization (Traditional Regularization). In order to make the generalization ability of the model better , The general practice is to do L2 Regularization ,SGD Next L2 Regularization is equivalent to Weight Decay.MixMaxtch Used Adam Optimizer , And an article found that Adam and L2 There will be problems when regularization is used at the same time , therefore MixMatch Use a separate Weight decay.
A recently invented data augmentation method is called Mixup 【6】, Randomly sample two samples from the training data , Construct mixed samples and mixed labels , As new augmented data ,
among lambda It's a 0 To 1 A positive number between , Represents the mixing ratio of two samples .MixMatch take Mixup It is used in both marked data and unlabeled data .
mixmatch Specific steps :
- Use MixMatch Algorithm , To a Batch Tag data for x And a Batch Unlabeled data for u Data expansion , Get one... Respectively Batch Augmented data x' and K individual Batch Of u'.

among T, K, It's a super parameter. , Later on .MixMatch The data augmentation algorithm is as follows ,

Algorithm description :for Loop to a Batch The marked pictures and unlabeled pictures of are expanded . To mark pictures , Only one augmentation , The label remains unchanged , Write it down as p . For unmarked data , do K Sub random augmentation ( Superparameters in the article K=2), Input classifier , Get the average classification probability , Application temperature Sharpen Algorithm (T It's a temperature parameter , This algorithm will be introduced later ), Get unlabeled data “ guess ” label . At this time, the expanded tag data There is one Batch, Expanded unlabeled data Yes K individual Batch. take and Mix it up , Randomly rearrange the data set . Final MixMatch The output of the augmentation algorithm , Yes, it will And Did MixUp() One of the Batch Tag data for , as well as And Did MixUp() Of K individual Batch Unmarked augmented data .
. For the expanded tag data x , And unmarked augmented data u Calculate the loss items separately ,

边栏推荐
- 【Cutout】《Improved Regularization of Convolutional Neural Networks with Cutout》
- [introduction to information retrieval] Chapter 6 term weight and vector space model
- Play online games with mame32k
- Get the uppercase initials of Chinese Pinyin in PHP
- Drawing mechanism of view (I)
- 【Random Erasing】《Random Erasing Data Augmentation》
- ABM论文翻译
- 腾讯机试题
- How do vision transformer work?【论文解读】
- yolov3训练自己的数据集(MMDetection)
猜你喜欢

Deep learning classification Optimization Practice
![[introduction to information retrieval] Chapter 1 Boolean retrieval](/img/78/df4bcefd3307d7cdd25a9ee345f244.png)
[introduction to information retrieval] Chapter 1 Boolean retrieval
![[CVPR‘22 Oral2] TAN: Temporal Alignment Networks for Long-term Video](/img/bc/c54f1f12867dc22592cadd5a43df60.png)
[CVPR‘22 Oral2] TAN: Temporal Alignment Networks for Long-term Video
![[introduction to information retrieval] Chapter 7 scoring calculation in search system](/img/cc/a5437cd36956e4c239889114b783c4.png)
[introduction to information retrieval] Chapter 7 scoring calculation in search system

ModuleNotFoundError: No module named ‘pytest‘

Faster-ILOD、maskrcnn_benchmark安装过程及遇到问题

程序的内存模型

【Paper Reading】

【Programming】

Implementation of yolov5 single image detection based on onnxruntime
随机推荐
SSM laboratory equipment management
Classloader and parental delegation mechanism
基于onnxruntime的YOLOv5单张图片检测实现
Delete the contents under the specified folder in PHP
ModuleNotFoundError: No module named ‘pytest‘
[introduction to information retrieval] Chapter 7 scoring calculation in search system
【Mixup】《Mixup:Beyond Empirical Risk Minimization》
《Handwritten Mathematical Expression Recognition with Bidirectionally Trained Transformer》论文翻译
Using MATLAB to realize: Jacobi, Gauss Seidel iteration
Implementation of yolov5 single image detection based on pytorch
Translation of the paper "written mathematical expression recognition with bidirectionally trained transformer"
Implementation of yolov5 single image detection based on onnxruntime
One field in thinkphp5 corresponds to multiple fuzzy queries
Common machine learning related evaluation indicators
论文tips
Use Baidu network disk to upload data to the server
[paper introduction] r-drop: regulated dropout for neural networks
Jordan decomposition example of matrix
【Hide-and-Seek】《Hide-and-Seek: A Data Augmentation Technique for Weakly-Supervised Localization xxx》
使用MAME32K进行联机游戏