当前位置:网站首页>On the back door of deep learning model
On the back door of deep learning model
2022-07-02 07:59:00 【MezereonXP】
About deep learning safety , It can be roughly divided into two pieces : Counter samples (Adversarial Example) as well as back door (Backdoor)
For the confrontation sample, please check my previous article ---- Against sample attacks
This time we mainly focus on the backdoor attack in deep learning . The back door , That is a hidden , A channel that is not easily found . In some special cases , This channel will be exposed .
Then in the deep learning , What about the back door ? Here I might as well take the image classification task as an example , We have a picture of a dog in our hand , By classifier , With 99% The degree of confidence (confidence) Classified as a dog . If I add a pattern to this image ( Like a small red circle ), By classifier , With 80% The confidence is classified as cat .
Then we will call this special pattern trigger (Trigger), This classifier is called a classifier with a back door .
Generally speaking , Backdoor attack is composed of these two parts , That is, triggers and models with backdoors
The trigger will trigger the classifier , Make it erroneously classified into the specified category ( Of course, it can also be unspecified , Just make it wrong , Generally speaking, we are talking about designated categories , If other , Special instructions will be given ).
We have already introduced the backdoor attack , Here we mainly focus on several issues :
- How to get the model with back door and the corresponding trigger
- How to make a hidden back door
- How to detect the back door in the model
This time we will focus on the first and second questions , How to get the model with back door and the corresponding trigger .
Generally speaking , We will operate the training data , The backdoor attack is realized by modifying the training data , Such means , be called Based on poisoning (poisoning-based) The back door of .
Here it is with Poison attack Make a difference , The purpose of poisoning attack is to poison data , Reduce the generalization ability of the model (Reduce model generalization), The purpose of backdoor attack is to invalidate the input of the model with trigger , The input without trigger behaves normally .
BadNet
First, let's introduce the most classic attacks , from Gu Et al , It's very simple , Is to randomly select samples from the training data set , Add trigger , And modify their real tags , Then put it back , Build a toxic data set .
Gu et al. Badnets: Evaluating backdooring attacks on deep neural networks
This kind of method often needs to modify the label , So is there a way not to modify the label ?
Clean Label
clean label The way is not to modify the label , As shown in the figure below , Just add a special transformation , At the same time, the trigger of this method is relatively hidden ( The trigger is the corresponding transformation ).
Barni et al. A new Backdoor Attack in CNNs by training set corruption without label poisoning
More subtle triggers Hiding Triggers
Liao Et al. Proposed a way to generate triggers , This method will restrict the impact of the trigger on the original image , As shown in the figure below :
Basically, the human eye can't distinguish whether a picture has a trigger .
Liao et al. Backdoor Embedding in Convolutional Neural Network Models via Invisible Perturbation
Dynamic triggers Dynamic Backdoor
This method consists of Salem And others raised it , Dynamically determine the location and style of triggers through a network , Enhanced the effect of the attack .
Salem et al. Dynamic Backdoor Attacks Against Machine Learning Models
边栏推荐
- Look for we media materials from four aspects to ensure your creative inspiration
- 解决jetson nano安装onnx错误(ERROR: Failed building wheel for onnx)总结
- 【Cutout】《Improved Regularization of Convolutional Neural Networks with Cutout》
- JVM instructions
- 将恶意软件嵌入到神经网络中
- w10升级至W11系统,黑屏但鼠标与桌面快捷方式能用,如何解决
- SQL server如何卸载干净
- How to clean up logs on notebook computers to improve the response speed of web pages
- Semi supervised mixpatch
- CONDA common commands
猜你喜欢
【MnasNet】《MnasNet:Platform-Aware Neural Architecture Search for Mobile》
【MnasNet】《MnasNet:Platform-Aware Neural Architecture Search for Mobile》
Vscode下中文乱码问题
Timeout docking video generation
CVPR19-Deep Stacked Hierarchical Multi-patch Network for Image Deblurring论文复现
利用Transformer来进行目标检测和语义分割
Real world anti sample attack against semantic segmentation
Graph Pooling 简析
Semi supervised mixpatch
【Programming】
随机推荐
In the era of short video, how to ensure that works are more popular?
Semi supervised mixpatch
Gensim如何冻结某些词向量进行增量训练
One book 1078: sum of fractional sequences
[CVPR‘22 Oral2] TAN: Temporal Alignment Networks for Long-term Video
【Cutout】《Improved Regularization of Convolutional Neural Networks with Cutout》
jetson nano安装tensorflow踩坑记录(scipy1.4.1)
应对长尾分布的目标检测 -- Balanced Group Softmax
Command line is too long
用MLP代替掉Self-Attention
Look for we media materials from four aspects to ensure your creative inspiration
用全连接层替代掉卷积 -- RepMLP
Mmdetection trains its own data set -- export coco format of cvat annotation file and related operations
Programmers can only be 35? The 74 year old programmer in the United States has been programming for 57 years and has not retired
Ppt skills
E-R画图明确内容
【双目视觉】双目矫正
Meta Learning 简述
Business architecture diagram
用于类别增量学习的动态可扩展表征 -- DER