当前位置:网站首页>1、 What is the difference between transfer learning and fine tuning?
1、 What is the difference between transfer learning and fine tuning?
2022-07-29 06:08:00 【My hair is messy】
One 、 The migration study
for instance , hypothesis boss Let you do some target detection , This data set is about optical fiber box inkjet character detection . The problem is , Data sets are few ( Only 1000 Data sheet ), There is much disturbing information in it , You find that starting from zero training yolo The effect is very poor , It's easy to over fit . What shall I do? , So you think of using Transfer Learning, Use what others have already trained Imagenet To make a model of .
Two 、 What are the methods of transfer learning
- hold Alexnet Inner convolution Take out the features of the output of the last layer , And use it directly SVM classification . This is a Transfer Learning, Because you used Alexnet Have learned in “ knowledge ”.
- Vggnet Take out the final output of convolution layer , Classify with Bayesian classifier . The thought is basically the same .
- Until you can put Alexnet、Vggnet The output of is taken out and combined , Design a classifier to classify by yourself . In this process, you not only use Alexnet Of “ knowledge ”, Also used. Vggnet Of “ knowledge ”.
- Last , You can also use it directly fine-tune This method , stay Alexnet On the basis of , Add the full connection layer again , Then train the network .
3、 ... and 、fine-tune Use policy
There are many factors that determine how to use transfer learning , This is the most important, only two : The size of the new dataset 、 And the similarity between the new data and the original data set . One thing to remember : The first few layers of the web learn about common features , The next few layers learn about the features related to categories . Here are four scenarios to use :
- The new data set is small and similar to the original data set . Because the new data set is smaller , If fine-tune It may be over fitting ; And because old and new data sets are similar , We expect them to have similar high-level characteristics , We can use the pre training network as a feature extractor , Train the linear classifier with the extracted features .
- The new data set is similar to the original data set . Because the new data set is big enough , Sure fine-tune The whole network .
- ** The new dataset is small and not similar to the original dataset .** The new data set is small , It's better not to fine-tune, Unlike the original data set , It's better not to use high-level features . At this time, we can use the features of the front layer to train SVM classifier .
- ** The new data set is large and not similar to the original data set .** Because the new data set is big enough , You can retrain . But in practice fine-tune The pre training model is useful . The new data set is big enough , Sure fine-tine The whole network .
Four 、 summary
Sum up ,Transfer Learning The concern is : What is? “ knowledge ” And how to better use what we got before “ knowledge ”. There can be many ways and means . and fine-tune Just one of the means .
边栏推荐
- 迁移学习——Robust Visual Domain Adaptation with Low-Rank Reconstruction
- 虚假新闻检测论文阅读(二):Semi-Supervised Learning and Graph Neural Networks for Fake News Detection
- 【Transformer】AdaViT: Adaptive Vision Transformers for Efficient Image Recognition
- GA-RPN:引导锚点的建议区域网络
- ASM piling: after learning ASM tree API, you don't have to be afraid of hook anymore
- ROS常用指令
- 二、深度学习数据增强方法汇总
- Beijing Baode & taocloud jointly build the road of information innovation
- 【pycharm】pycharm远程连接服务器
- Is flutter being quietly abandoned? On the future of flutter
猜你喜欢

这些你一定要知道的进程知识

pip安装后仍有解决ImportError: No module named XX

【pycharm】pycharm远程连接服务器

迁移学习——Robust Visual Domain Adaptation with Low-Rank Reconstruction

迁移学习——Transfer Joint Matching for Unsupervised Domain Adaptation

【ML】机器学习模型之PMML--概述

Detailed explanation of MySQL statistical function count

【语义分割】语义分割综述

Typical cases of xdfs & China Daily Online Collaborative Editing Platform
![[convolution kernel design] scaling up your kernels to 31x31: revising large kernel design in CNN](/img/71/f3fdf677cd5fddefffd4715e747297.png)
[convolution kernel design] scaling up your kernels to 31x31: revising large kernel design in CNN
随机推荐
pip安装后仍有解决ImportError: No module named XX
ReportingService WebService form authentication
fastText学习——文本分类
【Transformer】AdaViT: Adaptive Vision Transformers for Efficient Image Recognition
MarkDown简明语法手册
Isaccessible() method: use reflection techniques to improve your performance several times
迁移学习——Transitive Transfer Learning
Wechat built-in browser prohibits caching
第2周学习:卷积神经网络基础
一、常见损失函数的用法
虚假新闻检测论文阅读(四):A novel self-learning semi-supervised deep learning network to detect fake news on...
FFmpeg创作GIF表情包教程来了!赶紧说声多谢乌蝇哥?
个人学习网站
[pycharm] pycharm remote connection server
Are you sure you know the interaction problem of activity?
Transformer回顾+理解
clion+opencv+aruco+cmake配置
Android studio login registration - source code (connect to MySQL database)
研究生新生培训第二周:卷积神经网络基础
【语义分割】语义分割综述