当前位置:网站首页>Using transformer for object detection and semantic segmentation
Using transformer for object detection and semantic segmentation
2022-07-02 07:59:00 【MezereonXP】
Introduce
This time it's about Facebook AI An article from “End-to-End Object Detection with Transformers”
Just recently Transformer It's also popular , Here's how to use Transformer For object detection and semantic segmentation .
About Transformer, You can refer to my article article .
Let me briefly introduce Transformer, This is a model architecture for sequence to sequence modeling , It is widely used in natural language translation and other fields .Transformer Abandon the previous modeling of sequence RNN Form of network architecture , The attention mechanism is introduced , Achieved a good sequence modeling and transformation capabilities .
General structure and process

As shown in the figure above , It's mainly divided into two parts :
- Backbone: Mainly CNN, Used to extract advanced semantic features
- Encoder-Decoder: Make use of advanced semantic features and give target prediction
In more detail , The architecture is as follows

We give the process in sequence :
- Input picture , Shape is ( C 0 , H 0 , W 0 ) (C_0, H_0,W_0) (C0,H0,W0), among C 0 = 3 C_0 = 3 C0=3 Represents the number of channels
- CNN After feature extraction , obtain ( C , H , W ) (C,H,W) (C,H,W) The tensor of shape , among C = 2048 , H = H 0 32 , W = W 0 32 C=2048, H=\frac{H_0}{32}, W=\frac{W_0}{32} C=2048,H=32H0,W=32W0
- utilize 1x1 Convolution of , Reduce the size of the feature , obtain ( d , H , W ) (d, H, W) (d,H,W) Tensor , among d < < C d<< C d<<C
- Compress the tensor (squeeze), The shape becomes ( d , H W ) (d, HW) (d,HW)
- Got it d d d Vector sequence , Enter as a sequence into Encoder In
- Decoder Get the output vector sequence , adopt FFN(Feed Forward Network) Get the bounding box prediction and category prediction , among FFN It's simple 3 Layer perceptron , The bounding box prediction includes the normalized center coordinates and width and height .
The effect of target detection

As shown in the figure above , You can see DETR It's not a lot of calculations , however FPS It's not high , It's just in order .
So semantic segmentation ?
Here is the general framework of semantic segmentation , As shown in the figure below :

be aware , What's depicted in the picture , Bounding box embedding (Box Embedding) In essence decoder Output ( stay FFN Before ).
And then use a multi head attention mechanism , This mechanism is essentially right Q,K,V Do many linear transformations , In this ,K and V yes Encoder The input of ,Q yes decoder Output .
among M It's the number of heads for multi head attention .
after , Through a simple CNN, Get one Mask matrix , Used to generate the result of semantic segmentation .
Semantic segmentation results analysis

We can see that compared with PanopticFPN++ Come on , The improvement of effect is limited , especially AP It's not good , General performance .
Conclusion
The article will Transformer It is applied to the field of object detection and semantic segmentation , Good results have been achieved , But the performance is better than FastRCNN Architecture like approach , There is no obvious improvement , But it shows that this sequence model has good scalability . Using one architecture to solve multiple problems , The goal of a unified model is just around the corner .
边栏推荐
- 【Wing Loss】《Wing Loss for Robust Facial Landmark Localisation with Convolutional Neural Networks》
- 【Batch】learning notes
- Target detection for long tail distribution -- balanced group softmax
- TimeCLR: A self-supervised contrastive learning framework for univariate time series representation
- Jetson nano installation tensorflow stepping pit record (scipy1.4.1)
- What if the notebook computer cannot run the CMD command
- Business architecture diagram
- 【TCDCN】《Facial landmark detection by deep multi-task learning》
- [learning notes] matlab self compiled image convolution function
- [Sparse to Dense] Sparse to Dense: Depth Prediction from Sparse Depth samples and a Single Image
猜你喜欢

Semi supervised mixpatch

Thesis writing tip2

What if a new window always pops up when opening a folder on a laptop

【Mixed Pooling】《Mixed Pooling for Convolutional Neural Networks》

Look for we media materials from four aspects to ensure your creative inspiration

【Wing Loss】《Wing Loss for Robust Facial Landmark Localisation with Convolutional Neural Networks》

jetson nano安装tensorflow踩坑记录(scipy1.4.1)

【Mixup】《Mixup:Beyond Empirical Risk Minimization》
![[multimodal] clip model](/img/45/8501269190d922056ea0aad2e69fb7.png)
[multimodal] clip model

It's great to save 10000 pictures of girls
随机推荐
用MLP代替掉Self-Attention
【Mixup】《Mixup:Beyond Empirical Risk Minimization》
Open3d learning note 5 [rgbd fusion]
【TCDCN】《Facial landmark detection by deep multi-task learning》
Eklavya -- infer the parameters of functions in binary files using neural network
浅谈深度学习中的对抗样本及其生成方法
Open3d learning notes II [file reading and writing]
Mmdetection trains its own data set -- export coco format of cvat annotation file and related operations
It's great to save 10000 pictures of girls
业务架构图
【MobileNet V3】《Searching for MobileNetV3》
【Wing Loss】《Wing Loss for Robust Facial Landmark Localisation with Convolutional Neural Networks》
【双目视觉】双目矫正
Meta Learning 简述
包图画法注意规范
【Hide-and-Seek】《Hide-and-Seek: A Data Augmentation Technique for Weakly-Supervised Localization xxx》
关于原型图的深入理解
C # connect to MySQL database
使用C#语言来进行json串的接收
【Cutout】《Improved Regularization of Convolutional Neural Networks with Cutout》