当前位置:网站首页>[medical segmentation] attention Unet
[medical segmentation] attention Unet
2022-07-07 16:36:00 【Coke Daniel】
summary
attention-unet The main contribution is to propose attention gate, It's plug and play , It can be directly integrated into unet In the model , The function is to suppress irrelevant areas in the input image , At the same time, highlight the remarkable characteristics of specific local areas , And it uses soft-attention Instead of hard-attention, therefore attention Weights can be learned online , And there's no need for extra label, Only a small amount of calculation is added .
details
structure
The core or unet Structure , But doing skip-connection When , In the middle there's a attention gate, After this ag after , Proceed again concat operation . because encoder There is relatively more fine-grained information in , But many are unnecessary redundancy ,ag Quite so encoder The current layer of is filtered , Suppress irrelevant information in the image , Highlight important local features .
attention gate
The two inputs are encoder Current layer of x l x^l xl and decoder The next layer of g g g, They passed by 1x1 Convolution of ( After making the number of channels consistent ), Then add elements by elements , And then pass by relu,1x1 Convolution of ( Reduce the number of channels to 1) and sigmoid Get the attention coefficient , And then there's another resample The module restores the size , Finally, the attention coefficient can be used to weight the feature map .
notes : Here is 3D Of ,2D If you understand , Just remove the last dimension .

Some explanations : Why add two inputs instead of directly based on encoder The current layer of gets the attention coefficient ?
Probably because , First, two characteristic graphs with the same size and number of channels are processed , The extracted features are different . Then this operation can strengthen the signal of the same region of interest , At the same time, different areas can also be used as auxiliary , The two copies add up to more auxiliary information . Or the further emphasis on the core information , At the same time, don't ignore those details .
Why resample Well ?
because x l And g x^l And g xl And g The size of is different , obviously g g g Its size is x l x^l xl Half of , They cannot add element by element , So we need to make the two dimensions consistent , Either large down sampling or small up sampling , The experiment shows that the effect of large down sampling is good . But what you get after this operation is the attention coefficient , Want to be with x l x^l xl The weight must be the same size , So we have to re sample .
attention
Attention The essence of a function can be described as a query (query) To a series of ( key key- value value) Mapping to
In the calculation attention It is mainly divided into three steps :
- The first step is to query And each key Calculate the similarity to get the weight , The common similarity function is a little product , Splicing , Perceptron, etc ;
- The second step is usually to use a softmax Function normalizes these weights ;
- Finally, the weight and the corresponding key value value Weighted sum to get the final attention.
hard-attention: Select one area of an image at a time as attention , set 1, Others are set to 0. He can't differentiate , Standard back propagation is not possible , Therefore, Monte Carlo sampling is needed to calculate the accuracy of each back-propagation stage . Considering that the accuracy depends on the completion of sampling , Therefore, its Other technology ( For example, reinforcement learning ).
soft-attention: Each pixel of the weighted image . Multiply the high correlation area by a larger weight , Low correlation areas are marked with smaller weights . The weight range is (0-1). He is differentiable , Back propagation can be carried out normally .
边栏推荐
- 偶然升职的内心独白
- 121. 买卖股票的最佳时机
- OpenGL personal notes
- 记录Servlet学习时的一次乱码
- Tidb cannot start after modifying the configuration file
- Good news! Kelan sundb database and Hongshu technology privacy data protection management software complete compatibility adaptation
- How can laravel get the public path
- hellogolang
- Lecturer solicitation order | Apache seatunnel (cultivating) meetup sharing guests are in hot Recruitment!
- [flower carving experience] 15 try to build the Arduino development environment of beetle esp32 C3
猜你喜欢

Record the migration process of a project

【MySql进阶】索引详解(一):索引数据页结构

统计学习方法——感知机

C语言进阶——函数指针

Statistical learning method -- perceptron

Tragedy caused by deleting the console statement

Three. JS series (2): API structure diagram-2

使用JSON.stringify()去实现深拷贝,要小心哦,可能有巨坑

Xcode Revoke certificate

Opencv configuration 2019vs
随机推荐
The team of East China Normal University proposed the systematic molecular implementation of convolutional neural network with DNA regulation circuit
torch.numel作用
What is the difference between IP address and physical address
网关Gateway的介绍与使用
The differences between exit, exit (0), exit (1), exit ('0 '), exit ('1'), die and return in PHP
laravel中将session由文件保存改为数据库保存
全网“追杀”钟薛高
AutoLISP series (1): function function 1
prometheus api删除某个指定job的所有数据
Sysom case analysis: where is the missing memory| Dragon lizard Technology
AutoLISP series (3): function function 3
Vs2019 configuration matrix library eigen
pycharm 终端部启用虚拟环境
01tire+链式前向星+dfs+贪心练习题.1
Multiplication in pytorch: mul (), multiply (), matmul (), mm (), MV (), dot ()
Opencv personal notes
PHP中exit,exit(0),exit(1),exit(‘0’),exit(‘1’),die,return的区别
Tidb cannot start after modifying the configuration file
JS modularization
AutoLISP series (2): function function 2