当前位置:网站首页>CaEGCN: Cross-Attention Fusion based Enhanced Graph Convolutional Network for Clustering 2021
CaEGCN: Cross-Attention Fusion based Enhanced Graph Convolutional Network for Clustering 2021
2022-07-27 23:43:00 【Oriental shrimp】
problem : Existing deep clustering methods often ignore the relationship between data .
This paper proposes a deep clustering framework based on cross attention —— Enhanced graph convolution network based on cross attention fusion (CaEGCN) , The network consists of four main modules : Cross attention fusion module , Innovatively self coding the content related to individual data (CAE) And a graphic convolutional self coding module related to the relationship between layer by layer data (GAE) Connect ; And self-monitoring model , The model highlights the identification information of clustering tasks . The cross attention fusion module integrates two heterogeneous representations ,CAE The module complements GAE Content information of the module , Avoided GCN Over smoothing problem . stay GAE Module , Two new loss functions are proposed , Reconstruct the content of data and the relationship between data respectively . Last , Self monitoring module constraints CAE and GAE The distribution represented by the middle layer is consistent .
1) A deep clustering framework based on end-to-end cross attention fusion is proposed , Among them, the cross attention fusion module creatively connects the automatic picture coding module and the automatic content coding module ;
2) A cross attention fusion model is proposed , Pay attention to the weight distribution of the fused heterogeneous representation ;
3) In the graph convolution automatic coding module , Propose to reconstruct the content and relationship of data at the same time , Effectively enhanced CaEGCN Clustering performance ;
CROSS-ATTENTION FUSION BASED ENHANCED GRAPH CONVOLUTIONAL NETWORK
The overall framework is as shown in the figure :
It consists of four main modules : Self encoder module for extracting content information ; be based on GCN Self encoder module , Used to exploit the relationship between data ; Cross attention module , Used to connect the above two modules , The multi-level adaptive fusion strategy supplements effective content information as much as possible in the transmission process ; And a self-monitoring module for constraining the distribution consistency of the middle layer representation .
A. Constructing the Graph
Construct raw data X The method of drawing is similar to SDCN equally ,KNN Method .
B. Content Auto-encoder Module (CAE)
Adopt the full connection layer of the foundation , Loss adopts reconstruction loss , restructure X.
C. Cross-Attention Fusion Module
Cross attention fusion mechanism is used to integrate CAE The content information learned and GAE Integrate the learned data relationships .
The cross attention fusion mechanism is defined as :
R = F a t t ( Q , K , V ) (1) R=F_{att}(Q,K,V) \tag{1} R=Fatt(Q,K,V)(1)
Q:query,K:key,V:value
The initial input of the cross attention fusion module Y Defined as :
Y = γ Z l + ( 1 − r ) H l (2) Y=\gamma Z_l + (1-r)H_l \tag{2} Y=γZl+(1−r)Hl(2)
Z l , H l Z_l,H_l Zl,Hl Respectively GAE And CAE The first l Layer output . γ \gamma γ It's a trade-off parameter , In this experiment, it is set to 0.5.
Attention mechanism calculation : First calculate the fusion query And fusion key The similarity between :
s a b = q a ∗ k b (3) s_{ab} = q_a * k_b \tag{3} sab=qa∗kb(3)
Then proceed on the above basis softmax Normalized acquisition a a b a_{ab} aab:
a a b = s o f t m a x ( s a b ) = e x p ( s a b ) ∑ a = 0 D a t t e x p ( s a b ) (4) a_{ab} = softmax(s_{ab}) = \frac{exp(s_{ab})}{\sum^{D_{att}}_{a=0}exp(s_{ab})}\tag{4} aab=softmax(sab)=∑a=0Dattexp(sab)exp(sab)(4)
Cross attention fusion mechanism R The final output is :
r a = ∑ b = 0 N a a b v b (5) r_a = \sum^N_{b=0}a_{ab}v_b \tag{5} ra=b=0∑Naabvb(5)
To further perceive different aspects of data , The multi head mechanism is introduced :
R m = F a t t ( Q m , K m , V m ) , m = 1 , 2 , 3 , . . . , M . (6) R^m=F_{att}(Q_m,K_m,V_m) \tag{6},m=1,2,3,...,M. Rm=Fatt(Qm,Km,Vm),m=1,2,3,...,M.(6)
among Q m = W m q Q Q_m = W_m^qQ Qm=WmqQ,K,V Empathy .
R = W ⋅ C o n c a t ( R 1 , . . . , R M ) (7) R = W\cdot Concat(R_1,...,R_M)\tag{7} R=W⋅Concat(R1,...,RM)(7)
among Concat(·) Represents the matrix concatenation operation . This is the so-called multi head mechanism and cross attention fusion module .
D. Graph Convolutional Auto-Encoder Module (GAE)
The output obtained from the previous cross attention fusion module is represented R As GAE The input of , Perform spectral convolution .
Z l = G A E ( R l − 1 , A ) = a l ( D ^ − 1 / 2 A ^ D ^ − 1 / 2 R l − 1 U L ) (8) Z_l = GAE(R_{l-1},A)=a_l(\hat{D}^{-1/2}\hat{A}\hat{D}^{-1/2}R_{l-1}U_L )\tag{8} Zl=GAE(Rl−1,A)=al(D^−1/2A^D^−1/2Rl−1UL)(8)
Loss function reconstruction A ~ = S i g m o d i ( Z L T Z L ) \tilde{A} = Sigmodi(Z_L^TZ_L) A~=Sigmodi(ZLTZL) and X ~ = Z L \tilde{X}=Z_L X~=ZL.
E. Self-Supervised Module
In the previous study, it is difficult to judge whether it is the optimal representation of clustering , We need to give an optimization goal of clustering .
Here and such SDCN Like the previous methods Student’s t-distribution As a real label , The second normalization is used as the target label . Yes CAE and GAE At the same time, optimize .
边栏推荐
- Error:svn: E155010: ‘/Users/.../Desktop/wrokspace/xxx‘ is scheduled for addition, but is missing
- Find and leverage xxE – XML external entity injection
- TFRecord的Shuffle、划分和读取
- Tita 的OKR系统与其他同类型产品,或者是共享文档等相比,有什么优势?
- 编辑复制粘贴判定问题(bug?),所见即所得显示符号问题反馈。
- Blood spitting finishing nanny level series tutorial - playing Fiddler bag capturing tutorial (5) - detailed explanation of fiddler monitoring panel
- 【JS 逆向百例】某公共资源交易网,公告 URL 参数逆向分析
- How do I shut down oracle?
- 西门子PLC能否实时无线采集多处从站模拟量数据?
- QT with OpenGL(Shadow Mapping)(平行光篇)
猜你喜欢

Process and planned task management

习题 --- BFS

29.学习Highcharts 使用百分比的堆叠柱形图

【GNN报告】加拿大蒙特利尔唐建:Geometric Deep Learning For Drug Discovery

Preliminary understanding of Panda3D audio and advanced interactive components
![[elm classification] classification of UCI data sets based on nuclear limit learning machine and limit learning machine, with matlab code](/img/bd/fe105f0af3a24ad6f220fa35bb5e8d.png)
[elm classification] classification of UCI data sets based on nuclear limit learning machine and limit learning machine, with matlab code

华为鸿蒙 3 正式发布,这个安全功能解决了一大痛点

Interviewer: let's talk about the specific process of network data transmission

一位软件投资者的独白:我为什么不追逐快速增长的公司
Blood spitting finishing nanny level series tutorial - playing Fiddler bag capturing tutorial (5) - detailed explanation of fiddler monitoring panel
随机推荐
Redis distributed lock
Sudden, wechat important notice
Lua basic grammar learning
What are the advantages of Tita's OKR system over other similar products or shared documents?
编辑复制粘贴判定问题(bug?),所见即所得显示符号问题反馈。
Socket interaction process of four waves
我年薪100万,全身上下没有超过100块的衣服:存钱,是最顶级的自律
Huawei Hongmeng 3 was officially released, and this security feature has solved a major pain point
记录一下使用R语言中关于formatC的错误
proteus仿真arduino中调用DHT11/22温湿度传感器
华为鸿蒙 3 正式发布,这个安全功能解决了一大痛点
小程序容器技术超有料,可以让移动研发效率大幅提升
Record the errors about formatc in R language
用3dmax做折扇的思路方法与步骤
29.学习Highcharts 使用百分比的堆叠柱形图
What are the methods of process synchronization?
请求合并哪家强——接口请求合并的3种技巧,性能直接爆表
The total investment is 60billion! Foxconn semiconductor high-end package test project officially settled in Qingdao
怎么使用xshell免费版
Redis 哈希Hash底层数据结构