当前位置:网站首页>Self supervised heterogeneous graph neural network with CO comparative learning
Self supervised heterogeneous graph neural network with CO comparative learning
2022-07-06 18:30:00 【kc7w91】
Self-supervised Heterogeneous Graph Neural Network with Co-contrastive Learning Paper reading
One 、 Key points
Heterograph : The basic idea of heterograph is quick pass
Dual Perspective : network schema view & meta-path structure view, Finally, the node representation obtained from the two perspectives is studied through comparison (contrastive learning) To merge
Self supervision : No label information is required , Directly use the data itself as supervision information , It can be divided into comparative learning and generative learning . This article is comparative learning , The core idea is to compare positive samples and negative samples in feature space , Feature representation of learning samples
Two 、 Two perspectives
network schema Pictured 1(b), Represents the relationship between nodes of different categories
meta-path Pictured 1(c), Multiple meta paths can be defined by - article - author / article - The theme - article /…
3、 ... and 、 Model definition
1. Preprocessing
The node characteristics of different categories x Project into the same space , The feature length alignment is recorded as h
2.network schema
For an article node , stay network schema From the perspective of the author 、 Topics learn from nodes in two categories embedding, Different categories have different effects on the current node , This degree value is similar to GAT The way ( Introduce attention mechanism ) Obtained by autonomous learning of the model . Each category contains multiple nodes , The importance of different nodes varies , The importance of each point in the class is also similar GAT The way to learn .
Intra class importance alpha(node-level attention):
among fai_m Indicates the node category , common m Kind of ;a For the parameters to be learned
adopt node-level attention Aggregate neighbor information :
notes : Not all neighbors , There are sampling operations
Get different categories embedding To merge
Category importance beta(type-level attention):
among W、b For the parameters to be learned
adopt type-level attention Aggregate different categories embedding:
3. meta-path
Through the Metapath, we get isomorphic graph , For each isomorphic graph, use GCN Get the preliminary characterization of each node h:
Fuse the node representations obtained under different meta paths ( It's similar GAT, semantic-level attention):
among W、b For the parameters to be learned ,beta_fai_n Is the importance of graphs obtained under different meta paths
According to different degrees of importance , obtain meta-path Nodes from the perspective embedding
4. mask
network schema Nodes in do not aggregate their own information (masked),meta-path The information of the transit node is not aggregated in , This will distinguish the nodes connected with the current article by category , Don't double count .
5.contrastive learning
Fuse nodes drawn from two perspectives embedding And then judge , The first use of MLP mapping :
Next, introduce contrastive learning
Comparative learning : There should be a high degree of similarity between small variants of something (positive samples), The similarity between essentially different things is low (negative samples). The positive and negative samples in this paper are defined as follows :
positive: By multiple meta-path Connected node pairs ( Emphasize the importance of edges )
negative: others
With i Node network schema From the perspective of loss Function as an example , Align the nodes according to meta-path The number of connected pieces is arranged in descending order , Set the threshold to divide positive and negative samples . After the division, there are network schema view Under the contrastive loss(meta-path view Empathy ):
among sim by cos function , Indicates the similarity between two vectors . the reason being that network schema From the perspective of loss function , So target embedding(gt) yes network schema Medium embedding; Positive and negative samples embedding come from meta-path view. The corresponding value of positive samples should be as large as possible , The corresponding value of negative samples should be as small as possible ,loss To get smaller .
Two perspectives loss equilibrium :
6. model extension
Hard to distinguish negative samples is very helpful to improve the performance of the comparative learning model , Therefore, a new negative sample generation strategy is introduced :GAN & Mixup
边栏推荐
- C language college laboratory reservation registration system
- The third season of Baidu online AI competition is coming in midsummer, looking for you who love AI!
- Grafana 9.0 is officially released! It's the strongest!
- DOM Brief
- Prophet模型的简介以及案例分析
- Coco2017 dataset usage (brief introduction)
- Jielizhi obtains the customized background information corresponding to the specified dial [chapter]
- 1700C - Helping the Nature
- Penetration test information collection - CDN bypass
- Epoll () whether it involves wait queue analysis
猜你喜欢
Take you through ancient Rome, the meta universe bus is coming # Invisible Cities
第三季百度网盘AI大赛盛夏来袭,寻找热爱AI的你!
從交互模型中蒸餾知識!中科大&美團提出VIRT,兼具雙塔模型的效率和交互模型的性能,在文本匹配上實現性能和效率的平衡!...
小程序在产业互联网中的作用
Compilation principle - top-down analysis and recursive descent analysis construction (notes)
Interesting - questions about undefined
Penetration test information collection - CDN bypass
Excellent open source fonts for programmers
【Swoole系列2.1】先把Swoole跑起来
Implementation of queue
随机推荐
CSRF vulnerability analysis
测试1234
测试123
简单易用的PDF转SVG程序
Redis的五种数据结构
Interesting - questions about undefined
文档编辑之markdown语法(typora)
DOM简要
传输层 拥塞控制-慢开始和拥塞避免 快重传 快恢复
Five data structures of redis
当保存参数使用结构体时必备的开发技巧方式
Jerry's updated equipment resource document [chapter]
Codeforces Round #803 (Div. 2)
STM32+MFRC522完成IC卡号读取、密码修改、数据读写
阿里云国际版ECS云服务器无法登录宝塔面板控制台
面向程序员的精品开源字体
Wchars, coding, standards and portability - wchars, encodings, standards and portability
使用cpolar建立一个商业网站(1)
287. 寻找重复数
Tree-LSTM的一些理解以及DGL代码实现