当前位置:网站首页>Raki's notes on reading paper: neighborhood matching network for entity alignment
Raki's notes on reading paper: neighborhood matching network for entity alignment
2022-06-30 02:39:00 【Sleepy Raki】
Abstract & Introduction & Related Work
If you are not a knowledge map player , Please read the popular science of entity alignment first : Introduction to entity alignment
Research tasks
Entity alignedExisting methods and related work
- embedding-based methods
Facing the challenge
- Entity alignment is not easy , Because the knowledge map in real life is usually incomplete , And different knowledge maps usually have heterogeneous patterns . therefore , Equivalent entities from two knowledge maps may have different surface forms or different adjacency structures .
- The problem of different adjacency relations between equal entities is ubiquitous
- The difference of neighborhood size and topology brings great challenges to entity alignment methods

Innovative ideas
- The map sampling method extracts a neighborhood with identifiability for each entity
- Cross graph neighborhood matching module , Joint coding of neighborhood differences for a given entity pair
The experimental conclusion
sota
Our Approach
The two graphs are based on a pre aligned set of peer entities , The goal is to find the peer entity pairs in the two graphs 
Overview of NMN
- NMN The first use of GCNs Modeling the neighborhood topology information
- Neighborhood sampling is used to select neighbors with more information
- It uses a cross map matching module to capture the differences between neighbors

KG Structure Embedding
First use GCN To aggregate higher-order entity neighbor structure information , Use pre training word embedding to initialize GCN
Enter two graphs as a large graph into GCN Inside , Every GCN The layer takes a set of node features as input , And update the node representation to :
h i l h_i^{l} hil It's No L Node output characteristics of layer 
In order to control the accumulated noise , stay GCN A high-speed network is used between layers to effectively control noise propagation 
Neighborhood Sampling
One hop neighbor of an entity is the key to determine whether the entity should be aligned with other entities , But not all one hop neighbors contribute positively to entity alignment , Therefore, a down sampling process is used to select a neighbor with the largest amount of Central entity information
GCN The learned entities are embedded with rich context information ,for Neighborhood structure and entity semantics , The more context sensitive the central entity, the easier it is to sample

In essence, a discriminant neighborhood subgraph is constructed for each entity , This can achieve more accurate alignment through neighborhood matching 
Neighborhood Matching
The neighborhood subgraph generated by the sampling process determines which neighborhoods of the target entity should be considered in the later stage . let me put it another way ,NMN Handle pipeline The later stage of the will only operate on the neighbors in the subgraph . In the neighborhood matching stage , We want to be the corresponding KG For each candidate entity, find out which neighborhood of the entity is closely related to a neighborhood node in the subgraph of the target entity . This information is important for determining two entities ( From two KG) Whether it should be aligned is crucial 
Candidate selection
In order to reduce the computational overhead ,NMN First of all to E 1 E_1 E1 Medium e i e_i ei Sample an alignment candidate set C i = { c i 1 , c i 2 , . . . , c i t ∣ c i k ∈ E 2 } C_i = \{c_{i1}, c_{i2}, ..., c_{it} |c_{ik}∈E_2\} Ci={ ci1,ci2,...,cit∣cik∈E2}, And then calculate e i e_i ei And the subgraph similarity of these candidate sets . This is based on E 2 E_2 E2 In the embedded space e i e_i ei Closer entities are more likely to be associated with e i e_i ei Aligned observations . therefore , about E 2 E_2 E2 One of the entities in e j e_j ej, It is sampled as e i e_i ei The probability of a candidate can be calculated as 
Cross-graph neighborhood matching
p and q They are given E1 and E2 The neighbors of the nodes in the graph , Calculate an attention 
And then put them together 
For each target neighbor in the neighborhood subgraph , The attention mechanism in the matching module can accurately detect another KG Which neighbor in the subgraph of is most likely to match the target neighbor . Intuitively speaking , Matching vectors mp Captured the difference between the two closest neighbors . When the representations of two neighbors are similar , Matching vectors tend to be zero vectors , So their characterization remains similar . When the representations of neighbors are different , The matching vector will be amplified by propagation
Neighborhood Aggregation

My question is , Can you explain each parameter clearly ??? W g a t e W_{gate} Wgate What did you say ?
Experimental Setup


Experimental Results




Conclusion
NMN It's solved KG The ubiquitous neighborhood heterogeneity problem in . We achieve this goal by using a new sampling based method to select the most informative neighbor for each entity
NMN By considering topological structure and neighborhood similarity , Estimate the similarity of two entities at the same time . We have conducted extensive experiments on real-world data sets , And will NMN With the latest 12 Two embedding based methods are compared . Experimental results show that ,NMN Got the best 、 More robust performance , In different data sets and evaluation indicators, it is better than the competitive method
Remark
Can you make the symbol clear ?trainable Let's at least say the parameters 
Reading a little wordless paper
边栏推荐
- Jupyter notebook displays a collection of K-line graphs
- C语言 pivot_root的Invalid argument错误解决方案
- Time complexity analysis
- Global and Chinese market of ERP software for garment and textile industries 2022-2028: Research Report on technology, participants, trends, market size and share
- day33
- 堆排序
- Some configuration details about servlet initial development
- Detailed explanation of minimum stack
- 002 color classification
- 走进江苏作家诗人胭脂茉莉|世界读书日
猜你喜欢

FDA ESG规定:必须使用数字证书保证通信安全

快速排序

Unity3D UGUI强制刷新Layout(布局)组件

Shell Sort

Precautions for purchasing wildcard SSL certificate

2.< tag-动态规划和0-1背包问题>lt.416. 分割等和子集 + lt.1049. 最后一块石头的重量 II
![[Postgres] Postgres database migration](/img/45/7074aa766640160a3b6f00b109cb2f.png)
[Postgres] Postgres database migration

Time complexity analysis

Weekly recommended short video: why is the theory correct but can not get the expected results?

Can autojs code be encrypted? Yes, display of autojs encryption skills
随机推荐
Enlightenment from the revocation of Russian digital certificate by mainstream CA: upgrade the SSL certificate of state secret algorithm to help China's network security to be autonomous and controlla
1380. lucky numbers in matrices
Implementation of Sanzi chess with C language
FDA邮件安全解决方案
2.< tag-动态规划和0-1背包问题>lt.416. 分割等和子集 + lt.1049. 最后一块石头的重量 II
JS advanced -es6 syntax
Two methods of SSL certificate format conversion
CA数字证书包含哪些文件?如何查看SSL证书信息?
Dynamic SQL
Pytoch learning (II)
Cmake tutorial series -02- generating binaries using cmake code
CMake教程系列-03-依赖管理
Shenzhen CPDA Data Analyst Certification in July 2022
Network neuroscience——网络神经科学综述
Recheck on February 15, 2022
IBM websphere通道联通搭建和测试
SSL证书七大常见错误及解决方法
隐藏在科技教育中的steam元素
What files does a CA digital certificate contain? How to view SSL certificate information?
Creating exquisite skills in maker Education