Abstract | ||
---|---|---|
Graph representation learning or graph embedding is a classical topic in data mining. Current embedding methods are mostly non-parametric, where all the embedding points are unconstrained free points in the target space. These approaches suffer from limited scalability and an over-flexible representation. In this paper, we propose a parametric graph embedding by fusing graph topology information and node content information. The embedding points are obtained through a highly flexible non-linear transformation from node content features to the target space. This transformation is learned using the contrastive loss function of the siamese network to preserve node adjacency in the input graph. On several benchmark network datasets, the proposed GraPASA method shows a significant margin over state-of-the-art techniques on benchmark graph representation tasks. |
Year | DOI | Venue |
---|---|---|
2020 | 10.1016/j.ins.2019.10.027 | Information Sciences |
Keywords | Field | DocType |
Network embedding,Inductive representation learning,Siamese network,Information fusion | Adjacency list,Architecture,Embedding,Graph embedding,Theoretical computer science,Parametric statistics,Artificial intelligence,Topological graph theory,Graph (abstract data type),Machine learning,Mathematics,Scalability | Journal |
Volume | ISSN | Citations |
512 | 0020-0255 | 0 |
PageRank | References | Authors |
0.34 | 0 | 5 |
Name | Order | Citations | PageRank |
---|---|---|---|
Yujun Chen | 1 | 6 | 1.81 |
Ke Sun | 2 | 0 | 1.69 |
Juhua Pu | 3 | 50 | 11.90 |
Zhang Xiong | 4 | 1069 | 102.45 |
Xiangliang Zhang | 5 | 728 | 87.74 |