Title
Network Representation Learning: From Preprocessing, Feature Extraction to Node Embedding
Abstract
AbstractNetwork representation learning (NRL) advances the conventional graph mining of social networks, knowledge graphs, and complex biomedical and physics information networks. Dozens of NRL algorithms have been reported in the literature. Most of them focus on learning node embeddings for homogeneous networks, but they differ in the specific encoding schemes and specific types of node semantics captured and used for learning node embedding. This article reviews the design principles and the different node embedding techniques for NRL over homogeneous networks. To facilitate the comparison of different node embedding algorithms, we introduce a unified reference framework to divide and generalize the node embedding learning process on a given network into preprocessing steps, node feature extraction steps, and node embedding model training for an NRL task such as link prediction and node clustering. With this unifying reference framework, we highlight the representative methods, models, and techniques used at different stages of the node embedding model learning process. This survey not only helps researchers and practitioners gain an in-depth understanding of different NRL techniques but also provides practical guidelines for designing and developing the next generation of NRL algorithms and systems.
Year
DOI
Venue
2023
10.1145/3491206
ACM Computing Surveys
Keywords
DocType
Volume
Network representation learning, data preprocessing, feature extraction, node embedding
Journal
55
Issue
ISSN
Citations 
2
0360-0300
0
PageRank 
References 
Authors
0.34
0
4
Name
Order
Citations
PageRank
Jing-Ya Zhou16416.35
Ling Liu25020344.35
Wenqi Wei34810.69
Jianxi Fan471860.15