Title
Pruned Graph Scattering Transforms
Abstract
Graph convolutional networks (GCNs) have achieved remarkable performance in a variety of network science learning tasks. However, theoretical analysis of such approaches is still at its infancy. Graph scattering transforms (GSTs) are non-trainable deep GCN models that are amenable to generalization and stability analyses. The present work addresses some limitations of GSTs by introducing a novel so-termed pruned (p)GST approach. The resultant pruning algorithm is guided by a graph-spectrum-inspired criterion, and retains informative scattering features on-the-fly while bypassing the exponential complexity associated with GSTs. It is further established that pGSTs are stable to perturbations of the input graph signals with bounded energy. Experiments showcase that i) pGST performs comparably to the baseline GST that uses all scattering features, while achieving significant computational savings; ii) pGST achieves comparable performance to state-of-the-art GCNs; and iii) Graph data from various domains lead to different scattering patterns, suggesting domain-adaptive pGST network architectures.
Year
Venue
Keywords
2020
ICLR
Graph scattering transforms, pruning, graph convolutional networks, stability, deep learning
DocType
Citations 
PageRank 
Conference
0
0.34
References 
Authors
26
3
Name
Order
Citations
PageRank
Vassilis N. Ioannidis1147.34
Siheng Chen200.34
Georgios B. Giannakis300.34