Title
Graph Signal Sampling Via Reinforcement Learning
Abstract
We model the sampling and recovery of clustered graph signals as a reinforcement learning (RL) problem. The signal sampling is carried out by an agent which crawls over the graph and selects the most relevant graph nodes to sample. The goal of the agent is to select signal samples which allow for the most accurate recovery. The sample selection is formulated as a multi-armed bandit (MAB) problem, which lends naturally to learning efficient sampling strategies using the well-known gradient MAB algorithm. In a nutshell, the sampling strategy is represented as a probability distribution over the individual arms of the MAB and optimized using gradient ascent. Some illustrative numerical experiments indicate that the sampling strategies obtained from the gradient MAB algorithm outperform existing sampling methods.
Year
DOI
Venue
2018
10.1109/icassp.2019.8683181
2019 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING (ICASSP)
Keywords
Field
DocType
machine learning, reinforcement learning, multi-armed bandit, graph signal processing, total variation
Graph,Mathematical optimization,Gradient descent,Probability distribution,Sampling (statistics),Mathematics,Reinforcement learning
Journal
Volume
ISSN
Citations 
abs/1805.05827
1520-6149
0
PageRank 
References 
Authors
0.34
2
2
Name
Order
Citations
PageRank
Oleksii Abramenko100.34
Alexander Jung2133.46