Abstract | ||
---|---|---|
We model the sampling and recovery of clustered graph signals as a reinforcement learning (RL) problem. The signal sampling is carried out by an agent which crawls over the graph and selects the most relevant graph nodes to sample. The goal of the agent is to select signal samples which allow for the most accurate recovery. The sample selection is formulated as a multi-armed bandit (MAB) problem, which lends naturally to learning efficient sampling strategies using the well-known gradient MAB algorithm. In a nutshell, the sampling strategy is represented as a probability distribution over the individual arms of the MAB and optimized using gradient ascent. Some illustrative numerical experiments indicate that the sampling strategies obtained from the gradient MAB algorithm outperform existing sampling methods. |
Year | DOI | Venue |
---|---|---|
2018 | 10.1109/icassp.2019.8683181 | 2019 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING (ICASSP) |
Keywords | Field | DocType |
machine learning, reinforcement learning, multi-armed bandit, graph signal processing, total variation | Graph,Mathematical optimization,Gradient descent,Probability distribution,Sampling (statistics),Mathematics,Reinforcement learning | Journal |
Volume | ISSN | Citations |
abs/1805.05827 | 1520-6149 | 0 |
PageRank | References | Authors |
0.34 | 2 | 2 |
Name | Order | Citations | PageRank |
---|---|---|---|
Oleksii Abramenko | 1 | 0 | 0.34 |
Alexander Jung | 2 | 13 | 3.46 |