Abstract | ||
---|---|---|
Model-free agent in reinforcement learning (RL) generally performs well but inefficient in training process with sparse data. A practical solution is to incorporate a model-based module in model-free agent. State transition can be learned to make desirable prediction of next state based on current state and action at each time step. This paper presents a new learning representation for variational RL by introducing the so-called transition uncertainty critic based on the variational encoder-decoder network where the uncertainty of structured state transition is encoded in a model-based agent. In particular, an action-gating mechanism is carried out to learn and decode the trajectory of actions and state transitions in latent variable space. The transition uncertainty maximizing exploration (TUME) is performed according to the entropy search by using the intrinsic reward based on the uncertainty measure corresponding to different states and actions. A dedicate latent variable model with a penalty using the bias of state-action value is developed. Experiments on Cart Pole and dialogue system show that the proposed TUME considerably performs better than the other exploration methods for reinforcement learning. |
Year | DOI | Venue |
---|---|---|
2020 | 10.23919/Eusipco47968.2020.9287440 | 28TH EUROPEAN SIGNAL PROCESSING CONFERENCE (EUSIPCO 2020) |
Keywords | DocType | ISSN |
machine learning, reward optimization | Conference | 2076-1465 |
Citations | PageRank | References |
0 | 0.34 | 0 |
Authors | ||
3 |
Name | Order | Citations | PageRank |
---|---|---|---|
Jen-Tzung Chien | 1 | 918 | 82.45 |
Wei-Lin Liao | 2 | 0 | 0.34 |
Issam El-Naqa | 3 | 528 | 36.31 |