Abstract | ||
---|---|---|
Transfer Learning (TL) has shown great potential to accelerate Reinforcement Learning (RL) by leveraging prior knowledge from past learned policies of relevant tasks. Existing TL approaches either explicitly computes the similarity between tasks or select appropriate source policies to provide guided explorations for the target task. However, how to directly optimize the target policy by alternatively utilizing knowledge from appropriate source policies without explicitly measuring the similarity is currently missing. In this paper, we propose a novel Policy Transfer Framework (PTF) by taking advantage of this idea. PTF learns when and which source policy is the best to reuse for the target policy and when to terminate it by modeling multi-policy transfer as the option learning problem. PTF can be easily combined with existing deep RL approaches. Experimental results show it significantly accelerates the learning process and outperforms state-of-the-art policy transfer methods in both discrete and continuous action spaces.
|
Year | DOI | Venue |
---|---|---|
2020 | 10.5555/3398761.3399072 | AAMAS '19: International Conference on Autonomous Agents and Multiagent Systems
Auckland
New Zealand
May, 2020 |
DocType | ISBN | Citations |
Conference | 978-1-4503-7518-4 | 0 |
PageRank | References | Authors |
0.34 | 0 | 10 |
Name | Order | Citations | PageRank |
---|---|---|---|
Tianpei Yang | 1 | 13 | 6.43 |
Jianye Hao | 2 | 189 | 55.78 |
Zhaopeng Meng | 3 | 79 | 15.68 |
Zongzhang Zhang | 4 | 36 | 10.71 |
Yujing Hu | 5 | 21 | 2.51 |
yingfeng chen | 6 | 69 | 13.64 |
Changjie Fan | 7 | 57 | 21.37 |
Weixun Wang | 8 | 1 | 5.75 |
Zhaodong Wang | 9 | 14 | 1.66 |
Jiajie Peng | 10 | 132 | 17.70 |