Title
Reward function and initial values: better choices for accelerated goal-directed reinforcement learning
Abstract
An important issue in Reinforcement Learning (RL) is to accelerate or improve the learning process. In this paper, we study the influence of some RL parameters over the learning speed. Indeed, although RL convergence properties have been widely studied, no precise rules exist to correctly choose the reward function and initial Q-values. Our method helps the choice of these RL parameters within the context of reaching a goal in a minimal time. We develop a theoretical study and also provide experimental justifications for choosing on the one hand the reward function, and on the other hand particular initial Q-values based on a goal bias function.
Year
DOI
Venue
2006
10.1007/11840817_87
ICANN (1)
Keywords
Field
DocType
accelerated goal-directed reinforcement learning,goal bias function,experimental justification,important issue,hand particular initial q-values,initial q-values,initial value,rl parameter,rl convergence property,better choice,reinforcement learning,reward function,theoretical study
Convergence (routing),Computer science,Bellman equation,Initial value problem,Artificial intelligence,Estimation theory,Artificial neural network,Machine learning,Reinforcement learning
Conference
Volume
ISSN
ISBN
4131
0302-9743
3-540-38625-4
Citations 
PageRank 
References 
11
0.79
7
Authors
3
Name
Order
Citations
PageRank
Laëtitia Matignon1889.43
Guillaume J. Laurent29712.60
Nadine Le Fort-Piat37710.09