Abstract | ||
---|---|---|
Many two layer robot architectures have been proposed and implemented. While justification for the design can be well argued, how does one know it is really a good idea? In this paper, one describes a two layer architecture (reinforcement learning in the bottom layer and POMDP planning at the top) for a simulated robot and summarize a set of three experiments in which one evaluated the design. To address the many difficulties of evaluating robot architectures, one advocates an experimental approach in which design criteria are elucidated and then form the basis for the evaluation experiments. In our case, one tests the implementation for its reliability and generalization (our design criteria) by comparing our architecture to one in which a key component is substituted; in these experiments, one demonstrates significant performance gains on the design criteria for our architecture. |
Year | DOI | Venue |
---|---|---|
2000 | 10.1080/095281300409847 | JOURNAL OF EXPERIMENTAL & THEORETICAL ARTIFICIAL INTELLIGENCE |
Keywords | Field | DocType |
robot navigation,reinforcement learning,partially observable markov decision process | Robot learning,Architecture,Partially observable Markov decision process,Computer science,Robustness (computer science),Artificial intelligence,Robot,Machine learning,Robot architecture,Reinforcement learning | Journal |
Volume | Issue | ISSN |
12 | 2 | 0952-813X |
Citations | PageRank | References |
0 | 0.34 | 17 |
Authors | ||
2 |
Name | Order | Citations | PageRank |
---|---|---|---|
Larry D. Pyeatt | 1 | 66 | 11.11 |
Adele E. Howe | 2 | 561 | 65.70 |