Abstract | ||
---|---|---|
We present a new reinforcement learning (RL) approach that enables an autonomous agent to solve decision making problems under constraints. Our assured reinforcement learning approach models the uncertain environment as a high-level, abstract Markov decision process (AMDP), and uses probabilistic model checking to establish AMDP policies that satisfy a set of constraints defined in probabilistic temporal logic. These formally verified abstract policies are then used to restrict the RL agent's exploration of the solution space so as to avoid constraint violations. We validate our RL approach by using it to develop autonomous agents for a flag-collection navigation task and an assisted-living planning problem. |
Year | DOI | Venue |
---|---|---|
2017 | 10.5220/0006156001050117 | ICAART: PROCEEDINGS OF THE 9TH INTERNATIONAL CONFERENCE ON AGENTS AND ARTIFICIAL INTELLIGENCE, VOL 2 |
Keywords | Field | DocType |
Reinforcement Learning,Safety Constraint Verification,Abstract Markov Decision Processes | Software engineering,Computer science,Theoretical computer science,Artificial intelligence,Machine learning,Reinforcement learning | Conference |
Citations | PageRank | References |
2 | 0.39 | 0 |
Authors | ||
4 |
Name | Order | Citations | PageRank |
---|---|---|---|
George Mason | 1 | 2 | 0.39 |
Radu Calinescu | 2 | 905 | 63.01 |
Daniel Kudenko | 3 | 678 | 84.54 |
Alec Banks | 4 | 2 | 0.39 |