Abstract | ||
---|---|---|
Partially Observable Markov Decision Process (POMDP) is a popular framework for planning under uncertainty in partially observable domains. Yet, the POMDP model is risk-neutral in that it assumes that the agent is maximizing the expected reward of its actions. In contrast, in domains like financial planning, it is often required that the agent decisions are risk-sensitive (maximize the utility of agent actions, for non-linear utility functions). Unfortunately, existing POMDP solvers cannot solve such planning problems exactly. By considering piecewise linear approximations of utility functions, this paper addresses this shortcoming in three contributions: (i) It defines the Risk-Sensitive POMDP model; (ii) It derives the fundamental properties of the underlying value functions and provides a functional value iteration technique to compute them exactly and (c) It proposes an efficient procedure to determine the dominated value functions, to speed up the algorithm. Our experiments show that the proposed approach is feasible and applicable to realistic financial planning domains. |
Year | DOI | Venue |
---|---|---|
2010 | 10.5555/1838206.1838384 | AAMAS |
Keywords | Field | DocType |
observable environment,risk-sensitive planning,realistic financial planning domain,risk-sensitive pomdp model,underlying value function,pomdp model,financial planning,functional value iteration technique,planning problem,agent decision,agent action,non-linear utility function,value function,utility theory,piecewise linear,value iteration | Mathematical optimization,Observable,Computer science,Partially observable Markov decision process,Piecewise linear approximation,Markov decision process,Artificial intelligence,Financial plan,Machine learning,Utility theory,Speedup | Conference |
Citations | PageRank | References |
0 | 0.34 | 11 |
Authors | ||
2 |
Name | Order | Citations | PageRank |
---|---|---|---|
Janusz Marecki | 1 | 685 | 49.06 |
Pradeep Varakantham | 2 | 648 | 63.05 |