Skip to main content
U.S. Department of Energy
Office of Scientific and Technical Information

Computing optimal policies for partially observable decision processes using compact representations

Conference ·
OSTI ID:430799
 [1]
  1. Univ. of British Columbia, Vancouver, BC (Canada)
Partially-observable Markov decision processes provide a general model for decision theoretic planning problems, allowing trade-offs between various courses of actions to be determined under conditions of uncertainty, and incorporating partial observations made by an agent. Dynamic programming algorithms based on the belief state of an agent can be used to construct optimal policies without explicit consideration of past history, but at high computational cost. In this paper, we discuss how structured representations of system dynamics can be incorporated in classic POMDP solution algorithms. We use Bayesian networks with structured conditional probability matrices to represent POMDPs, and use this model to structure the belief space for POMDP algorithms, allowing irrelevant distinctions to be ignored. Apart from speeding up optimal policy construction, we suggest that such representations can be exploited in the development of useful approximation methods.
OSTI ID:
430799
Report Number(s):
CONF-960876--; CNN: Grant OGP0121843; Grant OGPOO44121
Country of Publication:
United States
Language:
English

Similar Records

Distributionally Robust Partially Observable Markov Decision Process with Moment-Based Ambiguity
Journal Article · Sun Jan 31 19:00:00 EST 2021 · SIAM Journal on Optimization · OSTI ID:1785682

Optimal replacement policy for a partially observable Markov decision-process model
Thesis/Dissertation · Tue Dec 31 23:00:00 EST 1985 · OSTI ID:5151930

Bayesian sequential optimal experimental design for nonlinear models using policy gradient reinforcement learning
Journal Article · Thu Aug 10 20:00:00 EDT 2023 · Computer Methods in Applied Mechanics and Engineering · OSTI ID:1996019