Published July 2020 | Version Submitted
Book Section - Chapter Open

Risk-Averse Planning Under Uncertainty

Abstract

We consider the problem of designing policies for partially observable Markov decision processes (POMDPs) with dynamic coherent risk objectives. Synthesizing risk-averse optimal policies for POMDPs requires infinite memory and thus undecidable. To overcome this difficulty, we propose a method based on bounded policy iteration for designing stochastic but finite state (memory) controllers, which takes advantage of standard convex optimization methods. Given a memory budget and optimality criterion, the proposed method modifies the stochastic finite state controller leading to sub-optimal solutions with lower coherent risk.

Additional Information

© 2020 AACC.

Attached Files

Submitted - 1909.12499.pdf

Files

1909.12499.pdf

Files (1.8 MB)

Name Size Download all
md5:40b4b0150a782c9a3f66677dff933cc4
1.8 MB Preview Download

Additional details

Identifiers

Eprint ID
100583
Resolver ID
CaltechAUTHORS:20200109-092433424

Related works

Dates

Created
2020-01-09
Created from EPrint's datestamp field
Updated
2021-11-16
Created from EPrint's last_modified field

Caltech Custom Metadata

Caltech groups
Division of Biology and Biological Engineering (BBE)