Published December 14, 2021 | Version Submitted
Book Section - Chapter Open

Risk-Averse Stochastic Shortest Path Planning

  • 1. ROR icon California Institute of Technology

Abstract

We consider the stochastic shortest path planning problem in MDPs, i.e., the problem of designing policies that ensure reaching a goal state from a given initial state with minimum accrued cost. In order to account for rare but important realizations of the system, we consider a nested dynamic coherent risk total cost functional rather than the conventional risk-neutral total expected cost. Under some assumptions, we show that optimal, stationary, Markovian policies exist and can be found via a special Bellman's equation. We propose a computational technique based on difference convex programs (DCPs) to find the associated value functions and therefore the risk-averse policies. A rover navigation MDP is used to illustrate the proposed methodology with conditional-value-at-risk (CVaR) and entropic-value-at-risk (EVaR) coherent risk measures.

Additional Information

© 2021 IEEE.

Attached Files

Submitted - 2103.14727.pdf

Files

2103.14727.pdf

Files (400.7 kB)

Name Size Download all
md5:b69e84986e92fa5781eb9d59f38bfabd
400.7 kB Preview Download

Additional details

Identifiers

Eprint ID
109065
DOI
10.1109/CDC45484.2021.9683527
Resolver ID
CaltechAUTHORS:20210511-082139184

Dates

Created
2021-05-11
Created from EPrint's datestamp field
Updated
2022-02-16
Created from EPrint's last_modified field