Published July 18, 2017 | Version Submitted
Discussion Paper Open

A Multi-Armed Bandit Approach for Online Expert Selection in Markov Decision Processes

Abstract

We formulate a multi-armed bandit (MAB) approach to choosing expert policies online in Markov decision processes (MDPs). Given a set of expert policies trained on a state and action space, the goal is to maximize the cumulative reward of our agent. The hope is to quickly find the best expert in our set. The MAB formulation allows us to quantify the performance of an algorithm in terms of the regret incurred from not choosing the best expert from the beginning. We first develop the theoretical framework for MABs in MDPs, and then present a basic regret decomposition identity. We then adapt the classical Upper Confidence Bounds algorithm to the problem of choosing experts in MDPs and prove that the expected regret grows at worst at a logarithmic rate. Lastly, we validate the theory on a small MDP.

Attached Files

Submitted - 1707.05714.pdf

Files

1707.05714.pdf

Files (562.9 kB)

Name Size Download all
md5:fad9369f8e768bf4149812c16642c2c4
562.9 kB Preview Download

Additional details

Identifiers

Eprint ID
110718
Resolver ID
CaltechAUTHORS:20210903-213649911

Related works

Dates

Created
2021-09-07
Created from EPrint's datestamp field
Updated
2023-06-02
Created from EPrint's last_modified field