Li, Kun and Burdick, Joel W. (2017) A Function Approximation Method for Model-based High-Dimensional Inverse Reinforcement Learning. . (Unpublished) https://resolver.caltech.edu/CaltechAUTHORS:20190410-120630278
![]() |
PDF
- Submitted Version
See Usage Policy. 262kB |
Use this Persistent URL to link to this item: https://resolver.caltech.edu/CaltechAUTHORS:20190410-120630278
Abstract
This works handles the inverse reinforcement learning problem in high-dimensional state spaces, which relies on an efficient solution of model-based high-dimensional reinforcement learning problems. To solve the computationally expensive reinforcement learning problems, we propose a function approximation method to ensure that the Bellman Optimality Equation always holds, and then estimate a function based on the observed human actions for inverse reinforcement learning problems. The time complexity of the proposed method is linearly proportional to the cardinality of the action set, thus it can handle high-dimensional even continuous state spaces efficiently. We test the proposed method in a simulated environment to show its accuracy, and three clinical tasks to show how it can be used to evaluate a doctor's proficiency.
Item Type: | Report or Paper (Discussion Paper) | ||||||
---|---|---|---|---|---|---|---|
Related URLs: |
| ||||||
Additional Information: | This work was supported by the National Institutes of Health, NIBIB. | ||||||
Funders: |
| ||||||
Record Number: | CaltechAUTHORS:20190410-120630278 | ||||||
Persistent URL: | https://resolver.caltech.edu/CaltechAUTHORS:20190410-120630278 | ||||||
Usage Policy: | No commercial reproduction, distribution, display or performance rights in this work are provided. | ||||||
ID Code: | 94632 | ||||||
Collection: | CaltechAUTHORS | ||||||
Deposited By: | George Porter | ||||||
Deposited On: | 11 Apr 2019 14:35 | ||||||
Last Modified: | 03 Oct 2019 21:05 |
Repository Staff Only: item control page