Published November 13, 2018 | Version Published + Submitted
Book Section - Chapter Open

A General Method for Amortizing Variational Filtering

Abstract

We introduce the variational filtering EM algorithm, a simple, general-purpose method for performing variational inference in dynamical latent variable models using information from only past and present variables, i.e. filtering. The algorithm is derived from the variational objective in the filtering setting and consists of an optimization procedure at each time step. By performing each inference optimization procedure with an iterative amortized inference model, we obtain a computationally efficient implementation of the algorithm, which we call amortized variational filtering. We present experiments demonstrating that this general-purpose method improves performance across several deep dynamical latent variable models.

Additional Information

© 2018 Neural Information Processing Systems Foundation, Inc. We would like to thank Matteo Ruggero Ronchi for helpful discussions. This work was supported by the following grants: JPL PDF 1584398, NSF 1564330, and NSF 1637598.

Attached Files

Published - 8011-a-general-method-for-amortizing-variational-filtering.pdf

Submitted - 1811.05090.pdf

Files

1811.05090.pdf

Files (1.7 MB)

Name Size Download all
md5:dfc8b196521e6ff3ff226b7acadacb7d
838.5 kB Preview Download
md5:d726c38ec9450afc43c5c3d758e45ca0
842.1 kB Preview Download

Additional details

Identifiers

Eprint ID
92660
Resolver ID
CaltechAUTHORS:20190205-101451102

Related works

Funding

JPL President and Director's Fund
1584398
NSF
IIS-1564330
NSF
CCF-1637598

Dates

Created
2019-02-05
Created from EPrint's datestamp field
Updated
2023-06-02
Created from EPrint's last_modified field

Caltech Custom Metadata

Series Name
Advances in Neural Information Processing Systems
Series Volume or Issue Number
31