Inference via low-dimensional couplings

We investigate the low-dimensional structure of deterministic transformations between random variables, i.e., transport maps between probability measures. In the context of statistics and machine learning, these transformations can be used to couple a tractable "reference" measure (e.g., a...

Full description

Bibliographic Details
Main Authors: Spantini, Alessio (Author), Bigoni, Daniele (Author), Marzouk, Youssef M (Author)
Other Authors: Massachusetts Institute of Technology. Department of Aeronautics and Astronautics (Contributor)
Format: Article
Language:English
Published: 2020-08-03T13:59:58Z.
Subjects:
Online Access:Get fulltext
LEADER 01869 am a22001813u 4500
001 126468
042 |a dc 
100 1 0 |a Spantini, Alessio  |e author 
100 1 0 |a Massachusetts Institute of Technology. Department of Aeronautics and Astronautics  |e contributor 
700 1 0 |a Bigoni, Daniele  |e author 
700 1 0 |a Marzouk, Youssef M  |e author 
245 0 0 |a Inference via low-dimensional couplings 
260 |c 2020-08-03T13:59:58Z. 
856 |z Get fulltext  |u https://hdl.handle.net/1721.1/126468 
520 |a We investigate the low-dimensional structure of deterministic transformations between random variables, i.e., transport maps between probability measures. In the context of statistics and machine learning, these transformations can be used to couple a tractable "reference" measure (e.g., a standard Gaussian) with a target measure of interest. Direct simulation from the desired measure can then be achieved by pushing forward reference samples through the map. Yet characterizing such a map-e.g., representing and evaluating it-grows challenging in high dimensions. The central contribution of this paper is to establish a link between the Markov properties of the target measure and the existence of low-dimensional couplings, induced by transport maps that are sparse and/or decomposable. Our analysis not only facilitates the construction of transformations in high-dimensional settings, but also suggests new inference methodologies for continuous non-Gaussian graphical models. For instance, in the context of nonlinear state-space models, we describe new variational algorithms for filtering, smoothing, and sequential parameter inference. These algorithms can be understood as the natural generalization-to the non-Gaussian case-of the square-root Rauch-Tung-Striebel Gaussian smoother. 
546 |a en 
655 7 |a Article 
773 |t Journal of machine learning research