Recursive compositional models: representation, learning, and inference

Recursive compositional models (RCMs) are hierarchical models which enable us to represent the shape/geometry and visual appearance of objects and images at different scales. The key design principle is recursive compositionality. Objects are represented by RCMs in a hierarchical form where complex...

Full description

Bibliographic Details
Main Authors: Yuille, Alan (Author), Zhu, Long (Contributor)
Other Authors: Massachusetts Institute of Technology. Computer Science and Artificial Intelligence Laboratory (Contributor)
Format: Article
Language:English
Published: Institute of Electrical and Electronics Engineers, 2010-11-12T18:39:10Z.
Subjects:
Online Access:Get fulltext
Description
Summary:Recursive compositional models (RCMs) are hierarchical models which enable us to represent the shape/geometry and visual appearance of objects and images at different scales. The key design principle is recursive compositionality. Objects are represented by RCMs in a hierarchical form where complex structures are composed of more elementary structures. Formally, they are represented by probability distributions defined over graphs with variable topology. Learning techniques are used to learn these models from a limited number of examples of the object by exploiting the recursive structure (some of our papers use supervised learning while others are unsupervised and induce the object structure). In addition, we can exploit this structure to develop algorithms that can perform inference on these RCMs to rapidly detect and recognize objects. This differs from more standard "flat models" of objects which have much less representational power if they wish to maintain efficient learning and inference. The basic properties of an RCM are illustrated in figures (1, 2). Because RCMs give a rich hierarchical description of objects and images they can be applied to a range of tasks including object detection, segmentation, parsing and image parsing. In all cases, we achieved state of the art results when evaluated on datasets with groundtruth.