Beyond the low-degree algorithm: mixtures of subcubes and their applications

© 2019 Association for Computing Machinery. We introduce the problem of learning mixtures of k subcubes over (0,1)n, which contains many classic learning theory problems as a special case (and is itself a special case of others). We give a surprising nO(log k)-time learning algorithm based on higher...

Full description

Bibliographic Details
Main Authors: Chen, Sitan (Author), Moitra, Ankur (Author)
Other Authors: Massachusetts Institute of Technology. Department of Electrical Engineering and Computer Science (Contributor), Massachusetts Institute of Technology. Computer Science and Artificial Intelligence Laboratory (Contributor), Massachusetts Institute of Technology. Department of Mathematics (Contributor)
Format: Article
Language:English
Published: ACM, 2022-07-06T21:03:19Z.
Subjects:
Online Access:Get fulltext
LEADER 03057 am a22001933u 4500
001 138050.2
042 |a dc 
100 1 0 |a Chen, Sitan  |e author 
100 1 0 |a Massachusetts Institute of Technology. Department of Electrical Engineering and Computer Science  |e contributor 
100 1 0 |a Massachusetts Institute of Technology. Computer Science and Artificial Intelligence Laboratory  |e contributor 
100 1 0 |a Massachusetts Institute of Technology. Department of Mathematics  |e contributor 
700 1 0 |a Moitra, Ankur  |e author 
245 0 0 |a Beyond the low-degree algorithm: mixtures of subcubes and their applications 
260 |b ACM,   |c 2022-07-06T21:03:19Z. 
856 |z Get fulltext  |u https://hdl.handle.net/1721.1/138050.2 
520 |a © 2019 Association for Computing Machinery. We introduce the problem of learning mixtures of k subcubes over (0,1)n, which contains many classic learning theory problems as a special case (and is itself a special case of others). We give a surprising nO(log k)-time learning algorithm based on higher-order multilinear moments. It is not possible to learn the parameters because the same distribution can be represented by quite different models. Instead, we develop a framework for reasoning about how multilinear moments can pinpoint essential features of the mixture, like the number of components. We also give applications of our algorithm to learning decision trees with stochastic transitions (which also capture interesting scenarios where the transitions are deterministic but there are latent variables). Using our algorithm for learning mixtures of subcubes, we can approximate the Bayes optimal classifier within additive error ϵ on k-leaf decision trees with at most s stochastic transitions on any root-to-leaf path in nO(s+log k) · poly(1/ϵ) time. In this stochastic setting, the classic nO(log k) · poly(1/ϵ)-time algorithms of Rivest, Blum, and Ehrenfreucht-Haussler for learning decision trees with zero stochastic transitions break down because they are fundamentally Occam algorithms. The low-degree algorithm of Linial-Mansour-Nisan is able to get a constant factor approximation to the optimal error (again within an additive ϵ) and runs in time nO(s+log(k/ϵ)). The quasipolynomial dependence on 1/ϵ is inherent to the low-degree approach because the degree needs to grow as the target accuracy decreases, which is undesirable when ϵ is small. In contrast, as we will show, mixtures of k subcubes are uniquely determined by their 2 logk order moments and hence provide a useful abstraction for simultaneously achieving the polynomial dependence on 1/ϵ of the classic Occam algorithms for decision trees and the flexibility of the low-degree algorithm in being able to accommodate stochastic transitions. Using our multilinear moment techniques, we also give the first improved upper and lower bounds since the work of Feldman-O'Donnell-Servedio for the related but harder problem of learning mixtures of binary product distributions. 
546 |a en 
655 7 |a Article 
773 |t 10.1145/3313276.3316375