Efficiently Learning Markov Random Fields from Dynamics

Abstract: An important task in high-dimensional statistics is learning the parameters or dependency structure of undirected graphical models, or Markov random fields (MRF), which have applications across computer science, machine learning, and statistics. Prior work on this problem assumes access to i.i.d. samples from the MRF distribution and state-of-the-art algorithms succeed with n^{\Theta(k)} runtime, where n is the dimension and k is the order of the interactions. However, well-known computational lower bounds imply that given i.i.d. samples from a sparse, order-k MRF, any learning algorithm likely requires n^{\Omega(k)} time. Moreover, in many economic, network, or statistical physics applications, even obtaining i.i.d. samples from the MRF distribution is unrealistic or itself computationally hard.

In this talk, we show that these fundamental barriers for learning MRFs can surprisingly be completely circumvented when learning from natural, dynamical observations. We prove that in bounded-degree MRFs, the dependency structure and parameters can be recovered using a trajectory of Glauber dynamics of length O(n \log n) with runtime O(n^2 \log n). The implicit constants depend only on the degree and non-degeneracy parameters of the model, but not the dimension n. Our results further evince the possibility of leveraging correlations across data to obtain provably better learning algorithms than are attainable from i.i.d. samples in important statistical settings.

Based on joint works with Ankur Moitra and Elchanan Mossel.

Bio: Jason is a Postdoctoral Associate in the Department of Mathematics at MIT, where he works with Elchanan Mossel. Previously, he received his PhD from the Department of Computer Science at Cornell University, where he was advised by Éva Tardos. He graduated from Yale University with degrees in Mathematics (B.S., with distinction) and Economics (B.A.).
He is broadly interested in theoretical computer science. In particular, his research focuses on theoretical problems at the intersection of algorithms, learning, game theory, networks, and randomness.