Nonconvex Statistical Learning and Estimation

To process high-dimensional data created by modern sensing modalities, it is necessary to exploit low-dimensional geometric structures of the information embedded in the data, including sparsity, low-rank structure, positivity, stationarity, and other structural constraints, in order to reduce the degrees of freedom to combat the curse of dimensionality. Notably, many of the most useful low-dimensional structures are naturally described using nonconvex constraints.

In recent years, statistical procedures have been developed to promote low-dimensional structures using convex relaxations, rather than directly attacking the nonconvex problems, exploiting the rich theory in convex analysis and convex optimization. However, by imposing convexity we pay our dues in computational guarantees. For example, in low-rank matrix estimation problems, convex relaxations typically require the run time to scale at least cubically in the matrix dimension. In contrast, a direct nonconvex approach only takes time (and also storage) proportional to reading the data, which is highly desirable when solving practical large-scale problems. The secret source that allows us to handle non-convexity is rooted in the fact that we are not interested in "generic" nonconvex problems, but rather, nonconvex problems arising in concrete signal and information processing scenarios. These scenarios typically provide us with much richer structural information. For example, it is discovered that, for several canonical information processing problems such as low-rank matrix completion, phase retrieval, and blind deconvolution, the natural non-convex formulations enjoy benign geometric structures, such that simple iterative algorithms, such as gradient descent and alternating minimization, can simultaneously achieve near-optimal statistical accuracy and low computational complexity.

Overview

Nonconvex Low-Rank Matrix Estimation

Nonconvex Low-Rank Tensor Estimation

Phase Retrieval and Blind Deconvolution via Nonconvex Optimization

Nonconvex Super Resolution

Neural Networks