Iterative Hard Thresholding for SparseLowrank Linear Regression Prateek

  • Slides: 35
Download presentation
Iterative Hard Thresholding for Sparse/Low-rank Linear Regression Prateek Jain Microsoft Research, India Ambuj Tewari

Iterative Hard Thresholding for Sparse/Low-rank Linear Regression Prateek Jain Microsoft Research, India Ambuj Tewari Univ of Michigan Purushottam Kar MSR, India Praneeth Netrapalli MSR, NE

Microsoft Research India Our work • Foundations • Systems • Applications • Interplay of

Microsoft Research India Our work • Foundations • Systems • Applications • Interplay of society and technology • Academic and government outreach Our vectors of impact • Research impact • Company impact • Societal impact

Machine Learning and Optimization @ MSRI • High-dimensional Learning & Optimization Manik Varma Prateek

Machine Learning and Optimization @ MSRI • High-dimensional Learning & Optimization Manik Varma Prateek Jain Ravi Kannan Amit Deshpande Navin Goyal Sundarajan S. Vinod Nair Sreangsu Acharyya Kush Bhatia Aditya Nori Raghavendra Udupa Purushottam Kar • Extreme Classification • Online Learning / Multi-armed Bandits • Learning with Structured Losses We are Hiring! • • Interns • Distributed Machine Learning Post. Docs • Probabilistic Programming Applied Researchers • Privacy Preserving Machine Learning Full-time Researchers • SVMs & Kernel Learning • Ranking & Recommendation

Learning in High-dimensions 2 2 • • Non-convex • Complexity: NP-Hard

Learning in High-dimensions 2 2 • • Non-convex • Complexity: NP-Hard

Overview • Most popular approach: convex relaxation • Solvable in poly-time • Guarantees under

Overview • Most popular approach: convex relaxation • Solvable in poly-time • Guarantees under certain assumptions • Slow in practice Practical Algorithms For High-d ML Problems Theoretically Provable Algorithms For High-d ML Problems

Results • 6

Results • 6

Outline • Sparse Linear Regression • Lasso • Iterative Hard Thresholding • Our Results

Outline • Sparse Linear Regression • Lasso • Iterative Hard Thresholding • Our Results • Low-rank Matrix Regression • Low-rank Matrix Completion • Conclusions

Sparse Linear Regression n d = =

Sparse Linear Regression n d = =

Sparse Linear Regression •

Sparse Linear Regression •

Convex Relaxation • Non-differentiable Lasso Problem

Convex Relaxation • Non-differentiable Lasso Problem

Our Approach : Projected Gradient Descent • [Jain, Tewari, Kar’ 2014]

Our Approach : Projected Gradient Descent • [Jain, Tewari, Kar’ 2014]

 sort Hard Thresholding

sort Hard Thresholding

Convex-projections vs Non-convex Projections • 1 st order Optimality condition 0 -th order Optimality

Convex-projections vs Non-convex Projections • 1 st order Optimality condition 0 -th order Optimality condition • 0 order condition sufficient for convergence of Proj. Grad. Descent? • In general, NO • But, for certain specially structured problems, YES!!!

Restricted Isometry Property (RIP) •

Restricted Isometry Property (RIP) •

Popular RIP Ensembles •

Popular RIP Ensembles •

Proof under RIP • Hard Thresholding Triangle inequality RIP [Blumensath & Davies’ 09, Garg

Proof under RIP • Hard Thresholding Triangle inequality RIP [Blumensath & Davies’ 09, Garg & Khandekar’ 09]

What if RIP is not possible? •

What if RIP is not possible? •

Iterative Hard Thresholding: Larger Sparsity •

Iterative Hard Thresholding: Larger Sparsity •

Stronger Projection Guarantee •

Stronger Projection Guarantee •

Statistical Guarantees • Same as Lasso [J. , Tewari, Kar’ 2014]

Statistical Guarantees • Same as Lasso [J. , Tewari, Kar’ 2014]

General Result for Any Function • [J. , Tewari, Kar’ 2014]

General Result for Any Function • [J. , Tewari, Kar’ 2014]

Extension to other Non-convex Procedures • IHT-Fully Corrective • HTP [Foucart’ 12] • Co.

Extension to other Non-convex Procedures • IHT-Fully Corrective • HTP [Foucart’ 12] • Co. SAMP [Tropp & Neadell’ 2008] • Subspace Pursuit [Dai & Milenkovic’ 2008] • OMPR [J. , Tewari, Dhillon’ 2010] • Partial hard thresholding and two-stage family [J. , Tewari, Dhillon’ 2010]

Empirical Results Hard Thresholding Greedy 350 x 90 x (d)

Empirical Results Hard Thresholding Greedy 350 x 90 x (d)

More Empirical Results

More Empirical Results

Empirical Results: poor condition number

Empirical Results: poor condition number

Low-rank Matrix Regression 2 2 •

Low-rank Matrix Regression 2 2 •

Low-rank Matrix Regression •

Low-rank Matrix Regression •

Statistical Guarantees • [J. , Tewari, Kar’ 2014]

Statistical Guarantees • [J. , Tewari, Kar’ 2014]

Low-rank Matrix Completion • Special case of low-rank matrix regression • However, assumptions required

Low-rank Matrix Completion • Special case of low-rank matrix regression • However, assumptions required by the regression analysis not satisfied

Guarantees • [J. , Netrapalli’ 2015]

Guarantees • [J. , Netrapalli’ 2015]

Tale of two Lemmas •

Tale of two Lemmas •

Empirical Results Matrix Regression Hard Thresholding Trace-norm Number of data points (n)

Empirical Results Matrix Regression Hard Thresholding Trace-norm Number of data points (n)

Summary •

Summary •

Future Work • Generalized theory for such provable non-convex optimization • Performance analysis on

Future Work • Generalized theory for such provable non-convex optimization • Performance analysis on different models • Empirical comparisons on “real-world” datasets

Questions?

Questions?