Skip to:Content
|
Bottom
Cover image for Sparse modeling : theory, algorithms, and applications
Title:
Sparse modeling : theory, algorithms, and applications
Personal Author:
Series:
Chapman & Hall/CRC machine learning & pattern recognition series
Publication Information:
Boca Raton, FL : CRC Press : Taylor & Francis Group, 2015
Physical Description:
xviii, 231 p. : ill. ; 24 cm.
ISBN:
9781439828694
Added Author:

Available:*

Library
Item Barcode
Call Number
Material Type
Item Category 1
Status
Searching...
30000010344074 TA342 R57 2015 Open Access Book Book
Searching...

On Order

Summary

Summary

Sparse models are particularly useful in scientific applications, such as biomarker discovery in genetic or neuroimaging data, where the interpretability of a predictive model is essential. Sparsity can also dramatically improve the cost efficiency of signal processing.

Sparse Modeling: Theory, Algorithms, and Applications provides an introduction to the growing field of sparse modeling, including application examples, problem formulations that yield sparse solutions, algorithms for finding such solutions, and recent theoretical results on sparse recovery. The book gets you up to speed on the latest sparsity-related developments and will motivate you to continue learning about the field.

The authors first present motivating examples and a high-level survey of key recent developments in sparse modeling. The book then describes optimization problems involving commonly used sparsity-enforcing tools, presents essential theoretical results, and discusses several state-of-the-art algorithms for finding sparse solutions.

The authors go on to address a variety of sparse recovery problems that extend the basic formulation to more sophisticated forms of structured sparsity and to different loss functions. They also examine a particular class of sparse graphical models and cover dictionary learning and sparse matrix factorizations.


Author Notes

Irina Rish, Genady Grabarnik


Table of Contents

List of Figuresp. xi
Prefacep. xvii
1 Introductionp. 1
1.1 Motivating Examplesp. 4
1.1.1 Computer Network Diagnosisp. 4
1.1.2 Neuroimaging Analysisp. 5
1.1.3 Compressed Sensingp. 8
1.2 Sparse Recovery in a Nutshellp. 9
1.3 Statistical Learning versus Compressed Sensingp. 11
1.4 Summary and Bibliographical Notesp. 12
2 Sparse Recovery: Problem Formulationsp. 15
2.1 Noiseless Sparse Recoveryp. 16
2.2 Approximationsp. 18
2.3 Convexity: Brief Reviewp. 19
2.4 Relaxations of (P 0 ) Problemp. 20
2.5 The Effect of l q -Regularizer on Solution Sparsityp. 21
2.6 l 1 -norm Minimization as Linear Programmingp. 22
2.7 Noisy Sparse Recoveryp. 23
2.8 A Statistical View of Sparse Recoveryp. 27
2.9 Beyond LASSO: Other Loss Functions and Regularizersp. 30
2.10 Summary and Bibliographical Notesp. 33
3 Theoretical Results (Deterministic Part)p. 35
3.1 The Sampling Theoremp. 36
3.2 Surprising Empirical Resultsp. 36
3.3 Signal Recovery from Incomplete Frequency Informationp. 39
3.4 Mutual Coherencep. 40
3.5 Spark and Uniqueness of (P 0 ) Solutionp. 42
3.6 Null Space Properly and Uniqueness of (P 1 ) Solutionp. 45
3.7 Restricted Isometry Property (RIP)p. 46
3.8 Square Root Bottleneck for the Worst-Case Exact Recoveryp. 47
3.9 Exact Recovery Based on RIPp. 48
3.10 Summary and Bibliographical Notesp. 52
4 Theoretical Results (Probabilistic Part)p. 53
4.1 When Does RIP Hold?p. 54
4.2 Johnson-Lindenstrauss Lemma and RIP for Subgaussian Random Matricesp. 54
4.2.1 Proof of the Johnson-Lindenstrauss Concentration Inequalityp. 55
4.2.2 RIP for Matrices with Subgaussian Random Entriesp. 56
4.3 Random Matrices Satisfying RIPp. 59
4.3.1 Eigenvalues and RIPp. 60
4.3.2 Random Vectors, Isotropic Random Vectorsp. 60
4.4 RIP for Matrices with Independent Bounded Rows and Matrices with Random Rows of Fourier Transformp. 61
4.4.1 Proof of URIp. 64
4.4.2 Tail Bound for the Uniform Law of Large Numbers (ULLN)p. 67
4.5 Summary and Bibliographical Notesp. 69
5 Algorithms for Sparse Recovery Problemsp. 71
5.1 Univariate Thresholding is Optimal for Orthogonal Designsp. 72
5.1.1 l -norm Minimizationp. 73
5.1.2 l 1 -norm Minimizationp. 74
5.2 Algorithms for l 0 -norm Minimizationp. 76
5.2.1 An Overview of Greedy Methodsp. 79
5.3 Algorithms for l 1 -norm Minimization (LASSO)p. 82
5.3.1 Least Angle Regression for LASSO (LARS)p. 82
5.3.2 Coordinate Descentp. 86
5.3.3 Proximal Methodsp. 87
5.4 Summary and Bibliographical Notesp. 92
6 Beyond LASSO: Structured Sparsityp. 95
6.1 The Elastic Netp. 96
6.1.1 The Elastic Net in Practice: Neuroimaging Applicationsp. 100
6.2 Fused LASSOp. 107
6.3 Group LASSO: l 1 / l 2 Penaltyp. 109
6.4 Simultaneous LASSO: l 1 /l ∞ Penaltyp. 110
6.5 Generalizationsp. 111
6.5.1 Block l 1 /l q -Norms and Beyondp. 111
6.5.2 Overlapping Groupsp. 112
6.6 Applicationsp. 114
6.6.1 Temporal Causal Modelingp. 114
6.6.2 Generalized Additive Modelsp. 115
6.6.3 Multiple Kernel Learningp. 115
6.6.4 Multi-Task Learningp. 117
6.7 Summary and Bibliographical Notesp. 118
7 Beyond LASSO: Other Loss Functionsp. 121
7.1 Sparse Recovery from Noisy Observationsp. 122
7.2 Exponential Family, GLMs, and Bregman Divergencesp. 123
7.2.1 Exponential Familyp. 124
7.2.2 Generalized Linear Models (GLMs)p. 125
7.2.3 Bregman Divergencep. 126
7.3 Sparse Recovery with GLM Regressionp. 128
7.4 Summary and Bibliographic Notesp. 136
8 Sparse Graphical Modelsp. 139
8.1 Backgroundp. 140
8.2 Markov Networksp. 141
8.2.1 Markov Network Properties: A Closer Lookp. 142
8.2.2 Gaussian MRFsp. 144
8.3 Learning and Inference in Markov Networksp. 145
8.3.1 Learningp. 145
8.3.2 Inferencep. 146
8.3.3 Example: Neuroimaging Applicationsp. 147
8.4 Learning Sparse Gaussian MRFsp. 151
8.4.1 Sparse Inverse Covariance Selection Problemp. 152
8.4.2 Optimization Approachesp. 153
8.4.3 Selecting Regularization Parameterp. 160
8.5 Summary and Bibliographical Notesp. 165
9 Sparse Matrix Factorization: Dictionary Learning and Beyondp. 167
9.1 Dictionary Learningp. 168
9.1.1 Problem Formulationp. 169
9.1.2 Algorithms for Dictionary Learningp. 170
9.2 Sparse PCAp. 174
9.2.1 Backgroundp. 174
9.2.2 Sparse PCA: Synthesis Viewp. 176
9.2.3 Sparse PCA: Analysis Viewp. 178
9.3 Sparse NMF for Blind Source Separationp. 179
9.4 Summary and Bibliographical Notesp. 182
Epiloguep. 185
Appendix: Mathematical Backgroundp. 187
A.1 Norms, Matrices, and Eigenvaluesp. 187
A.1.1 Short Summary of Eigentheoryp. 188
A.2 Discrete Fourier Transformp. 190
A.2.1 The Discrete Whittaker-Nyquist-Kotelnikov-Shannon Sampling Theoremp. 191
A.3 Complexity of l 0 -norm Minimizationp. 192
A.4 Subgaussian Random Variablesp. 192
A.5 Random Variables and Symmetrizacion in R np. 197
A.6 Subgaussian Processesp. 199
A.7 Dudley Entropy Inequalityp. 200
A.8 Large Deviation for the Bounded Random Operatorsp. 202
Bibliographyp. 203
Indexp. 227
Go to:Top of Page