TY - JOUR
T1 - Sparse Subspace Clustering via Two-Step Reweighted L1-Minimization
T2 - Algorithm and Provable Neighbor Recovery Rates
AU - Wu, Jwo-Yuh
AU - Huang, Liang Chi
AU - Yang, Ming Hsun
AU - Liu, Chun Hung
N1 - Publisher Copyright:
© 1963-2012 IEEE.
PY - 2021/2
Y1 - 2021/2
N2 - Sparse subspace clustering (SSC) relies on sparse regression for accurate neighbor identification. Inspired by recent progress in compressive sensing, this paper proposes a new sparse regression scheme for SSC via two-step reweighted ℓ1 -minimization, which also generalizes a two-step ℓ1 -minimization algorithm introduced by E. J. Candès et al. in [The Annals of Statistics, vol. 42, no. 2, pp. 669-699, 2014] without incurring extra algorithmic complexity. To fully exploit the prior information offered by the computed sparse representation vector in the first step, our approach places a weight on each component of the regression vector, and solves a weighted LASSO in the second step. We propose a data weighting rule suitable for enhancing neighbor identification accuracy. Then, under the formulation of the dual problem of weighted LASSO, we study in depth the theoretical neighbor recovery rates of the proposed scheme. Specifically, an interesting connection between the locations of nonzeros of the optimal sparse solution to the weighted LASSO and the indexes of the active constraints of the dual problem is established. Afterwards, under the semi-random model, analytic probability lower/upper bounds for various neighbor recovery events are derived. Our analytic results confirm that, with the aid of data weighting and if the prior neighbor information is accurate enough, the proposed scheme with a higher probability can produce many correct neighbors and few incorrect neighbors as compared to the solution without data weighting. Computer simulations are provided to validate our analytic study and evidence the effectiveness of the proposed approach.
AB - Sparse subspace clustering (SSC) relies on sparse regression for accurate neighbor identification. Inspired by recent progress in compressive sensing, this paper proposes a new sparse regression scheme for SSC via two-step reweighted ℓ1 -minimization, which also generalizes a two-step ℓ1 -minimization algorithm introduced by E. J. Candès et al. in [The Annals of Statistics, vol. 42, no. 2, pp. 669-699, 2014] without incurring extra algorithmic complexity. To fully exploit the prior information offered by the computed sparse representation vector in the first step, our approach places a weight on each component of the regression vector, and solves a weighted LASSO in the second step. We propose a data weighting rule suitable for enhancing neighbor identification accuracy. Then, under the formulation of the dual problem of weighted LASSO, we study in depth the theoretical neighbor recovery rates of the proposed scheme. Specifically, an interesting connection between the locations of nonzeros of the optimal sparse solution to the weighted LASSO and the indexes of the active constraints of the dual problem is established. Afterwards, under the semi-random model, analytic probability lower/upper bounds for various neighbor recovery events are derived. Our analytic results confirm that, with the aid of data weighting and if the prior neighbor information is accurate enough, the proposed scheme with a higher probability can produce many correct neighbors and few incorrect neighbors as compared to the solution without data weighting. Computer simulations are provided to validate our analytic study and evidence the effectiveness of the proposed approach.
KW - compressive sensing
KW - discovery rate
KW - neighbor recovery
KW - performance guarantees
KW - sparse representation
KW - sparse subspace clustering
KW - Subspace clustering
KW - weighted LASSO
UR - http://www.scopus.com/inward/record.url?scp=85097181494&partnerID=8YFLogxK
U2 - 10.1109/TIT.2020.3039114
DO - 10.1109/TIT.2020.3039114
M3 - Article
SN - 0018-9448
VL - 67
SP - 1216
EP - 1263
JO - IEEE Transactions on Information Theory
JF - IEEE Transactions on Information Theory
IS - 2
M1 - 9262938
ER -