Low-rank matrix recovery (LRMR) has been becoming an increasingly popular technique for analyzing data with missing entries, gross corruptions, and outliers. As a significant component of LRMR, the model of low-rank representation (LRR) seeks the lowest-rank representation among all samples and it is robust for recovering subspace structures. This paper attempts to solve the problem of LRR with partially observed entries. Firstly, we construct a nonconvex minimization by taking the low rankness, robustness, and incompletion into consideration. Then we employ the technique of augmented Lagrange multipliers to solve the proposed program. Finally, experimental results on synthetic and real-world datasets validate the feasibility and effectiveness of the proposed method.
1. Introduction
In the community of pattern recognition, machine learning, and computer vision, the investigated datasets usually have intrinsically low-rank structure although they are probably high-dimensional. Low-rank matrix recovery (LRMR) [1–3] is just a type of model which utilizes the crucial low-complexity information to complete missing entries, recover sparse noise, identify outliers, and build an affinity matrix. It also can be regarded as the generalization of compressed sensing from one order to two orders due to the fact that the low rankness of a matrix is equivalent to the sparsity of its singular values. Recently, LRMR has received more and more attentions in the fields of information science and engineering and achieved great success in video background modeling [4, 5], collaborative filtering [6, 7], and subspace clustering [3, 8, 9], to name just a few.
Generally, LRMR is mainly composed of three appealing types, that is, matrix completion (MC) [1], robust principal component analysis (RPCA) [2, 4], and low-rank representation (LRR) [3, 8]. Among them, MC aims to complete the missing entries with the aid of the low-rank property and is initially described as an affine rank minimization problem. In the past few years, the affine rank minimization is convexly relaxed into a nuclear norm minimization [10] and it is proven that if the number of sampled entries and singular vectors satisfy some conditions, then most low-rank matrices can be perfectly recovered by solving the aforementioned convex program [1].
Classical principal component analysis (PCA) [11] is very effective to small Gaussian noise, but it does not work well in practice when data samples are corrupted by outliers or large sparse noise. For this purpose, several robust variants of PCA have been proposed successively during the past two decades [12, 13]. Since the seminal research work [4], the principal component pursuit (PCP) approach has become a standard for RPCA. This approach minimizes a weighted combination of the nuclear norm and the l1 norm with linear equality constraints. It is proven that both the low-rank and the sparse components can be recovered exactly with dominant probability under some conditions by solving PCP [4].
In subspace clustering, a commonly used assumption is that the data lie in the union of multiple low-rank subspaces and each subspace has sufficient samples compared with its rank. Liu et al. [3] proposed a robust subspace recovery technique via LRR. Any sample in each subspace can be represented as the linear combination of the bases. The low complexity of the linear representation coefficients is very useful in exploiting the low-rank structure. LRR attempts to seek the lowest-rank representation of all data jointly and it is demonstrated that the data contaminated by outliers can be exactly recovered under certain conditions by solving a convex program [8]. If the bases are chosen as the columns of an identity matrix and the l1 norm is employed to measure the sparsity, then LRR is changed into the PCP of RPCA.
For the datasets with missing entries and large sparse corruption, the robust recovery of subspace structures may be a challenging task. The available algorithms to MC are not robust to gross corruption. Moreover, a large quantity of missing values will bring out the degeneration of recovering performance for LRR or RPCA. In this paper, we attempt to address the problem of low-rank subspace recovery in the presence of missing values and sparse noise. Specifically speaking, we present a model of incomplete low-rank representation (ILRR) which is a direct generalization of LRR. The ILRR model can be boiled down to a nonconvex optimization model which minimizes the combination of the nuclear norm and the l2,1-norm. To solve this program, we design an iterative scheme by applying the method of inexact augmented Lagrange multipliers (ALM).
The rest of this paper is organized as follows. Section 2 briefly reviews preliminaries and related works on LRMR. The model and algorithm for ILRR are presented in Section 3. In Section 4, we discuss the extension of ILRR and its relationship with the existing works. We compare the performance of ILRR with the state-of-the-art algorithms on synthetic data and real-world datasets in Section 5. Finally, Section 6 draws some conclusions.
2. Preliminaries and Related Works
This section introduces the relevant preliminary material concerning matrices and representative models of low-rank matrix recovery (LRMR).
The choice of matrix norms plays a significant role in LRMR. In the following, we present four important types of matrix norms. For arbitrary X∈Rm×n, the Frobenius norm of X is expressed by XF=∑i=1m∑j=1nXij2, the l2,1-norm is X2,1=∑j=1n∑i=1mXij2, thel1-norm is X1=∑i=1m∑j=1n|Xij|, and the nuclear norm is X*=∑i=1min(m,n)σi, where Xij is the (i,j)th entry of X and σi is the ith largest singular value. Among them, the matrix nuclear norm is the tightest convex relaxation of the rank function, and the l2,1-norm and the l1-norm are frequently used to measure the sparsity of a noise matrix.
Consider a data matrix D∈Rm×n stacked by n training samples, where each column of D indicates a sample with the dimensionality of m. Within the field of LRMR, the following three proximal minimization problems [3, 4] are extensively employed:
(1)minXλX1+12X-DF2,minXλX*+12X-DF2,minXλX2,1+12X-DF2,
where λ is a positive constant used to balance the regularization term and the approximation error. For given λ and D, we define three thresholding operators Sλ(D), Dλ(D), and Wλ(D) as follows:
(2)Sλ(D)ij=maxDij-λ,0,DλD=USλΣVT,Wλ(D)ij=max∑k=1mDkj2-λ,0Dij∑k=1mDkj2,
where UΣVT is the singular value decomposition (SVD) of D. It is proven that the aforementioned three optimization problems have closed-form solutions denoted by Sλ(D) [4], Dλ(D) [14], and Wλ(D) [8], respectively.
We assume that D is low-rank. Because the degree of freedom of a low-rank matrix is far less than its number of entries, it is possible to recover exactly all missing entries from partially observed entries as long as the number of sampled entries satisfies certain conditions. Formally, the problem of matrix completion (MC) [1] can be formulated as follows:
(3)minXX*s.t.vvivXij=Dij,(i,j)∈Ω,
where Ω⊂m×[n], [m]={1,2,…,m}, [n]={1,2,…,n}. We define a linear projection operator PΩ(•):Rm×n→Rm×n as follows:
(4)PΩ(X)ij=Xij,if(i,j)∈Ω0,otherwise.
Hence, the constraints in problem (3) can be rewritten as PΩ(X)=PΩ(D).
PCA obtains the optimal estimate for small additive Gaussian noise but breaks down for large sparse contamination. Here, the data matrix D is assumed to be the superposition of a low-rank matrix A and a sparse noise matrix E. In this situation, robust principal component analysis (RPCA) is very effective to recover both the low-rank and the sparse components by solving a convex program. Mathematically, RPCA can be described as the following nuclear norm minimization [2]:
(5)minA,EA*+λE1s.t.vivvD=A+E,
where λ is a positive weighting parameter. Sequentially, RPCA is generalized into a stable version which is simultaneously stable to small perturbations and robust to gross sparse corruption [15].
We further assume that the dataset is self-expressive and the representation coefficients matrix is also low-rank. Based on the above two assumptions, the model of low-rank representation (LRR) [3] is expressed as
(6)minZ,EZ*+λE2,1s.t.vvvD=DZ+E,
where Z∈Rn×n is the coefficient matrix, E is the noise matrix, and λ is a positive trade-off parameter. This model is very effective to detect outliers and the optimal Z is in favor of the robust subspace recovery. In subspace clustering, the affinity matrices can be constructed by the optimal Z to problem (6).
Problems (3), (5), and (6) belong to the nuclear norm minimizations. The existing algorithms to the preceding optimizations mainly include the iterative thresholding, the accelerated proximal gradient, the dual approach, and the augmented Lagrange multipliers (ALM) [16]. These algorithms are scalable owing to the adoption of first-order information. Among them, ALM, also called alternating direction method of multipliers (ADMM) [17], is a very popular and effective method to solve the nuclear norm minimizations.
3. Model and Algorithm of Incomplete Low-Rank Representation
This section proposes a model of low-rank representation for incomplete data and develops a corresponding iterative scheme for this model.
3.1. Model
We consider an incomplete data matrix M∈Rm×n and denote the sampling index set by Ω. The (i,j)th entry Mij of M is missing if and only if (i,j)∉Ω. For the sake of convenience, we set all missing entries of M to zeros. To recover simultaneously the missing entries and the low-rank subspace structure, we construct an incomplete low-rank representation (ILRR) model:
(7)minZ,E,DZ*+λE2,1s.t.vvivD=DZ+EvvvvvvPΩ(D)=M,
where λ is a positive constant and D corresponds to the completion argument of M. If there is not any missing entry, that is, Ω=[m]×[n], then the above model is equivalent to LRR. In other words, LRR is a special case of ILRR.
In order to solve conveniently the nonconvex nuclear norm minimization (7), we introduce two auxiliary matrix variables X∈Rm×n and J∈Rn×n. Under this circumstance, the above optimization problem is reformulated as
(8)minZ,E,D,J,XJ*+λE2,1s.t.vvvvvvD=XZ+EvvvvvvvvZ=J,X=D,PΩ(D)=M.
This minimization problem is equivalent to
(9)minZ,E,D,J,XJ*+λE2,1vvvvvvvvvvv+μ2D-XZ-EF2+Z-JF2+X-DF2s.t.vvvviivD=XZ+EvvvvvviivvZ=J,X=D,PΩ(D)=M,
where the factor μ>0. Without considering the constraint PΩ(D)=M, we construct the augmented Lagrange function of problem (9) as follows:
(10)L=J*+λE2,1+μ2D-XZ-EF2+Z-JF2+X-DF2+Y1,D-XZ-E+Y2,Z-J+Y3,X-D,
where •,• is the inner product operator between matrices and Yi is a Lagrange multiplier matrix, i=1,2,3. In the next part, we will propose an inexact augmented Lagrange multipliers (IALM) method to solve problem (8) or problem (9).
3.2. Algorithm
Inexact ALM (IALM) method employs an alternating update strategy and it minimizes or maximizes the function L with respect to each block variable at each iteration. Let Y={Y1,Y2,Y3}.
Computing J. When J is unknown and other variables are fixed, the calculation procedure of J is as follows:
(11)J≔argminJL=argminJJ*+Y2,Z-J+μ2Z-JF2=argminJ1μJ*+12Z+Y2μ-JF2=D1/μZ+Y2μ.
Computing Z. If matrix Z is unknown and other variables are given, Z is updated by minimizing L:
(12)Z≔argminZL=argminZY1,D-XZ-E+Y2,Z-J+μ2D-XZ-EF2+Z-JF2=argminZD-E+Y1μ-XZF2+Z-J-Y2μF2.
Let g(Z)=(D-E+Y1/μ)-XZF2+Z-(J-Y2/μ)F2. By setting the derivative of g(Z) to zero, we have
(13)∇gZ=2XTXZ-2XTD-E+Y1μ+2Z-2J-Y2μ=0
or, equivalently,
(14)Z≔XTX+I-1XTD-E+Y1μ+J-Y2μ,
where I is an n-order identity matrix.
Computing D. The update formulation of matrix D is calculated as follows:
(15)D≔argminDL=argminDY1,D-XZ-E+Y3,X-D+μ2D-XZ-EF2+X-DF2=argminDD-XZ+E-Y1μF2+X+Y3μ-DF2=12XZ+E-Y1μ+X+Y3μ.
Considering the constraint PΩ(D)=M, we further obtain the iteration formulation of D(16)D≔M+12PΩ-XZ+E-Y1μ+X+Y3μ,
where Ω- is the complementary set of Ω.
Computing E. FixZ,D,J,X, and Y and minimize L with respect to E:(17)E≔argminEL=argminEλμE2,1+12D-XZ-E+1μY1F2=Wλ/μD-XZ+Y1μ.
Computing X. Fix Z,D,J,E, and Y to calculate X as follows:
(18)X≔argminXL=argminXhX,
where h(X)=D-XZ-E+Y1/μF2+X-D+Y3/μF2. The derivative of h(X) is ∇h(X)=2XZZT-2(D-E+Y1/μ)ZT+2X-2(D-Y3/μ). Hence, we obtain the update of X by setting ∇h(X)=0:
(19)X≔D-E+Y1μZT+D-Y3μZZT+I-1.
Computing Y. Given Z,E,D,J, and X, we calculate Y as follows
(20)Y≔argmaxYL.
In the detailed implementation, Y is updated according to the following formulations:
(21)Y1≔Y1+μD-XZ-E,Y2≔Y2+μZ-J,Y3≔Y3+μX-D.
We denote by Om×n the m×n zeros matrix. The whole iterative procedure is outlined in Algorithm 1. The stopping condition of Algorithm 1 can be set as
(22)maxD-XZ-EF,Z-JF,X-DF<ε,
where ε is a sufficiently small positive number.
Algorithm 1: Solving ILRR by IALM.
Input: Data matrix M, sampling index set Ω, compromising parameter λ.
When solving ILRR via the IALM method, the block variables J,Z,D,E,X are updated alternatively. Now, we update simultaneously these five block variables; namely,
(23)J,Z,D,E,X≔argminJ,Z,D,E,XL.
The modified method is called the exact ALM method. Since the objective function in problem (8) is continuous, the exact ALM method is convergent [18]. However, it is still difficult to prove the convergence of the IALM. There are two reasons for the difficulty: one is the existence of nonconvex constraints in (8) and the other is that the number of block variables is more than two. Nevertheless, the experimental results of Section 5 demonstrate the validity and effectiveness of Algorithm 1 in practice.
4. Model Extensions
In ILRR model, the main aim of the term E2,1 is to enhance the robustness to noise and outliers. If we do not consider outliers, then E2,1 should be replaced with E1. For the new ILRR model, we can design an algorithm by only revising Step 4 of Algorithm 1 as follows:
(24)E≔argminEλμE1+12D-XZ-E+1μY1F2=Sλ/μD-XZ+Y1μ.
If we substitute EF2 for E2,1 and set Z=ZT, then problem (7) is the incomplete version of low-rank subspace clustering (LRSC) with uncorrupted data [9]. If we replace Z* and E2,1 by Z1 and E1, respectively, and incorporate diag(Z)=0 into the constraints, then problem (7) is the incomplete version of sparse subspace clustering (SSC) without dense errors [19].
ILRR uses the data D itself as the dictionary. Now, we extend the dictionary and noise sparsity to the more general cases. As a result, we obtain a comprehensive form of ILRR:
(25)minZ,E,A,DZ*+λEls.t.vvvivD=AZ+E,PΩ(D)=M,
where A∈Rm×r represents the dictionary, Z∈Rr×n is the coefficients matrix, and El indicates a certain norm of E. If A is an m-order identity matrix and El is chosen as E1, then problem (25) corresponds to the incomplete version of RPCA [20]. If we further reinforce λ=0, then problem (25) becomes the equivalent formulation of MC [16]. Moreover, if A is an unknown orthogonal matrix and λ=0, then problem (25) is equivalent to matrix decomposition method for MC [21]. Finally, if we let D and A be stacked by the testing samples and the training samples, respectively, let Z* be replaced by Z1, and let λ=0, then problem (25) is changed into the incomplete version of robust pattern recognition via sparse representation [22].
5. Experiments
In this section, we validate the effectiveness and efficiency of the proposed method by carrying out experiments on synthetic data and real-world datasets. The experimental results of incomplete low-rank representation (ILRR) are compared with that of other state-of-the-art methods: sparse subspace clustering (SSC), low-rank subspace clustering (LRSC), and low-rank representation (LRR). For the latter three methods, the missing values are replaced by zeros. For SSC and LRSC, their parameters are tuned to achieve the best performance. The tolerant error ε is set to 10-8 in all experiments.
5.1. Synthetic Data
We generate randomly an orthogonal matrix U1∈R200×4 and a rotation matrix T∈R200×200. Four other orthogonal matrices are constructed as Ui+1=TUi, i=1,2,3,4. Thus, five independent low-rank subspaces {Si}i=15⊂R200 are spanned by the columns of Ui, respectively. We draw 40 data vectors from each subspace by Mi=UiQi, i=1,…,5, where the entries of Qi∈R4×40 are independent of each other and they obey standard normal distributions. We set M′=(M1,…,M5) and choose randomly its 20 column vectors to be corrupted. In this part, the chosen column vector m is contaminated by Gaussian noise with zero mean and standard deviation 0.1m2. The resulting matrix is expressed as M′′.
We draw samples on M′′ according to a uniform distribution and denote by Ω the sampling index set. The sampling ratio (SR) is defined as SR=Ω/2002, where Ω is the cardinality of Ω and SR=1 means that no entry is missing. Hence, an incomplete matrix is generated by M=PΩ(M′′). The trade-off parameter λ in both LRR and ILRR is set to 0.1. After running Algorithm 1, we obtain the optimal low-rank representation matrix Z. On the basis of Z, we construct an affinity matrix Z~=(Z+ZT)/2, where • means the absolute value operator. Then we choose spectral clustering [23] as the clustering algorithm and evaluate the clustering performance by normalized mutual information (NMI). Let C be a set of true cluster labels and let C′ be a set of clusters obtained from the spectral clustering algorithm. NMI is calculated as
(26)NMI=MI(C,C′)H(C)H(C′),
where MI(C,C′) is the mutual information metric and H(C) is the entropy of C. NMI takes values in [0,1] and a larger NMI value indicates a better clustering performance.
In the experimental implementation, we vary the SR from 0.2 to 1 with an interval of 0.1. For each fixed SR, we repeat the experiments 10 times and report the average NMI. We first compare the affinity matrices Z~ produced by SSC, LRSC, LRR, and ILRR, as shown partially in Figure 1. From Figure 1, we observe that our method exhibits obvious block structures compared to other three methods, whereas SSC, LRSC, and LRR show no block-diagonal structure with SR=0.3. This observation means that only ILRR can keep compact representation for the same subspace and divergent representation for different subspaces when a large number of values are missing.
Affinity matrices comparison on synthetic data.
SSC (SR = 0.5)
LRSC (SR = 0.5)
LRR (SR = 0.5)
ILRR (SR = 0.5)
SSC (SR = 0.3)
LRSC (SR = 0.3)
LRR (SR = 0.3)
ILRR (SR = 0.3)
Secondly, we compare NMI values of ILRR with that of other three methods at different SR, as shown in Figure 2. It can be seen from this figure that NMI values of SSC, LRR, and ILRR are almost one if SR≥0.7, while ILRR has higher NMI values than other three methods if SR≤0.6. In other words, the NMI values of SSC, LRSC, and LRR are reduced drastically with the decreasing of SR. These observations verify that ILRR is more robust than other methods in the presence of missing values.
NMI comparison on synthetic data.
5.2. Face Clustering
We carry out face clustering experiments on a part of Extended Yale Face Database B [24] with large corruptions. This whole database consists of 38 objects and each object has about 64 images. We choose the first 10 objects and each image is resized into 32 × 32. Thus, the face images dataset is represented by a matrix with size of 1024 × 640. Each column of the data matrix is normalized to an l2 unit length in consideration of variable illumination conditions and poses.
The generation manner of sampling set Ω is similar to that of the previous part. We vary the values of SR from 0.1 to 1 with an interval of 0.1 and set λ=1.2 in LRR and ILRR. The comparison of NMI values is compared among SSC, LRSC, LRR, and ILRR, as shown in Figure 3, where each NMI value is the average result of ten repeated experiments. It can be seen from Figure 3 that ILRR achieves relatively stable NMI values if SR≥0.2, while other three methods obtain much worse NMI values than ILRR if SR≤0.9. This observation shows that SSC, LRSC, and LRR are more sensitive than ILRR on SR.
NMI comparison on Extended Yale B.
Furthermore, ILRR has an advantage in recovering the low-rank components over other three methods. In the following, we give the recovery performance contrast of ILRR. Here, we only consider two different SR; that is, SR=0.3 and SR=0.6. For these two given SR, the original images, sampled images, recovered images, and noise images are shown partially in Figures 4 and 5, respectively. Among these sampled images, the unsampled entries are shown in white. From these two figures, we can see that ILRR not only corrects automatically the corruptions (shadow and noise) but also recovers efficiently the low-rank components.
Face images recovery of ILRR on Extended Yale B with SR=0.3.
Original images
Sampled images
Recovered images
Noise images
Face images recovery of ILRR on Extended Yale B with SR=0.6.
Original images
Sampled images
Recovered images
Noise images
5.3. Motion Segmentation
In this part, we test the proposed ILRR method for the task of motion segmentation. We consider eight sequences of outdoor traffic scenes, a subset of the Hopkins 155 dataset [25]. These sequences were taken by a moving handheld camera and they tracked two cars translating and rotating on a street, as shown partly in Figure 6. These sequences are drawn from two or three motions, where one motion corresponds to one subspace. There are 8 clustering tasks in total since motion segmentation of each sequence is a sole clustering task. The tasks of motion segmentation are carried out according to the given features extracted and tracked in multiple frames of the above video sequences.
Images from traffic sequences on the Hopkins 155 dataset.
The extracted feature points of each sequence can be reshaped into an approximately low-rank matrix with n columns, where n equals the number of feature points. In the detailed experimental implementation, we consider different SR varying from 0.2 to 1 with an interval 0.2. The regularization parameter set λ in LRR and ILRR is set to 2. For fixed sequence and fixed SR, each method is repeated 10 times and the mean NMI values are recorded. We report the mean and the standard derivation of NMI values among 8 sequences, as shown in Table 1.
NMI (%) comparison of different methods on traffic sequences.
SR
SSC
LRSC
LRR
ILRR
0.2
5.15 ± 4.97
8.76 ± 4.66
9.97 ± 5.26
66.18 ± 15.40
0.4
13.74 ± 7.00
15.61 ± 10.50
15.07 ± 11.45
89.98 ± 13.24
0.6
21.45 ± 15.88
18.48 ± 11.84
17.78 ± 13.71
90.60 ± 13.72
0.8
50.97 ± 25.52
18.85 ± 14.06
17.46 ± 15.71
96.71 ± 6.84
1.0
100 ± 0.00
96.38 ± 7.60
96.83 ± 6.41
96.83 ± 6.41
From Table 1, we can see that ILRR obtains very high NMI values especially for SR≥0.2, while NMI values of other three methods are unacceptable for SR≤0.8. Although SSC segments exactly each subspace when SR=1.0, small SR has a fatal influence on SSC’s segmentation performance. In summary, ILRR is more robust to a large number of missing values than other three methods.
6. Conclusions
In this paper, we investigate the model of low-rank representation for incomplete data, which can be regarded as the generalization of low-rank representation and matrix completion. For the model of incomplete low-rank representation, we propose an iterative scheme based on augmented Lagrange multipliers method. Experimental results show that the proposed method is feasible and efficient in recovering low-rank structure, completing missing entries, and removing noise. It still needs further research on how to construct a general model of low-rank matrix recovery and design the corresponding algorithm.
Conflict of Interests
The authors declare that there is no conflict of interests regarding the publication of this paper.
Acknowledgments
This work is partially supported by the National Natural Science Foundation of China (no. 61403298, no. 11326204, no. 11401457, and no. 11401357), by the Natural Science Basic Research Plan in Shaanxi Province of China (no. 2014JQ8323 and no. 2014JQ1019), by the Shaanxi Provincial Education Department (no. 2013JK0587 and no. 2013JK0588), and by the Hanzhong Administration of Science & Technology (no. 2013hzzx-39).
CandèsE. J.RechtB.Exact matrix completion via convex optimization20099671777210.1007/s10208-009-9045-52-s2.0-71049116435MR2565240WrightJ.PengY.MaY.GaneshA.RaoS.Robust principal component analysis: exact recovery of corrupted low-rank matrices by convex optimization22Proceedings of the 23rd Annual Conference on Neural Information Processing Systems (NIPS '09)December 2009208020882-s2.0-84863367863LiuG.LinZ.YuY.Robust subspace segmentation by low-rank representationProceedings of the 27th International Conference on Machine Learning2010663670CandèsE. J.LiX.MaY.WrightJ.Robust principal component analysis?2011583, article 1110.1145/1970392.1970395MR28110002-s2.0-79960675858BouwmansT.ZahzahE. H.Robust PCA via principal component pursuit: a review for a comparative evaluation in video surveillance201412222342-s2.0-8489807983910.1016/j.cviu.2013.11.009RennieJ. D. M.SrebroN.Fast maximum margin matrix factorization for collaborative predictionProceedings of the 22nd International Conference on Machine Learning (ICML '05)August 200571371910.1145/1102351.11024412-s2.0-31844451557NieF.WangH.CaiX.Robust matrix completion via joint Schatten p-norm and lp-norm minimizationProceedings of the 12th IEEE International Conference on Data Mining2012566574LiuG.LinZ.YanS.SunJ.YuY.MaY.Robust recovery of subspace structures by low-rank representation201335117118410.1109/tpami.2012.882-s2.0-84870197517VidalR.FavaroP.Low rank subspace clustering (LRSC)2014431476110.1016/j.patrec.2013.08.0062-s2.0-84899446059RechtB.FazelM.ParriloP. A.Guaranteed minimum-rank solutions of linear matrix equations via nuclear norm minimization201052347150110.1137/070697835MR26805432-s2.0-78549288866JolliffeI. T.20022ndSpringer Series in StatisticsDe La TorreF.BlackM. J.Robust principal component analysis for computer visionProceedings of the 8th International Conference on Computer VisionJuly 20013623692-s2.0-0034849799DingC.ZhouD.HeX.R1-PCA: rotational invariant L1-norm principal component analysis for robust subspace factorizationProceedings of the 23rd International Conference on Machine Learning2006281288CaiJ.-F.CandèsE. J.ShenZ.A singular value thresholding algorithm for matrix completion201020419561982MR26002482-s2.0-7795129104610.1137/080738970ZhouZ.LiX.WrightJ.CandèsE.MaY.Stable principal component pursuitProceedings of the IEEE International Symposium on Information Theory (ISIT '10)June 2010Austin, Tex, USA1518152210.1109/isit.2010.55135352-s2.0-77955670622LinZ.ChenM.WuL.The augmented Lagrange multiplier method for exact recovery of corrupted low-rank matrices2010UILU-ENG-09-2215UIUCYuanX.YangJ.Sparse and low-rank matrix decomposition via alternating direction methods2009BertsekasD. P.1982New York, NY, USAAcademic PressComputer Science and Applied MathematicsMR690767ElhamifarE.VidalR.Sparse subspace clustering: algorithm, theory, and applications201335112765278110.1109/tpami.2013.572-s2.0-84884541998ShiJ.ZhengX.YongL.Incomplete robust principal component analysis20145615311538LiuY.ShangF.An efficient matrix factorization method for tensor completion201320430731010.1109/LSP.2013.22454162-s2.0-84874642792WrightJ.YangA. Y.GaneshA.SastryS. S.MaY.Robust face recognition via sparse representation200931221022710.1109/TPAMI.2008.792-s2.0-61549128441NgA.JordanM.WeissY.On spectral clustering: analysis and an algorithm200214849856GeorghiadesA. S.BelhumeurP. N.KriegmanD. J.From few to many: illumination cone models for face recognition under variable lighting and pose200123664366010.1109/34.9274642-s2.0-0035363672VidalR.HartleyR.Three-view multibody structure from motion200830221422710.1109/TPAMI.2007.11792-s2.0-37549058472