A Novel Robust Principal Component Analysis Algorithm of Nonconvex Rank Approximation

. Noise exhibits low rank or no sparsity in the low-rank matrix recovery, and the nuclear norm is not an accurate rank approximation of low-rank matrix. In the present study, to solve the mentioned problem, a novel nonconvex approximation function of the low-rank matrix was proposed. Subsequently, based on the nonconvex rank approximation function, a novel model of robust principal component analysis was proposed. Such model was solved with the alternating direction method, and its convergence was veriﬁed theoretically. Subsequently, the background separation experiments were performed on the Wallﬂower and SBMnet datasets. Furthermore, the eﬀectiveness of the novel model was veriﬁed by numerical experiments.


Introduction
As fuelled by the advancement of big data and artificial intelligence industry, the data in the image processing, semantic analysis, and other application fields exhibit extremely large scale and dimension, leading to a significant difficulty of data processing and analysis. It has been reported from research that the nature of the mentioned data is commonly characterized by low-rank subspace [1]. For this reason, the problems in the fields mentioned can be transformed into the problems such as low-rank matrix representation [2] or recovery [3]. In industrial production, the collected data usually contain outliers or noises for the factors (e.g., environment and equipment). For instance, as impacted by cloud or other objects' covering, the geomorphological map collected by the sensors contains shadows; face image exhibits differences in brightness for the effect of the light; the quality of nuclear magnetic resonance image degrades because of object motion, and the defects of imaging systems, as well as inherent noise and external interference of recording equipment. Under the impact of the mentioned practical problems, image denoising, image reconstruction, image restoration, and image completion have become research hotpots.
Image reconstruction and image denoising aim at discovering the low-rank part of the image-related data, i.e., the background in the image. Moreover, image recognition of face recognition and motion segmentation can exploit the subspace clustering method to recover the low-rank part of the image data, i.e., seeking the low-rank matrix representation of the data. Seeking the low-rank part and low-rank matrix representation in the data, denoising, and completing can be essentially considered the low-rank matrix recovery from the perspective of mathematics. Accordingly, the lowrank matrix recovery model creates a unified framework for the mentioned problems existing in image processing.
From the mathematic perspective, each piece of data is arranged into a column vector, and then a data matrix X ∈ R m×n is formed. It is assumed that X can be decomposed as where X, Y, E ∈ R m×n ; Y denotes the low-rank matrix, representing the intrinsic low-rank subspace characteristics of the real data; and E is the sparse matrix, representing the outliers or noises in the real data. To solve the low-rank matrix recovery problem, classical principal component analysis and robust principal component analysis can be conducted, whereas numerous problems may exist in the problem-solving process. Problem 1. Classical principal component analysis [4,5] method is adopted to solve the model below: where ‖·‖ 2 F denotes the Frobenius norm of the matrix and rank(Y) represents the rank of the matrix. First, the rank of low-rank matrix was assessed, which was not larger than k. Second, E should be the only small independent identically distributed Gaussian noises. Equation (2) exploits the singular value matrix decomposition to determine the singular value matrix decomposition of data matrix X, k main singular values, and corresponding left singular vectors, which were subsequently employed to indicate the major features of the analysis or recognition pattern. Principal component analysis (PCA), as an effective tool for data analysis and dimensional reduction, has been extensively applied in various fields. However, if the data are seriously polluted (e.g., large Gaussian noise or only sparse non-Gaussian noise), conventional PCA will fail. Problem 2. In conventional subspace learning models, principal component analysis (PCA), linear discriminant analysis (LDA), and independent component analysis [6] (ICA) can significantly denoise small Gaussian noise contained in data, whereas they are very sensitive to the data containing outliers or sparse large noises. To recover the low-rank structure from the data containing sparse large noises, Chandrasekaran et al. [7] and Wright et al. [8] also built a robust principal component analysis model, which is expressed below: where ‖E‖ 0 denotes the ℓ 0 -norm of matrix E, i.e., the number of nonzero elements in E, and η represents the positive trade-off parameter. e objective function of equation (3) is a nonlinear nonconvex function, and its solution is NP problem and relatively difficult. e compression sensing theory was introduced in this problem to solve the convex relaxation problem: where ‖Y‖ * denotes the kernel norm of matrix Y (trace norm or Ky-Fan norm), i.e., ‖Y‖ * � r i�1 σ i (Y), where c denotes the rank of Y and σ i (Y) is the i-th singular value of Y. ‖E‖ 1 represents the ℓ 1 -norm of matrix E, i.e., ‖E‖ 1 � i,j |E ij | is the sum of the absolute values of all elements of E.
When the singular vector of the low-rank matrix distributes reasonably and the nonzero elements of the sparse matrix distribute uniformly, equation (4) recovers the lowrank part and the noisy part from the matrix of the real data exhibiting a probability close to 1. However, in the definition of nuclear norm, all singular values are treated unequally. All singular values in the problem should upregulate the rank by 1 regardless of their values. However, in terms of the nuclear norm, the contribution of the nonzero singular values to the nuclear norm is directly determined by its value. Accordingly, the nuclear norm is commonly regulated by some large singular values, even if there are only a few large singular values, whose effect cannot be overlooked. us, in the presence of a large singular value, the nuclear norm will deviate from the rank of the low-rank matrix, so it cannot accurately approximate the rank, and the degree of inaccuracy is upregulated with the increase in the singular value.

Problem 3.
e situation that the nuclear norm cannot well approximate the rank of the matrix can be considered a problem arising from the compression sensing. Since ℓ 1 -norm is the minimum convex envelope of ℓ 0 -norm [9], conventional compression sensing technology exploits ℓ 1 -norm instead of ℓ 0 -norm to recover the original data. However, under several conditions, Y is both a low-rank matrix and a sparse matrix, while E denotes a sparse matrix exhibiting low-rank properties. It was demonstrated that ℓ 1 -norm is not an exact approximation of ℓ 0 -norm [10,11], and in most cases, this approximation bias is not negligible.
Over the past few years, to improve the approximation rank of nuclear norm, numerous scholars have been committed to develop nonconvex rank approximation functions [12][13][14] to remedy the defect of nuclear norm convex rank approximation functions, and they have been extensively employed. Considerable experiments proved that the nonconvex rank approximation can more accurately and fastly recover the low-rank part of the data than the nuclear norm. For instance, Sun et al. [11] proposed a nonconvex model of robust principal component analysis with capped trace norm and capped ℓ 1 -norm; Kang et al. [15] built a more accurate nonconvex rank approximation function, termed as log-determinant function. In literature [16], Kang et al. defined a novel norm as the nonconvex rank approximation, which is termed as c-norm, and the experimental results revealed that it is a better rank approximation than the nuclear norm. Xie et al. [17] built a nonconvex regularization tool, termed as weighted Schatten p-norm, which acted as the rank approximation to achieve a better approximation. Lu et al. [12] introduced a class of nonconvex approximation functions of ℓ 0 -norm to the singular values of low-rank matrices and built a nonconvex low-rank minimization problem with the nonconvex approximation functions thus obtained. Yu et al. [18] achieved a threshold representation as a recursive function for general ℓ p regularization problems; subsequently, they designed a filtering algorithm suitable for image reconstruction. Xu et al. [19] introduced an algorithm of background subtraction based on low rank and structured sparse decomposition. e algorithm process fully complied with Gao's algorithm [20], block-sparse RPCA. e difference between algorithms of Gao and Liu was that the section of first-pass RPCA introduced structure sparse norm considering the foreground structural continuity. It utilized the adaptive regularization parameter based on the significant adjustment of motion detection method to do group-sparsity operation, and thereby better results were obtained. Liu et al. [21] proposed an efficient numerical solution, Grassmannian Online Subspace Updates with Structured-Sparsity (GOSUS), for online subspace learning in the context of sequential observations involving structured perturbations. Xu modelled both homogeneous perturbations of the subspace and structural contiguities of outliers by Grassmannian fluid, and after certain manipulations, they solved via an alternating direction method of multipliers. Javed et al. [22] introduced a spatiotemporal low-rank modelling method to overcome occlusions by foreground objects and redundancy in video data. e proposed method encodes spatiotemporal constraints by regularizing spectral graphs and uses optical flow information to remove redundant data and to create a set of dynamic frames from the input video sequence. rough their experiments, it can be confirmed that the algorithm has a high efficiency in processing background reconstruction. However, the 11 parameters required by the two models proposed by them need to be optimized, which greatly affects the robustness of the algorithm.
Since the rank approximation of the low-rank matrix refers to the critical problem in the low-rank matrix recovery, how to build a more accurate approximation function should be studied [12,13,23], and on that basis, the corresponding low-rank matrix recovery model is built, which is of great practical significance to solving image processing problems (e.g., image reconstruction, image denoising, and image recognition).
Based on the mentioned inspirations, starting from the construction of nonconvex rank approximation function of low-rank matrix, the present study built a low-rank matrix recovery model suitable for nonconvex approximation in different image processing problems and then designed the corresponding algorithm to obtain more accurate low-rank matrix part from the original data, as an attempt to solve the specific problems (e.g., image reconstruction and image denoising).

Relevant Work
e so-called low-rank matrix is presented as follows: assuming X � (x ij ) m×n ∈ R m×n , its rank is r, and r ≪ min m, n { }. Subsequently, X is the low rank. Each row and column of low-rank matrix covers considerable redundant information, which can be exploited for image recovery and image feature extraction.

Evolution of Low-Rank Matrix Recovery Model.
e present study involved the relevant application of low-rank representation; its model evolution and solving algorithm were first presented. Starting from the simplest case, assuming that the data were clean, i.e., the data were not polluted or not covered with any noise and outliers, considering the following rank minimization problems where P is termed as the dictionary matrix and can also be considered the basis matrix of the data space; Y denotes the low-rank matrix; and the lowest rank of the data matrix was expressed by solving equation (5). e general solution of equation (5) is to replace the rank with the nuclear norm, which is transformed into the following problem: Liu et al. proved the properties of the solution of equation (6) in literature [24].

Optimal Solution Is
Unique. Since the kernel norm is not a strictly convex function, equation (6) may achieve multiple optimal solutions, whereas the theorem below can still prove that equation (6) achieves a unique optimal solution.
Theorem 1 (see [24]). Assuming P ≠ 0 and X � PY shows a feasible solution, i.e., X ∈ span(P), then Y * � P † X denotes the unique solution to equation (6), where P † expresses the generalized inverse of P.

Optimal Solution Is Block Diagonal.
In the selection of a right dictionary, the lowest rank representation will reveal the true segmentation result. In other words, when the columns of P and X are indeed from independent subspaces, the optimal solution of equation (6) can reveal the subspace relationship between sample points. e mentioned conclusion can be expressed by the theorem below. Theorem 2. (see [24]). Set S 1 , S 2 , . . . , S k as the set of k subspaces, and the rank (or dimension) of S i is r i , P � [P 1 , P 2 , . . . , P k ] and X � [X 1 , X 2 , . . . , X k ]. Assuming that P i denotes the set of m i sample points of subspace S i , X i represents the set of n i sample points of subspace S i , and P i satisfies rank(P i ) � r i . If the subspace S 1 , S 2 , . . . , S k is independent, the optimal solution Y * of problem (6) is block diagonal:

Mathematical Problems in Engineering
Clean data building is mentioned above. Besides, a general situation is given below. For data with noise, the following rank minimization model should be built: where η denotes a positive compromise factor, i.e., the equilibrium parameter. ‖·‖ ℓ represents ℓ-norm, where ℓ denotes a parameter of norm, i.e., ℓ can be taken as (1,2), (2,1) etc., and E denotes the noise matrix. Obviously, when P � I (I denotes the unit matrix), equation (9) is converted into the form of equation (4). Accordingly, the low-rank representation can be considered as the generalization of robust principal component analysis with orthonormal basis as a dictionary. By taking a proper basis, the data from different subspaces can be extracted with low-rank representation, thereby remedying the defect that the robust principal component analysis can only process the data from the identical subspace. For equation (9), the kernel norm acts as the convex approximation of the rank function; since it is more robust to noises and outliers, it is converted into the following model: Equation (10) can be solved by the augmented Lagrange multiplier method. For the separation of variables, a novel variable matrix J is introduced, and constraints Y � J are added. Equation (10) has the following equivalent form: e augmented Lagrange function of equation (11) is Equation (11) can be solved with precise and imprecise Lagrange multiplier methods. e convergence of the former has been verified in literature [25]. For the latter, convergence is not ensured for cases with more than two pieces for Y. e primary steps to solve equation (12) with the precise Lagrange multiplier method are presented below: Step 1: with singular value thresholding operator [26] (SVT operator), the optimal solution J k+1 of the (k + 1)th iteration is solved by Step 2: the optimal solution Y k+1 of the (k + 1)-th iteration is solved: Step 3: based on the conclusion in literature [27], the optimal solution E k+1 of the (k + 1)-th iteration is solved by using the following formula: Step 4: the multiplier matrices Λ 1 and Λ 2 and parameter ρ are updated: where β and ρ max denote the predetermined positive numbers; generally, β � 1.1 and ρ max � 106. According to equations (13)- (16), each variable is alternately updated until the termination condition of the algorithm is reached. For instance, a previously selected small positive number ε (generally, ε � 10 − 7 , 10 − 8 ) can be taken in the following form [24]:

Principle of Constructing Nonconvex Rank Approximation
Function. Since the kernel norm is the minimum convex envelope of the rank function, the conventional low-rank matrix recovery model employs the kernel norm as the rank approximation function of the low-rank matrix; as a result, the problem is transformed into a convex optimization problem, e.g., equations (4), (6), and (10). However, the kernel norm approximation rank function has some defects, so the conventional low-rank matrix recovery model should be optimized. Since ℓ 1 -norm is only loose approximation of ℓ 0 -norm, only the suboptimal solution of the original problem can be obtained. us, a function better than the ℓ 1 -norm should be developed to replace the ℓ 0 -norm. Under this context, scholars proposed numerous nonconvex approximations of the ℓ 0 -norm, and some nonconvex approximate functions are presented in literature [12] (e.g., ℓ p -norm [18] (0 < p < 1), Geman and Young [28], smoothly clipped absolute deviation (SCAD) [29], minimax concave penalty (MCP) [30], Laplace [31], capped ℓ 1 [32], exponential-type penalty (ETP) [33], and logarithmic function [34]). e specific function forms are listed in Table 1.
e images of the eight nonconvex approximate functions are shown in Figure 1.
As shown in Figure 1, the graphs of the mentioned functions show a similar trend, which also inspires us to develop novel nonconvex approximate functions. Moreover, by combining with the problem of approximate rank of nuclear norm, to develop the approximate function of the rank function more accurately than the nuclear norm, our principle aims to satisfy the following properties: (1) e contribution of a singular value complying with a large value σ the rank approximation function should be bounded; its contribution should be significantly reduced compared with the nuclear norm, i.e., it satisfies where f(σ) denotes the function of singular value σ, representing the contribution of the singular value, and σ is a positive constant.

A Novel Low-Rank Matrix Recovery Model with Nonconvex Rank Approximation.
According to the definition of nuclear norm, under a too large value of a certain σ i (X), the rank estimation will be excessively large, affecting the approximation effect. e present section will give a novel nonconvex approximation function to approximate the rank function of the matrix. Lu et al. [35] summarized the existing nonconvex approximation model and developed the following general form model: where f denotes a monotonically increasing continuous concave function defined in [0, ∞). e gradient function of g ∇g Lipschitz is continuous.
denotes a certain form of nonconvex approximation function. Subsequently, the rank approximation function in equations (3) and (5) can be approximately replaced with F(Y). us, a model based on nonconvex rank approximation can be developed.
For the robust principal component analysis model, the nonconvex rank approximation model of equation (3) is expressed as where Y, E ∈ R m×n ; ‖E‖ 0 denotes the ℓ 0 -norm of matrix E; and η is the positive compromise factor. A general low-rank matrix recovery model based on nonconvex rank approximation is mentioned above. For instance, equation (4) indicates a model with sparse noise only in the original data, while equation (5) builds the model with no noise in the original data. For the low-rank matrix recovery model under other noises, the rank function of the low-rank matrix in the model can also be replaced by several appropriate nonconvex rank approximation functions.
In the nonconvex rank approximation model, by adopting nonconvex rank approximation function F(Y), exhibiting higher accuracy than the nuclear norm, the use of the nonconvex rank approximation model can more effectively restore low-rank matrix part in the original data, which can enhance the application effect of the original low-rank matrix recovery model in image processing (e.g., image reconstruction, image denoising, and image recognition).

Name of alternate function
Expression of alternate function f(x), x ≥ 0, μ > 0 ℓ p [18] μx p Geman and Young [28] (μx/(x + c)) SCAD [29] μx, [34] (μ/log(c + 1))log(cx + 1) Mathematical Problems in Engineering 5  Subsequently, a novel low-rank matrix recovery model with nonconvex rank approximation was proposed, and the rank function of low-rank matrix in the existing model was substituted with the nonconvex rank approximation function. Next, a correlation algorithm was developed to address the problem.

Nonconvex Rank Approximation Function.
Consistent with the approximation function of ℓ 0 -norm given by Geman and Young [28], a novel nonconvex rank approximation function was developed in the present study: where μ denotes the equilibrium factor, and μ > 0; c represents the convergence rate parameter of f(x), and c > 0. Next, it will be verified that f(x) satisfies the three properties in the previous section: (1) σ i , (i � 1, 2, · · · , l) is assumed as the singular value of matrix Y, and Y ∈ R m×n , and 0 ≤ σ 1 ≤ σ 2 ≤ · · · ≤ σ l : In other words, the contribution of the singular value exhibiting a large value to the rank approximation function should be bounded.
at is, under the zero singular value σ, the contribution is zero.
f(x) satisfies the property of nonconvex approximate rank function. Based on the function in equation (20), the following New-norm can be obtained.
where σ i (Y) denotes the i-th singular value of matrix Y. When parameter c takes different values, the true rank of the matrix is approximated by the nonconvex function derived from Definition 1, as presented in Figure 2.
In Figure 2, given the rank and the rank approximation function, the horizontal coordinate x denotes the singular value, while the true rank is 1 under the nonzero singular  e curve indicates the nonconvex approximation function f(x). With the increase in the singular value x, its function value will not surge, and it is relatively consistent with the real rank. e difference from the rank can indicate the approximation effect. With the increase in the singular value x, the function value of f(x) is close to 1, which differs slightly from the real rank. Accordingly, as a rank approximation function, its approximation effect is significantly high. Under the relatively small value of c, for instance, c � 0.1 or c � 0.5, a relatively large difference is identified between the curve obtained and the real rank. However, with the increase in the value of c, the obtained curve approximates the real rank more efficiently. For instance, when c � 2, the curve's approximation effect has been prominent.
Moreover, Figure 2 presents the simplification in the presence of only one singular value. e nuclear norm complies with a line starting from the origin with a slope of 1. Since the nuclear norm is the sum of all singular values, under only one singular value, the nuclear norm is exactly the singular value. Furthermore, the figure indicates the obvious defect of approximate rank of nuclear norm. e greater the singular value, the greater the degree of deviation of nuclear norm from true rank. us, when the nuclear norm is adopted to approximate the rank, even if there are only a few large singular values, the deviation of approximation should not be overlooked.
Next, this study proves theoretically and tests equation (24) as a more accurate nonconvex rank approximation function based on experimental results. It is known that the mentioned norm equation (24) does not satisfy the direct proportionality of the matrix norm (for arbitrary constant c, ‖cY‖ � |c|‖Y‖), whereas it is only a virtual norm, or a pseudonorm. However, it can also be verified that they satisfy the properties below.

Proof
(1) For the following is obtained: In other words, when c ⟶ 0, the function value f(σ i (Y)) is 1 for the nonzero singular value, and 0 for the singular value of zero. On the contrary, given the rank of the matrix rank(Y), it is also the case that each singular value with the value of nonzero upregulates the rank by 1, while each zero singular value does not increase the rank of the matrix. ‖Y‖ N � rank(Y) is proved to be true.
(2) It is assumed that (·) H , (·) − 1 , and λ, respectively, denote the conjugate transpose, inverse, and eigenvalues of the matrix and I represents the unit matrix of the appropriate order. Given the characteristic polynomial, Based on the derivation, it is suggested that the matrices (UYV) H UYV and Y H Y show the identical characteristic polynomial, so they achieve the same eigenvalue. Subsequently, the singular value of UYV and Y refers to the nonnegative square root of the eigenvalues (UYV) H UYV and Y H Y, respectively, and the singular value of UYV is equal to that of Y.  Mathematical Problems in Engineering us, ‖Y‖ N � ‖UYV‖ N is true, and ‖Y‖ N is unitarily invariant.

RPCA of New Nonconvex Rank Approximation.
As impacted by the limitations of PCA and RPCA, in the present study, a novel nonconvex rank approximation robust principal component analysis (NNCRPCA) was developed, capable of effectively dealing with the overall corrosion and exhibiting a good generalization ability. For any novel data damaged by gross error, the learning projection can effectively eliminate the possible damage by projecting the data into its low-rank subspace. As revealed from the results of considerable experiments performed on three static continuous image datasets and one dynamic video dataset, NNCRPCA is robust to serious corruption and can effectively process novel data.
In the low-rank matrix recovery model, the rank function rank(Y) is approximately replaced by New-norm of Y to obtain the following model: where X, Y, E ∈ R m×n , and η > 0 is a positive compromise factor. For equation (31), we can change it into an unconstrained optimization form by convex relaxation: where η 1 and η 2 are nonnegative parameters used to balance data consistency. For the unconstrained optimization problem (32), we can use the alternating direction method to solve large-scale optimization problems that are commonly encountered and dealt with in the field of machine learning.
Using the alternating direction method, in the (k + 1)-th iteration, equation (32) requires solving the following two subproblems: To solve equation (33), first of all, its objective function is the combination of nonconvex functions and convex functions is needed to be noticed: the first term ‖Y‖ N is the nonconvex function, and the second term ‖X − PY − E k ‖ 2 F is the convex function. us, the solution can be obtained with the help of the difference of convex function (DC) [36]. e convex function algorithm proposed by Tao in Literature [37] refers to an algorithm based on duality theory and local optimality conditions of convex function optimization, and the convergence proof is presented in Literature [36].

Lemma 1.
e subdifferential of ‖Y‖ N : where the columns of U and V, respectively, denote left and right singular vectors of Y, and the following formula is yielded from equation (25): According to Lemma 1, the (k + 1)-th iteration of Y in equation (33) is written as In equation (37), the partial derivative of the objective function is with respect to Y, and then it is set as zero, so the following formula can be obtained: Equation (38) refers to the iterative formula for solving Y.
Next, the solution of E is to be discussed. Soft thresholding of singular values of E can be adopted iteratively. Accordingly, the (k + 1)-th iteration in equation (34) can be expressed as where X 0 � P H X and matrix X k � Y k + E k . e iterative formula of matrix X is as follows: In equation (40), to keep the data consistency, the novel X k+1 is derived by subtracting the residuals Mathematical Problems in Engineering e mentioned solution process is summarized in Algorithm 1.

Algorithm Convergence.
In the present section, the convergence proof of the algorithm is provided.

Lemma 2.
Set Y k and E k as the sequence generated by the algorithm, then Y k and E k are bounded.
Proof. First, the objective function of equation (32) is denoted as the following form: Based on equation (33) and (34), us, the sequence Ψ(Y k , E k ) is nonincreasing. Next, it is verified that the sequences Y k and E k are bounded, and the boundedness of Y k is subsequently proved by the proof by contradiction.
Assuming Y k is unbounded, then for any real number M > 0, there is always a positive integer N that satisfies ‖Y k ‖ ≥ M for k ≥ N and any norm ‖·‖, i.e., ‖Y k ‖ ⟶ ∞. Accordingly, the first term of equation (41) is As the second term ‖Y‖ N in equation (41) is a pseudonorm, next, we will discuss the boundedness of this term under the following two conditions: according to equation (24), the following formula can be obtained: In a review of the nonconvex rank approximation, to better approximate the rank function, the value of c approaches ∞, so it is suggested that it is bounded in this case.
according to equation (24), it is obvious that, in this case, ‖Y‖ N is also bounded.
In conclusion, ‖Y‖ N is always bounded. Consistent with the proof method in Y k , assuming E k is unbounded, the first and third terms in equation (41) are then unbounded as follows: In brief, both Y k and E k are unbounded, so Ψ(Y k , E k ) is unbounded. Combined with its nonnegative property, it is suggested that Ψ(Y k , E k ) ⟶ ∞, which is not consistent with the conclusion in equation (43) that the sequence Ψ(Y k , E k ) is nonincremental. Accordingly, the sequences Y k and E k are bounded.
Based on the mentioned lemma, the below theorem can be developed. □ Theorem 3. Assuming that Y k and E k are the sequences generated by the algorithm, then any clustering point of sequence (Y k , E k ) is a local minimum point of equation (32).
Proof. ased on Lemma 2, it can be known that (Y k , E k ) is bounded. So (Y k , E k ) has at least one gathering point, i.e., (Y * , E * ). Combining with the conclusion that Ψ(Y k , E k ) is a nonincremental sequence in equation (43), it can be known that there is lim k⟶∞ Ψ(Y k , E k ), and the following formula is true:
subsequently, it was compared with existing GoDec [38], ALM, RegL1-ALM [24], NCRPCA, GOSUS [19], and SLM [22] methods. Video background separation, a process to remove foreground and separate background from video shot with a fixed camera, has been extensively employed in computer moving target detection. In the present study, the experimental operating environment is Intel(R) Core ™ i7-8565U CPU @1.80ghz 2.00ghz and memory 16.00 GB. All algorithms are written by MATLAB R2018b.

Parameter
Analysis. e novel algorithm was tested and analysed on Wallflower and SBMnet, respectively. By referencing the stop criterion of the four comparison algorithms, the relative error value t of this algorithm in the Wallflower background separation experiment reaches 2.5 × 10 − 3 and 4 × 10 − 4 in the SBMnet background separation experiment, which is expressed as For Wallflower, the effect of the parameter values on the experimental effect was analysed. To be specific, a common quantitative index, i.e., root-mean-square error (RMSE), was adopted as the measurement standard, which is defined as follows: where X, Y, and E, respectively, denote the original observation data matrix, low-rank matrix, and sparse matrix. e value of P encoding operator refers to the unit matrix. Next, with Wallflower as an example, the relationship between the value of parameter c and RMSE is illustrated. e initial value c � 0.1, step size � 0.01, and then stops when c � 10.
(1) When c ∈ [4,10], the value of RMSE approaches 1, suggesting that the algorithm cannot recover lowrank matrix Y from data X at this time, i.e., the experimental effect is not ideal (2) When c ∈ [4,10], the value of RMSE maintains stably between 0.5 and 0.9, and meantime, the experimental effect of the algorithm is ideal Accordingly, a conclusion was also drawn that the larger the c, the better the approximation degree of the approximation rank, whereas the image reconstruction effect of the algorithm in the actual implementation is unsatisfactory. Among the experiments on Wallflower, when the RMSE value is the smallest, and the rank is approximately the optimal, c � 2. e statistical information of Wallflower exploited in the experiment is listed in Table 2 (e.g., image dimension, image frame number, and dimension of stretch matrix). e statistical information of SBMnet exploited in the experiment is listed in Table 3 (e.g., image dimension, image frame number, and dimension of stretch matrix).

Experimental Results.
Besides the experimental renderings, the quantitative indexes to test the algorithm consist of calculation time, iteration times, and other common indexes. Moreover, since the premise of the low-rank matrix recovery model adopted for image reconstruction is that the matrix corresponded by the background part is low rank, the rank of matrix Y recovered by the algorithm should be maximally small, and the most ideal result is that the rank is 1. By equation (33), the RMSE was calculated after applying various algorithms to Wallflower and SBMnet, as an attempt to indicate the background separation effect. Besides the mentioned indexes, the RE is another vital index to reflect the accuracy of the algorithm.
Let us start with Wallflower first. Many challenges are placed in the dataset to evaluate the performance of the algorithms. e Bootstrap refers to a scene of taking food from a restaurant. e characters in the scene are constantly updated with frequent activities, and there exist considerable characters. e Camouflage refers to a scene of electronic computer screen. When recording the Camouflage dataset, the computer screen refreshes at a certain frequency, and someone flashes by in front of the computer screen, so the background is difficult to separate. In LightSwitch, both the electronic computer refreshing, and the behaviour process of characters are identified; meantime, the transformation of lights is added in the video shooting process. e Waving Trees refers to a scene in which a man passes through the shaking trees. e Foreground Aperture refers to a scene of one person sleeping on the table in a low-light environment.
e Move Object refers to a scene in which a person enters the classroom, picks up the phone to talk, and leaves. e Time of Day refers to a scene in which a living room passes from night to day and then from day to night. We used many algorithms (i. e. GoDec, ALM, RegL1-ALM, NCRPCA, GOSUS, SLM, and NNCRPCA) to do background separation experiments on these datasets and evaluated their performance. Table 4 lists the algorithms' experimental quantitative index for Wallflower, and Figure 3 illustrates the algorithms' rendering for the background separation experiment. Table 4 indicates that the performance of the algorithms mentioned above is quite different in the background separation experiment. e earlier algorithms, ALM and RegL1-ALM, have higher iteration times and more running time, but they own better relative error and RMSE than   NCRPCA is inferior to SLM in execution time and iteration. Although NNCRPCA is worse than SLM in execution time, it is better than SLM in other aspects. RMSE and relative error are kept with a good level; especially the rank of tensor matrix of dataset is kept at a low level, which is lower than other algorithms. Figure 3 shows the processing effect of background separation by different algorithms, and it is suggested that the background calculated by the NNCRPCA algorithm is clearer.
In real life, influenced by the variations in quantity, lighting, weather, and other conditions, video background changes constantly. e rank of the background matrix reaches over 1, and the separation process is relatively complex, which is beyond the reach of ordinary algorithms. In the following section, the approximate model proposed in the present study was applied in SBMnet, and the feasibility of the model was verified by numerical experiments.
In the SBMnet dataset, many challenging scenarios which were illustrated in Table 3 were set up and imposed great challenges on the calculation of the algorithm. rough many experiments, we can observe how the backgrounds of scenarios could be separated by the above algorithms and the performance of the algorithms verified again. e algorithms' background separation experimental quantitative index for SBMnet is listed in Table 5, and Figure 4 shows the algorithm's rendering for the background separation. e Toscana dataset has only 6 frames, which makes GOSUS unable to run. According to Table 5, the performance of NNCRPCA is better than other algorithms. In terms of relative error and RMSE, the results of NNCRPCA are lower than or close to those of SLM, but lower than those of other algorithms. NNCRPCA is still lower in rank approximation than other algorithms, including SLM, but its CPU time is still inferior to some algorithms. Figure 4 shows the processing effect of background separation by different algorithms, and it is suggested that the background calculated by the NNCRPCA algorithm is clearer.
Tables 4 and 5 suggests that the NNCRPCA algorithm achieves prominent experimental results for RE and RMSE in the background separation application. e NNCRPCA algorithm modifies the existing convex approximation that employs nuclear norm as the low-rank matrix and adopts a more accurate approximation method of nonconvex rank approximation function. In the application of background separation, its effect is higher than that of nuclear norm, and it is significantly improved in time and rank. Besides, the improvement of the RMSE and relative error is also the result of adopting a better rank approximation. A low rank is achieved by the NNCRPCA algorithm, demonstrating that the low-rank matrix to extract the background part of the image is more accurate, so the sparse part extraction of the experiment is more complete. As revealed from Figures 3 and  4, the low-rank part extracted by our algorithm is clearer, and the information of the sparse part is more complete.

Conclusions
A novel algorithm for robust principal component analysis of nonconvex approximation rank function was proposed. Unlike the previous models, a novel nonconvex approximation rank function acted as the rank approximation of the low-rank matrix. Subsequently, the robust principal component analysis based on this nonconvex rank approximation function was established, and the NNCRPCA algorithm based on the alternating direction method was developed. Next, the numerical experiments were performed on Wallflower as well as SBMnet. From the perspective of image effect, as well as the main quantitative indexes (e.g., running time, iteration times, RMSE, and RE), the algorithm effectively improved the image background separation effect. Our major contributions are revealed in the following three aspects: (1) A novel nonconvex rank approximation function was constructed. Different from the conventional methods in which the nuclear norm acted as convex approximation, the new rank approximation function approximated the rank of the low-rank matrix very well, the construction and solution of the novel algorithm for robust principal component analysis were achieved, and the convergence of the algorithm was proved theoretically. (2) It is combined with the unconstrained problem of alternating direction method and convex function after solving the differential of convex relaxation.
Since the objective function of the optimization problem to be solved is the combination of convex function and nonconvex function, the differential of convex function refers to an effective technique to solve this type of optimization problems. (3) e low-rank matrix recovery model was applied for background separation, and the experimental results assessed the application prospect of applying the nonconvex approximation method in Mathematical Problems in Engineering 15 this problem. Moreover, in terms of data processing mode, the background separation done in the present study requires further studies. With the continuous increase in the data size in the video dataset, the model and algorithm proposed in this study will exhibit a broader application prospect.

Data Availability
e data used to support the findings of the study are available from the corresponding author upon request.

Conflicts of Interest
e authors declare no potential conflicts of interest with respect to the research, authorship, and/or publication of this article.