Next Article in Journal
Peculiarities of Assessing Body Strength When Converting a Bus from Diesel to Electric Traction Following the UNECE R100 Regulation
Previous Article in Journal
PD Detection and Analysis Triggered by Metal Protrusion in GIS Through Various Methods
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Accelerated Tensor Robust Principal Component Analysis via Factorized Tensor Norm Minimization

by
Geunseop Lee
Division of Global Business and Technology, Hankuk University of Foreign Studies, Yongin 17035, Republic of Korea
Appl. Sci. 2025, 15(14), 8114; https://doi.org/10.3390/app15148114
Submission received: 18 June 2025 / Revised: 18 July 2025 / Accepted: 18 July 2025 / Published: 21 July 2025

Abstract

In this paper, we aim to develop an efficient algorithm for the solving Tensor Robust Principal Component Analysis (TRPCA) problem, which focuses on obtaining a low-rank approximation of a tensor by separating sparse and impulse noise. A common approach is to minimize the convex surrogate of the tensor rank by shrinking its singular values. Due to the existence of various definitions of tensor ranks and their corresponding convex surrogates, numerous studies have explored optimal solutions under different formulations. However, many of these approaches suffer from computational inefficiency primarily due to the repeated use of tensor singular value decomposition in each iteration. To address this issue, we propose a novel TRPCA algorithm that introduces a new convex relaxation for the tensor norm and computes low-rank approximation more efficiently. Specifically, we adopt the tensor average rank and tensor nuclear norm, and further relax the tensor nuclear norm into a sum of the tensor Frobenius norms of the factor tensors. By alternating updates of the truncated factor tensors, our algorithm achieves efficient use of computational resources. Experimental results demonstrate that our algorithm achieves significantly faster performance than existing reference methods known for efficient computation while maintaining high accuracy in recovering low-rank tensors for applications such as color image recovery and background subtraction.

1. Introduction

Robust principal component analysis (RPCA) has been extensively studied to address the issue of extreme outliers contained in data, which causes significant degradation of the performance when applying classical principal component analysis [1]. Specifically, RPCA aims to separate outliers from the true data by reconstructing a low-rank matrix L R m × n from an observed matrix X, which is contaminated by sparse noise S of arbitrary magnitude, such that
X = L + S .
This RPCA model arises in numerous real-world applications, including image denoising, video background subtraction, subspace clustering, or feature selection in bioinformatics [2,3,4,5,6]. However, identifying the locations of sparse noise, i.e., the nonzero entries of S, is often challenging. Under the assumptions that L is low-rank and S is sufficiently sparse, it has been shown that one can accurately recover L by solving the following convex problem:
min L * + λ S 1 , s . t . X = L + S ,
where · * denotes the nuclear norm, · 1 denotes the l 1 norm, and  λ is a regularization parameter [1]. A typical algorithm for solving the problem (2) is to apply singular value thresholding, which modifies the singular values of a matrix by applying a threshold operator [7]. However, as the data structure become increasingly complex, the matrix representation may no longer be sufficient for capturing the intricacies of real-world data. Since tensors can represent higher-order data while preserving spatio-temporal and multi-dimensional correlations, it is natural to extend RPCA to tensor setting. This motivates the extension of the RPCA model in (2) to the tensor RPCA (TRPCA) formulation, where the input is an n-th order tensor X R I 1 × × I n . The optimization problem of TRPCA is formulated as follows:
min | | L | | * + λ | | S | | 1 , s . t . X = L + S ,
where L denotes a low-rank tensor, S is a sparse tensor, and  λ is a regularization parameter. However, this extension introduces new challenges, particularly due to the lack of a unified definition of tensor rank and tensor nuclear norm, unlike in the matrix case. For example, under Canonical Polyadic Decomposition (CPD), the tensor rank is defined as the minimum number of rank-1 tensors whose sum reconstructs the given tensor [8]. However the rank minimization problem with CP rank is generally NP-hard [9]. Another widely used definition is the Tucker rank, which consists of a tuple of ranks obtained by unfolding the tensor along each mode. Its convex surrogate can be formulated by summing the nuclear norms of the mode-wide unfoldings, leveraging ideas from matrix convex optimization [10]. More recently, tensor singular value decomposition (T-SVD) and tensor–tensor product (T-product) frameworks introduced by Kilmer et al. [11] led to definitions such as tensor multi-rank and tensor tubal rank. These ranks provide a more structured and compact representation of tensors and avoid the loss of structural information that typically occurs during tensor matricization. They are particularly useful for preserving the inherent low-rank structure in tensor data. Nevertheless, it should be noted that computing the T-SVD-based tensor ranks can be computationally demanding as the size of the tensor increases.
In this paper, we propose a novel and computationally efficient TRPCA algorithm based on tensor average rank minimization. Specifically, our main contribution is the further relaxation of the tensor nuclear norm—a convex envelope of the tensor average norm defined in (3)—into an optimization problem involving the sum of Frobenius norms of factor tensors. By alternatingly updating the truncated factor tensors from a given tensor, our method significantly improves computational efficiency while maintaining competitive reconstruction accuracy on real-world data compared to existing TRPCA algorithms. The remainder of this paper is organized as follows. In Section 2 we define the notations and preliminaries used throughout this paper. Section 3 introduces well-known TRPCA methods. In Section 4, we present our proposed TRPCA algorithm based on tensor average rank. Section 5 provides experimental results on real-world images and video sequences. Finally, Section 6 concludes this paper.

2. Notations and Preliminaries

We first summarize the symbols and terminologies consistently used throughout this paper. Capital calligraphy letters, e.g.,  A , and capital letters, e.g., A, are used to denote tensors and matrices, respectively. Boldface lowercase letters, e.g.,  a , and lowercase letters a represent vectors and scalars, respectively. For a third-order tensor A , we adopt MATLAB-style notation: A ( i , : , : ) , A ( : , i , : ) , and  A ( : , : , i ) denote the i-th horizontal, lateral, and frontal slices of A , respectively. For convenience, we abbreviate the i-th frontal slice A ( : , : , i ) as A ( i ) . The MATLAB functions A ¯ = fft ( A , [ ] , 3 ) and A = ifft ( A ¯ , [ ] , 3 ) are used to compute the Discrete Fourier Transform (DFT) and its inverse along the third dimension of A , respectively, yielding A ¯ and reconstructing A . Additionally, we define the function A ¯ = bdiag ( A ¯ ) as
A ¯ = bdiag ( A ¯ ) = A ¯ ( 1 ) A ¯ ( 2 ) A ¯ ( n 3 ) ,
which rearranges the frontal slices of the tensor A ¯ R n 1 × n 2 × n 3 to the block diagonal matrix A ¯ R n 1 n 3 × n 2 n 3 . The function bcirc ( · ) reconstructs the block circulant matrix B R n 1 n 3 × n 2 n 3 from the tensor B R n 1 × n 2 × n 3 as follows:
bcirc ( B ) = B ( 1 ) B ( n 3 ) B ( 2 ) B ( 2 ) B ( 1 ) B ( 3 ) B ( n 3 ) B ( n 3 1 ) B ( 1 ) .
A function D = unfold ( D ) transforms the tensor D R n 1 × n 2 × n 3 into the matrix D R n 1 n 3 × n 2 , where D = [ D ( 1 ) ; D ( 2 ) ; ; D ( n 3 ) ] . Conversely, the function D = fold ( D ) transforms the matrix D to the tensor D . These functions satisfy the identity fold ( unfold ( D ) ) = D . Based on these notations, we now define several important concepts of tensor algebra used in this paper.
Definition 1
((T-product) [11]). Let A R n 1 × n 2 × n 3 and B R n 2 × n × n 3 be two third-order tensors. The T-product A B produces a tensor of size n 1 × n × n 3 , defined as
A B = fold ( bcir ( A ) · unfold ( B ) ) .
Additionally, the result of the T-product defined in (6) is equivalent to matrix multiplication in the Fourier domain. Therefore, it can be computed efficiently as
A B = ifft ( fold ( bdiag ( A ¯ ) , unfold ( B ¯ ) , [ ] , 3 ) ,
where A ¯ = fft ( A , [ ] , 3 ) and B ¯ = fft ( B , [ ] , 3 ) .
Definition 2
((Identity tensor) [11]). The identity tensor I R n × n × n 3 is defined such that I ( 1 ) is the identity matrix, and  I ( i ) = 0 for all 2 i n 3 .
Definition 3
((Orthogonal tensor) [11]). A tensor P R n × n × n 3 is said to be orthogonal if it satisfies
P T * P = P * P T = I .
Definition 4
((Inverse tensor)). Assume that all frontal slices of a Tensor A R n × n × n 3 are invertible. Then, a tensor B R n × n × n 3 is called the inverse of A if A B = I .
Theorem 1.
Let A R n 1 × n 2 × n 3 . Then it can be factorized as
A = U * C * V T ,
where U R n 1 × n 1 × n 3 and V R n 2 × n 2 × n 3 are orthogonal tensors and  C R n 1 × n 2 × n 3 is an f-diagonal tensor, meaning that each frontal slice of C contains the singular values of the corresponding frontal slice of A [11]. This factorization is known as T-SVD. To obtain the best low-rank approximation of A , we use truncated tensor singular value decomposition (TT-SVD), which truncates the factor tensors as U R n 1 × r × n 3 , C R r × r × n 3 , and  V R n 2 × r × n 3 , where r min ( n 1 , n 2 ) [12].
Definition 5
((Tensor average rank) [13]). Given a tensor A R n 1 × n 2 × n 3 , the tensor average rank is defined as
rank ( A ) = 1 n 3 rank ( bcirc ( A ) .
Definition 6
((Tensor nuclear norm) [14]). A tensor nuclear norm of a tensor A R n 1 × n 2 × n 3 is defined as
| | A | | * = 1 n 3 bcirc ( A ) * = 1 n 3 i = 1 n 3 | | A ¯ ( i ) | | * .

3. Related Works

Because typical computation of TRPCA requires substantial computational resources, especially dealing with large-scale and high-order tensor data, numerous studies have focused on enhancing the computational efficiency of TRPCA using various definitions of tensor rank and tensor nuclear norm. Lu et al. defined the tensor average rank defined in Definition 5 and showed that their proposed tensor nuclear norm serves a convex surrogate for the tensor average rank minimization problem [13]. By applying the ADMM technique, Lu et al. solved the TRPCA problem and demonstrated the effectiveness of their method through applications in image denoising and background modeling. Gao et al. introduced a weighted tensor Schatten p-norm minimization approach to explicitly account for the disparity among singular values [15]. Dong et al. formulated the TRPCA model using Tucker decomposition [16]. To reduce the computational burden, they proposed a scaled gradient descent method that initializes iterations by directly recovering low-dimensional tensor factors. Qiu et al. proposed the alternative projection algorithm, which applies truncated T-SVD to compute the low-rank tensor efficiently [17]. Additionally to speed up computation of TRPCA, Qiu et al. utilized the property of the tangent space of low rank. Geng et al. introduced the tensor adjustable logarithmic norm, which relaxes the tensor nuclear norm as follows:
| | A | | log = 1 n 3 i = 1 n 3 j = 1 r g ( σ j ( A ¯ ( i ) ) ) ,
where g ( x ) = l o g ( θ x + 1 ) is a nonconvex function with an adjustable positive parameter θ [18]. Qiu et al. proposed a fast TRPCA algorithm using the tensor train norm (TTN), approximated through compressed Tucker decomposition, where the TTN is defined as
| | A | | T T N = i = 1 n 3 | | A i | | * ,
and A i denotes the mode-i unfolding matrix of A [19]. Cai et al. employed fiber CUR decomposition to significantly reduce the computational complexity by approximating a tensor using a small subset of its fibers [20]. For scenarios involving frequent data updates, Salut and Anderson proposed the incremental T-SVD approach, which efficiently updates TRPCA solutions as new data arrives [21]. Since our proposed algorithm also aims for computational efficiency of TRPCA, several of the aforementioned methods will be used as references for performance comparison in Section 5.

4. Proposed Algorithm

In this section, we introduce an accelerated TRPCA method based on a factorized low-rank tensor representation, which modifies the original TRPCA model in (3). Specifically in the proposed method, factorized tensors are used to further relax the complex tensor norm minimization problem into a sum of tensor Frobenius norms. This reformulation eliminates the need to compute T-SVD at each iteration, thereby significantly improving the execution speed for solving (3).

4.1. Optimization Model

Assume that we have a tensor X R n 1 × n 2 × n 3 , which can be expressed as the sum of a low-rank tensor and a sparse tensor such that X = L + S . If the low-rank component L can be factorized as L = A B T , then solving the TRPCA problem defined in (3) is equivalent to finding the solution of the following optimization problem:
arg min A , B 1 2 n 3 A F 2 + B F 2 + λ | | S | | 1 , s . t . X = L + S ,
where A R n 1 × r × n 3 and B R n 2 × r × n 3 . Due to the low-rank property of L , it satisfies that r min ( n 1 , n 2 ) .
Theorem 2.
Let L R n 1 × n 2 × n 3 be factorized as L = A B T . Then, the nuclear norm minimization problem of L can be relaxed to the following optimization problem such that
arg min | | L | | * relax arg min A , B 1 2 n 3 ( A F 2 + B F 2 ) .
Proof. 
By Definition 6, the tensor nuclear norm is defined as
| | L | | * = 1 n 3 i = 1 n 3 | | L ¯ ( i ) | | * .
From [22], if we assume L ¯ ( i ) , 1 i n 3 can be factorized as A ( i ) and B ( i ) such that L ¯ ( i ) = A ( i ) B ( i ) T , then the following relaxation holds:
1 n 3 i = 1 n 3 | | L ¯ ( i ) | | * = 1 n 3 i = 1 n 3 | | A ¯ ( i ) B ¯ ( i ) T | | * relax 1 2 n 3 i = 1 n 3 ( A ¯ ( i ) F 2 + B ¯ ( i ) F )
Since A F = A ¯ F and B F = B ¯ F , this completes the proof of (14).    □
Note that if the rank of the solution obtained from (3) is equal to that of the solution obtained from (13), then the solution of (13) is also a solution to (3) [23]. The augmented Lagrange function corresponding to the optimization problem in (13) is given by
L ( A , B , P , λ ) = 1 2 n 3 ( A F 2 + B F ) + λ | | S | | 1 + < P , A B T + S X > + μ 2 A B T + S X F 2 ,                                                            
where P denotes the augmented Lagrange multiplier and  μ denotes a penalty parameter. Note that the inner product between tensors < A , B > is
< A , B > = 1 n 3 < A ¯ , B ¯ > .
The low-rank tensor L = A B T is then recovered by solving the optimization problem of (17).

4.2. Solution Algorithm

The solution to the optimization problem in (17) can be obtained efficiently using an alternating direction method of multiplier (ADMM)-based algorithm.
(1)
Computation of A .
Finding the optimal A k at iteration k, while keeping the other variables fixed, involves solving the following sub-problem derived from (17) such that
A k + 1 = arg min A k 1 2 n 3 A k F 2 + μ A k B k T Q F 2 ,
where Q = X S k μ 1 P k . By taking the derivative of (19) with respect to A k , and rearranging terms, we obtain the closed-form solution for A k + 1 as follows:
A k + 1 = Q B k ( 1 n 3 I + μ B k T B k ) 1 ,
where I R r × r × n 3 denotes the identity tensor.
(2)
Computation of B .
Similar to (19), we fix A k + 1 , S k and P k , and then formulate the sub-problem for updating B k + 1 such that
B k + 1 = arg min B k 1 2 n 3 B F 2 + μ A k + 1 B k T Q F 2 .
By taking the derivative of (21) with respect to B k , and rearranging terms, we obtain the closed-form solution for B k + 1 such that
B k + 1 = Q A k + 1 ( 1 n 3 I + μ A k + 1 T A k + 1 ) 1 .
(3)
Computation of S .
The optimal sub-problem for updating S k , while keeping the other terms in (17) fixed, is defined as follows:
S k + 1 = arg min S k λ μ | | S | | 1 + S k H F 2 ,
where H = X A k + 1 B k + 1 T μ 1 P k . Since the second term in (23) is convex and differentiable, the closed-form solution of (23) can be obtained using the soft-thresholding operator D τ ( x ) defined as [24]
D τ ( X i 1 , i 2 , i 3 ) = sign ( X i 1 , i 2 , i 3 ) ( | X i 1 , i 2 , i 3 | τ ) if | X i 1 , i 2 , i 3 | > τ , 0 if | X i 1 , i 2 , i 3 | τ ,
where X i 1 , i 2 , i 3 denotes the ( i 1 , i 2 , i 3 ) th element of X . Thus, the update for S k + 1 is obtained by applying soft-thresholding as S k + 1 = D λ / μ ( H ) .
The procedure for finding A k , B k and S k is summarized in Algorithm 1. Before computing A 1 , B 1 , we initialize the factor tensors A 0 and B 0 by performing T-SVD of X with truncation level r, such that X A 0 * B 0 T .
Algorithm 1 X k + 1 = TRPCAAL ( X , r , λ , μ , ϵ )
1:
P 0 = 0
2:
A 0 = U 0 C 0 and B 0 = V 0 C 0 , where [ U 0 , C 0 , V 0 ] = TT - SVD ( X , r )
3:
for k = 1,2,…, do
4:
    Update A k + 1 via (20)
5:
    Update B k + 1 via (22)
6:
    Update S k + 1 via (23)
7:
     L k + 1 = A k + 1 B k + 1 T
8:
     P k + 1 = P k + μ ( L k + 1 + S k + 1 X )
9:
    if  L k + 1 L k F L k F ϵ  then
10:
     break
11:
   end if
12:
end for
13:
return  L k + 1

4.3. Computational Complexity

Unlike many reference algorithms, Algorithm 1 uses T-SVD only for the initialization of factor tensors and does not employ it during the iterative steps, thereby avoiding the significant computational overhead typically incurred by repeated T-SVD computations, especially for large-scale tensors. In Algorithm 1, the main per-iteration computational cost arises from computing the inverse tensor and performing tensor–tensor products, which require O ( n 1 n 2 n 3 log n 3 + r 3 n 3 ) and O ( n 1 n 2 n 3 ( log n 3 + r ) ) flops, respectively. When r min ( n 1 , n 2 ) , the overall computational complexity of Algorithm 1 becomes O ( n 1 n 2 n 3 ( log n 3 + r ) . Moreover, since most computations on Algorithm 1 can be performed using BLAS level 3 operations, the actual computational speed is significantly improved with modern parallel computing architectures.

5. Experimental Results

In this section, we conduct numerical experiments to evaluate the performance of the proposed algorithm. To assess the improvements, we compare the execution time and accuracy of Algorithm 1 (hereafter referred to as TRPCAAL) with the methods proposed by Cai et al. [20] (hereafter IRCUR (https://github.com/huangl3/RTCUR (accessed on 16 July 2025))), Qiu et al. [19] (hereafter FTTNN (https://github.com/ynqiu/fast-TTRPCA (accessed on 16 July 2025))), Geng et al. [18] (hereafter N-TRPCA (https://github.com/qguo2010/NN-TRPCA (accessed on 16 July 2025))), Lu et al. [13] (hereafter TRPCA_TNN (https://github.com/canyilu/Tensor-Robust-Principal-Component-Analysis-TRPCA (accessed on 16 July 2025))), and Qie et al. [17] (hereafter EAPT-DCT (https://github.com/ucker/EAPT (accessed on 16 July 2025))). All experiments are conducted on a machine with an Intel i9-11900k processor and 64GB memory, using MATLAB version 9.10.00.1710957. The reported results are the average over 20 independent trials.

5.1. Color Image Recovery

Color images may contain sparse and impulse noise due to sensor malfunction, external interference, or transmission errors caused by network issues. Such noise significantly degrades the performance of image processing and computer vision algorithms, necessitating the separation of sparse noise from the image as a pre-processing step. One effective solution for recovering color images is to apply TRPCA models, which approximate the clean image as a low-rank tensor under sparse noise corruption. In this section, we compare the performance of various TRPCA algorithms in the context of color image recovery. To evaluate their effectiveness, we use three popular test images: “peppers”, “airplane”, and “house”, shown in Figure 1a, Figure 2a, and Figure 3a, respectively. The image sizes are 128 × 128 , 256 × 256 , and 512 × 512 . Impulse noise with random magnitudes and positions is added to the test images at different noise levels (10%, 20%, and 40%). Note that the noise level indicates the proportion of image pixels corrupted by impulse noise relative to the total number of pixels. Example images corrupted with 20% noise are shown in Figure 1b, Figure 2b, and Figure 3b.
To compare the performance, we measured the execution time, the number of iterations to converge, and the peak signal-to-noise ratio (PSNR), defined as
PSNR = 10 log 10 | | X | | 2 1 n 1 n 2 n 3 X L F 2 ,
where X represents the original image and L is the low-rank approximated image recovered by the algorithms. The PSNR value quantifies the overall differences between the original and recovered image. For further evaluation of recovery accuracy, we use the feature similarity index (FSIM) [25], which emphasizes structural similarity between the original and recovered image. For all algorithms, we set the stopping criterion ϵ = 1 e 4 and the maximum iteration number as 500. The stopping criterion is defined as
L k L k 1 F L k 1 F ϵ .
The hyperparameters is empirically chosen, for example r , λ and μ for TRPCAAL. Figure 4 illustrates PSNR and execution time as functions of μ and r using the “airplane” image with 20% random noise. The trends in the plots indicates that the choice of r is more important to the accuracy of TRPCAAL, while its effect on execution time is minimal. This implies that the proposed algorithm maintains computational stability. Throughout the additional experiments, we found that varying λ had a negligible impact on performance, so we fixed λ = 0.1 . For the other algorithms, we empirically selected the hyperparameters to optimize recovery accuracy.
Table 1, Table 2 and Table 3 present the experimental results for all algorithms. In the tables, the top two performing results in each category are highlighted for easier comparison. Additionally, example outputs with 20% random noise are shown in Figure 1, Figure 2 and Figure 3. From the results, we observe that N-TRPCA consistently achieves the highest accuracy in terms of PSNR and FSIM across most scenarios in the color image recovery application. However, as evident from the data, its execution time is significantly longer than that of the other algorithms. Apart from N-TRPCA, TRPCAAL generally ranks as the second most accurate algorithm, except in the cases of the “peppers” image with 20% and 40% noise, where FTTNN and TRPCA_TNN show slightly better accuracy. Nevertheless, even in these cases, the performance of TRPCAAL is only marginally lower. What makes TRPCAAL stand out is its execution time—it is the fastest across all test cases. Notably, TRPCAAL’s computational efficiency becomes increasingly apparent as image size increases. For example, when recovering a 128 × 128 image, TRPCAAL is approximately 33 times faster than N-TRPCA, the slowest among the tested algorithms. This gap widens to 43 times for a 512 × 512 image. Compared to the other algorithms known for their faster execution, TRPCAAL still outperforms them in execution time.
An additional experiment was conducted using the “airplane” image with impulse noises containing extremely large magnitudes—50 times greater than the maximum pixel brightness—to simulate gross corruptions or outlier entries in sparse locations. All hyperparameters for algorithms were selected empirically. Table 4 presents the experimental results. Similar to the experiments with random pixel noise, Algorithm 1 and N-TRPCA achieved the highest accuracy, as indicated by high PNSR and FSIM values. However, aside from these two algorithms, the others failed to produce meaningful outputs in the presence of outliers, even though some of them exhibited a faster execution time. Overall, the proposed algorithm achieves reasonably high recovery accuracy while offering the fastest execution speed, making it a highly appealing solution for large-scale color image recovery problems.

5.2. Background Subtraction

Another popular application of TRPCA is background subtraction. Since the video sequences can be modeled as a combination of a static background (low-rank component) and moving objects (sparse component), the background subtraction problem can be effectively addressed using TRPCA algorithms [26]. To evaluate the performance of the algorithms, three video sequences are used, which include the “highway”, “Bus station”, and “park”. We extract 300 frames from the video sequences with frame sizes of 240 × 320 , 240 × 360 , and 288 × 352 , respectively. To assess the accuracy of background subtraction, we compute precision, recall, and F-score based on the foreground detection results and corresponding ground-truth data. As in the color image recovery experiment in Section 5.1, we set the stopping criterion ϵ = 1.0 e 4 for all algorithms. The hyperparameters for each algorithms are chosen empirically to achieve the highest possible F-score value.
Table 5, Table 6 and Table 7 summarize the performance comparison of the algorithms. Note that the precision, recall, and F-score values presented in Table 5, Table 6 and Table 7 are averaged over all video frames. Figure 5a–c illustrate example frames from the video sequences, along with their ground-truth background subtraction results and the outputs from each algorithm. From the results, TRPCAAL consistently achieved the fastest execution time while maintaining competitive or the best F-score values across all video sequences, demonstrating strong overall performance in both speed and accuracy. Unlike the experimental results in color image recovery, N-TRPCA produced unsatisfactory results in both accuracy and execution time. While IRCUR achieved slightly higher F-score values in some cases, it suffered from extremely high execution times, making it impractical for real-time or large-sized video sequences. Therefore, TRPCAAL emerges as the most attractive choice, providing accurate results with minimal computational cost in background subtraction.

6. Conclusions

TRPCA is widely used in various applications due to its ability to separate a low-rank approximated tensor and sparse noise from the data. However, because the problem involves manipulating large-scale and complexly structured data, TRPCA typically requires substantial computational resources. To address this limitation, we propose a novel TRPCA computation approach by relaxing the tensor nuclear norm minimization into an optimization problem involving the sum of Frobenius norms of factor tensors. Since the low-rank approximated tensor can be represented using truncated factor tensors, we alternatingly update these factor tensors to efficiently compute the low-rank approximation with reduced computational cost. Experimental results show that the proposed method significantly outperforms other state-of-the-art TRPCA algorithms in terms of execution time while maintaining highly accurate recovery performance in practical applications such as color image recovery and background subtraction.

Funding

This work was supported by Hankuk University of Foreign Studies Research Fund.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Data presented in this study are available on request from the corresponding author.

Conflicts of Interest

The authors declare no conflicts of interest.

Abbreviations

The following abbreviations are used in this manuscript:
TRPCAtensor robust principal component analysis
T-SVDtensor singular value decomposition
T-producttensor–tensor product

References

  1. Candès, E.J.; Li, X.; Ma, Y.; Wright, J. Robust Principal Component Analysis? J. ACM 2011, 58, 1–37. [Google Scholar] [CrossRef]
  2. Bouwmans, T.; Javed, S.; Zhang, H.; Lin, Z.; Otazo, R. On the Applications of Robust PCA in image and Video Processing. Proc. IEEE 2018, 160, 1427–1457. [Google Scholar] [CrossRef]
  3. Cao, W.; Wang, Y.; Sun, J.; Meng, D.; Yan, C.; Cichocki, A.; Xu, Z. Total Variation Regularized Tensor RPCA for Background Subtraction from Compressive Measurement. IEEE Trans. Image Proc. 2016, 25, 4075–4090. [Google Scholar] [CrossRef] [PubMed]
  4. Hu, Y.; Liu, J.; Gao, Y.; Shang, J. DSTPCA: Double-Sparse Constrained Tensor Principal Component Analysis Method for Feature Selection. IEEE/ACM Trans. Comput. Biol. Bioinform. 2021, 18, 1481–1491. [Google Scholar] [CrossRef]
  5. Markowitz, S.; Snyder, C.; Eldar, Y.C.; Do, M.N. Mutilmodal unrolled robust RPCA for background foreground separation. IEEE Trans. Image Proc. 2022, 31, 3553–3564. [Google Scholar] [CrossRef]
  6. Zhong, G.; Pun, C.M. RPCA-induced self-representation for subspace clustering. Neurocomputing 2021, 437, 249–260. [Google Scholar] [CrossRef]
  7. Cai, J.F.; Candès, E.J.; Shen, Z. A singular value thresholding algorithm for matrix completion. SIAM J. Optim. 2010, 20, 1956–1982. [Google Scholar] [CrossRef]
  8. Kolda, T.G.; Bader, B.W. Tensor decomposition and applications. SIAM Rev. 2009, 51, 455–500. [Google Scholar] [CrossRef]
  9. Hillar, J.C.; Lim, L.H. Most tensor problems are NP-Hard. J. ACM 2013, 60, 1–39. [Google Scholar] [CrossRef]
  10. Liu, J.; Musialski, P.; Wonka, P.; Ye, J. Tensor completion for estimating missing values in visual data. IEEE Trans. Pattern Anal. Mach. Intell. 2013, 35, 208–220. [Google Scholar] [CrossRef]
  11. Kilmer, M.E.; Martin, C.D. Factorization strategies for third-order tensors. Linear Algebra Appl. 2011, 435, 641–658. [Google Scholar] [CrossRef]
  12. Kilmer, M.E.; Horesh, L.; Avron, H.; Newman, E. Tensor-tensor algebra for optimal representation and compression of multiway data. Proc. Natl. Acad. Sci. USA 2021, 118, e205851118. [Google Scholar] [CrossRef]
  13. Lu, C.; Feng, J.; Chen, Y.; Liu, W.; Lin, Z.; Yan, S. Tensor Robust Principal Component Analysis with a New Tensor Nuclear Norm. IEEE Trans. Pattern Anal. Mach. Intell. 2020, 42, 925–938. [Google Scholar] [CrossRef]
  14. Zhang, Z.; Ely, G.; Aeron, S.; Hao, N.; Kilmer, M.E. Novel methods for multilinear data completion and de-noising based on tensor-SVD. IEEE Conf. Comput. Vis. Pattern Recognit. 2014, 3842–3849. [Google Scholar] [CrossRef]
  15. Gao, Q.; Zhang, P.; Xia, W.; Xie, D.; Gao, X.; Tao, D. Enhanced Tensor RPCA and its Application. IEEE Trans. Pattern Anal. Machine Intell. 2021, 43, 2133–2140. [Google Scholar] [CrossRef] [PubMed]
  16. Dong, H.; Tong, T.; Ma, C.; Chi, Y. Fast and provable tensor robust principal component analysis via scaled gradient descent. Inform. Inference J. IMA 2023, 12, iaad019. [Google Scholar] [CrossRef]
  17. Qui, H.; Wang, Y.; Tang, S.; Meng, D.; Yao, Q. Fast and Provable Nonconvex Tensor RPCA. In Proceedings of the 39th International Conference on Machine Learning, Baltimore, MI, USA, 17–23 July 2022. [Google Scholar]
  18. Geng, X.; Guo, Q.; Hui, S.; Yang, M.; Zhang, C. Tensor robust PCA with nonconvex and nonlocal regularization. Comput. Vision Image Understand 2024, 243, 104007. [Google Scholar] [CrossRef]
  19. Qui, Y.; Zhou, G.; Huang, Z.; Zhao, Q.; Xie, S. Efficient Tensor Robust PCA Under Hybrid Model of Tucker and Tensor Train. IEEE Sig. Proc. Lett. 2022, 29, 627–631. [Google Scholar]
  20. Cai, H.; Chao, Z.; Huang, L.; Needell, D. Fast Robust Tensor Principal Component Analysis via Fiber CUR Decomposition. IEEE/CVF Int. Conf. Comput. Vision Work. 2021, 189–197. [Google Scholar] [CrossRef]
  21. Salut, M.M.; Anderson, D.V. Online Tensor Robust Principal Component Analysis. IEEE Access 2022, 10, 69354–69363. [Google Scholar] [CrossRef]
  22. Hastie, T.; Mazumder, R.; Lee, J.D.; Zadeh, R. Matrix Completion and Low-Rank SVD via Fast Alternating Least Squares. J. Machin. Learn. Res. 2015, 16, 3367–3402. [Google Scholar]
  23. Mazumder, R.; Hastie, T.; Tibshirani, R. Spectral Regularization Algorithms for Learning Large Incomplete Matrices. J. Machine Learn. Res. 2010, 11, 2287–2322. [Google Scholar]
  24. Tibshirani, R. Regression shrinkage and selection via the lasso: A retrospective. J. R. Stat. Soc. 2011, 73, 273–282. [Google Scholar] [CrossRef]
  25. Zhang, H.; Gong, C.; Qian, J.; Zhang, B.; Xu, C.; Yang, J. Efficient recovery of low-rank via double nonconvex nonsmooth rank minimization. IEEE Trans. Neural Net. Learn. Syst. 2019, 30, 2916–2925. [Google Scholar] [CrossRef]
  26. Bouwmans, T.; Zahzah, E.H. Robust PCA via Principal Component Pursuit: A review for a comparative evaluation in video surveillance. Comput. Vision Image Understand 2014, 122, 22–34. [Google Scholar] [CrossRef]
Figure 1. Test images used in color image recovery when the “peppers” image with 20% random noise is used.
Figure 1. Test images used in color image recovery when the “peppers” image with 20% random noise is used.
Applsci 15 08114 g001
Figure 2. Test images used in color image recovery when the “airplane” image with 20% random noise is used.
Figure 2. Test images used in color image recovery when the “airplane” image with 20% random noise is used.
Applsci 15 08114 g002
Figure 3. Test images used in color image recovery when the “house” image with 20% random noise is used.
Figure 3. Test images used in color image recovery when the “house” image with 20% random noise is used.
Applsci 15 08114 g003
Figure 4. Empirical hyperparameter selection based on the variation in μ and r. Note that the “airplane” image with 20% random noise is used.
Figure 4. Empirical hyperparameter selection based on the variation in μ and r. Note that the “airplane” image with 20% random noise is used.
Applsci 15 08114 g004
Figure 5. Background subtraction results from video sequences. From left to right, original frame and output from TRPCAAL, FTTNN, IRCUR, TRPCA_TNN, N-TRPCA, and EAPT-DCT, respectively.
Figure 5. Background subtraction results from video sequences. From left to right, original frame and output from TRPCAAL, FTTNN, IRCUR, TRPCA_TNN, N-TRPCA, and EAPT-DCT, respectively.
Applsci 15 08114 g005
Table 1. Experimental results for the “peppers” image with 10%, 20%, and 40% noise levels. The top two results in execution time and accuracy are bolded.
Table 1. Experimental results for the “peppers” image with 10%, 20%, and 40% noise levels. The top two results in execution time and accuracy are bolded.
Noise IterationsExecution TimePSNR (dB)FSIM
10%noisy image--18.00720.8318
TRPCAAL9.000.040425.60080.9305
FTTNN27.000.134323.82660.8816
IRCUR5.951.137622.93370.8581
TRPCA_TNN59.450.226123.53370.9051
N-TRPCA-1.104625.42160.9369
EATP-DCT12.300.049422.63690.8710
20%noisy image--14.96300.7333
TRPCAAL11.000.042622.05610.8594
FTTNN27.000.137620.76890.8095
IRCUR5.951.142019.82460.7855
TRPCA_TNN58.750.223221.99830.8615
N-TRPCA-1.330623.23040.8950
EATP-DCT11.700.048420.59740.8285
40%noisy image--11.95940.6209
TRPCAAL12.900.024017.32870.7286
FTTNN25.000.114417.58320.7196
IRCUR5.851.111316.51590.6955
TRPCA_TNN58.550.231216.92190.7254
N-TRPCA-1.140618.87620.7780
EATP-DCT11.450.044216.02440.7030
Table 2. Experimental results for the “airplane” image with 10%, 20%, and 40% noise levels. The top two results in execution time and accuracy are bolded.
Table 2. Experimental results for the “airplane” image with 10%, 20%, and 40% noise levels. The top two results in execution time and accuracy are bolded.
Noise IterationsExecution TimePSNR (dB)FSIM
10%noisy image--17.40300.7873
TRPCAAL7.000.192230.88060.9564
FTTNN26.000.321126.54450.8668
IRCUR5.953.260826.02130.8629
TRPCA_TNN58.101.041428.13010.9327
N-TRPCA-5.050632.25810.9691
EATP-DCT19.700.350626.85690.8964
20%noisy image--14.39600.6763
TRPCAAL12.000.144427.03070.9083
FTTNN26.000.331124.68610.8249
IRCUR5.903.226623.12910.7837
TRPCA_TNN58.101.040026.45280.9040
N-TRPCA-5.061329.31790.9409
EATP-DCT19.050.341123.89890.8517
40%noisy image--11.40250.5574
TRPCAAL14.000.109021.46350.7780
FTTNN24.000.280719.48780.6968
IRCUR5.852.338417.95450.6502
TRPCA_TNN57.601.046820.08160.7619
N-TRPCA-5.034423.61630.8286
EATP-DCT14.450.278715.36440.6345
Table 3. Experimental results for the “house” image with 10%, 20%, and 40% noise levels. The top two results in execution time and accuracy are bolded.
Table 3. Experimental results for the “house” image with 10%, 20%, and 40% noise levels. The top two results in execution time and accuracy are bolded.
Noise IterationsExecution TimePSNR (dB)FSIM
10%noisy image--18.64440.8690
TRPCAAL8.000.343537.72540.9915
FTTNN26.001.019633.95940.9725
IRCUR6.8016.183433.51310.9698
TRPCA_TNN58.054.327835.61280.9838
N-TRPCA-19.425942.15220.9976
EATP-DCT24.101.419735.27860.9740
20%noisy image--15.63220.7935
TRPCAAL13.150.522334.35710.9830
FTTNN25.000.842529.58790.9186
IRCUR5.9010.610128.11230.9047
TRPCA_TNN57.254.279833.02070.9730
N-TRPCA-19.486438.83580.9940
EATP-DCT27.351.559833.94890.9716
40%noisy image--12.62720.6913
TRPCAAL17.100.503427.00850.9195
FTTNN24.000.768223.61060.8241
IRCUR19.2018.177924.31230.8327
TRPCA_TNN56.604.399324.72240.9139
N-TRPCA-20.249231.53740.9640
EATP-DCT44.802.376220.38440.8389
Table 4. Experimental results for the “airplane” image with 10%, 20%, and 40% noise levels. The top two results in execution time and accuracy are bolded.
Table 4. Experimental results for the “airplane” image with 10%, 20%, and 40% noise levels. The top two results in execution time and accuracy are bolded.
Noise IterationsExecution TimePSNR (dB)FSIM
10%noisy image--14.50790.0965
TRPCAAL65.950.603929.61320.9305
FTTNN20.000.27964.77970.0898
IRCUR22.2510.069415.55290.6053
TRPCA_TNN63.001.262112.91680.6503
N-TRPCA-6.264632.34810.9704
EATP-DCT24.250.368012.98460.3860
20%noisy image--11.50190.0263
TRPCAAL67.450.615827.12140.9012
FTTNN20.000.29264.56660.0615
IRCUR21.359.607515.05620.5743
TRPCA_TNN62.151.232610.69870.5592
N-TRPCA-6.257229.53710.9449
EATP-DCT23.150.34838.99720.1999
40%noisy image--8.48110.0135
TRPCAAL87.850.682821.91680.7935
FTTNN20.000.27974.50930.0427
IRCUR20.209.187313.83730.4780
TRPCA_TNN62.001.22498.16580.4618
N-TRPCA-6.268924.11660.8463
EATP-DCT24.000.35974.66470.0194
Table 5. Experimental results for the “highway” video sequence. The top two results in terms of execution time and F-score are bolded for comparison.
Table 5. Experimental results for the “highway” video sequence. The top two results in terms of execution time and F-score are bolded for comparison.
Noise IterationsExecution TimePrecisionRecallF-Score
TRPCAAL43.73910.52950.56720.5407
FTTNN2532.61190.47490.14680.2041
IRCUR24513.97140.48550.58980.5269
TRPCA_TNN4382.74040.02440.43750.0456
N-TRPCA-249.20750.09540.37530.1474
EATP-DCT2532.96070.43770.55870.4826
Table 6. Experimental results for the “bus station” video sequence. The top two results in terms of execution time and F-score are bolded for comparison.
Table 6. Experimental results for the “bus station” video sequence. The top two results in terms of execution time and F-score are bolded for comparison.
Noise IterationsExecution TimePrecisionRecallF-Score
TRPCAAL44.21060.37190.54770.4356
FTTNN2536.87980.33220.15210.2047
IRCUR24585.05440.17040.64910.2565
TRPCA_TNN4393.04260.11950.15490.1310
N-TRPCA-287.72770.73930.02430.0470
EATP-DCT2638.58580.16350.55860.2455
Table 7. Experimental results for the “park” video sequence. The top two results in terms of execution time and F-score are bolded for comparison.
Table 7. Experimental results for the “park” video sequence. The top two results in terms of execution time and F-score are bolded for comparison.
Noise IterationsExecution TimePrecisionRecallF-Score
TRPCAAL44.95720.27270.44360.3294
FTTNN2435.09600.31050.07380.1157
IRCUR23658.97670.25410.58750.3475
TRPCA_TNN40112.55050.64250.02010.0387
N-TRPCA-346.12220.05830.03290.0396
EATP-DCT2645.25970.12810.56210.2040
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Lee, G. Accelerated Tensor Robust Principal Component Analysis via Factorized Tensor Norm Minimization. Appl. Sci. 2025, 15, 8114. https://doi.org/10.3390/app15148114

AMA Style

Lee G. Accelerated Tensor Robust Principal Component Analysis via Factorized Tensor Norm Minimization. Applied Sciences. 2025; 15(14):8114. https://doi.org/10.3390/app15148114

Chicago/Turabian Style

Lee, Geunseop. 2025. "Accelerated Tensor Robust Principal Component Analysis via Factorized Tensor Norm Minimization" Applied Sciences 15, no. 14: 8114. https://doi.org/10.3390/app15148114

APA Style

Lee, G. (2025). Accelerated Tensor Robust Principal Component Analysis via Factorized Tensor Norm Minimization. Applied Sciences, 15(14), 8114. https://doi.org/10.3390/app15148114

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop