Robust principal component analysis

From HandWiki
Revision as of 22:36, 6 February 2024 by Rjetedi (talk | contribs) (over-write)
(diff) ← Older revision | Latest revision (diff) | Newer revision → (diff)

Robust Principal Component Analysis (RPCA) is a modification of the widely used statistical procedure of principal component analysis (PCA) which works well with respect to grossly corrupted observations. A number of different approaches exist for Robust PCA, including an idealized version of Robust PCA, which aims to recover a low-rank matrix L0 from highly corrupted measurements M = L0 +S0.[1] This decomposition in low-rank and sparse matrices can be achieved by techniques such as Principal Component Pursuit method (PCP),[1] Stable PCP,[2] Quantized PCP,[3] Block based PCP,[4] and Local PCP.[5] Then, optimization methods are used such as the Augmented Lagrange Multiplier Method (ALM[6]), Alternating Direction Method (ADM[7]), Fast Alternating Minimization (FAM[8]), Iteratively Reweighted Least Squares (IRLS [9][10][11]) or alternating projections (AP[12][13][14]).

Algorithms

Non-convex method

The 2014 guaranteed algorithm for the robust PCA problem (with the input matrix being [math]\displaystyle{ M=L+S }[/math]) is an alternating minimization type algorithm.[12] The computational complexity is [math]\displaystyle{ O\left(m n r^2 \log\frac{1}{\epsilon} \right) }[/math] where the input is the superposition of a low-rank (of rank [math]\displaystyle{ r }[/math]) and a sparse matrix of dimension [math]\displaystyle{ m \times n }[/math] and [math]\displaystyle{ \epsilon }[/math] is the desired accuracy of the recovered solution, i.e., [math]\displaystyle{ \| \widehat{L}-L \|_F \leq \epsilon }[/math] where [math]\displaystyle{ L }[/math] is the true low-rank component and [math]\displaystyle{ \widehat{L} }[/math] is the estimated or recovered low-rank component. Intuitively, this algorithm performs projections of the residual onto the set of low-rank matrices (via the SVD operation) and sparse matrices (via entry-wise hard thresholding) in an alternating manner - that is, low-rank projection of the difference the input matrix and the sparse matrix obtained at a given iteration followed by sparse projection of the difference of the input matrix and the low-rank matrix obtained in the previous step, and iterating the two steps until convergence.

This alternating projections algorithm is later improved by an accelerated version, coined AccAltProj.[13] The acceleration is achieved by applying a tangent space projection before project the residue onto the set of low-rank matrices. This trick improves the computational complexity to [math]\displaystyle{ O\left(m n r \log\frac{1}{\epsilon} \right) }[/math] with a much smaller constant in front while it maintains the theoretically guaranteed linear convergence.

Another fast version of accelerated alternating projections algorithm is IRCUR.[14] It uses the structure of CUR decomposition in alternating projections framework to dramatically reduces the computational complexity of RPCA to [math]\displaystyle{ O\left(\max\{m,n\} r^2 \log(m) \log(n) \log\frac{1}{\epsilon} \right) }[/math]

Convex relaxation

This method consists of relaxing the rank constraint [math]\displaystyle{ rank(L) }[/math] in the optimization problem to the nuclear norm [math]\displaystyle{ \|L\|_* }[/math] and the sparsity constraint [math]\displaystyle{ \|S\|_0 }[/math] to [math]\displaystyle{ \ell_1 }[/math]-norm [math]\displaystyle{ \|S\|_1 }[/math]. The resulting program can be solved using methods such as the method of Augmented Lagrange Multipliers.

Deep-learning augmented method

Some recent works propose RPCA algorithms with learnable/training parameters.[15] Such a learnable/trainable algorithm can be unfolded as a deep neural network whose parameters can be learned via machine learning techniques from a given dataset or problem distribution. The learned algorithm will have superior performance on the corresponding problem distribution.

Applications

RPCA has many real life important applications particularly when the data under study can naturally be modeled as a low-rank plus a sparse contribution. Following examples are inspired by contemporary challenges in computer science, and depending on the applications, either the low-rank component or the sparse component could be the object of interest:

Video surveillance

Given a sequence of surveillance video frames, it is often required to identify the activities that stand out from the background. If we stack the video frames as columns of a matrix M, then the low-rank component L0 naturally corresponds to the stationary background and the sparse component S0 captures the moving objects in the foreground.[1][16]

Face recognition

Images of a convex, Lambertian surface under varying illuminations span a low-dimensional subspace.[17] This is one of the reasons for effectiveness of low-dimensional models for imagery data. In particular, it is easy to approximate images of a human's face by a low-dimensional subspace. To be able to correctly retrieve this subspace is crucial in many applications such as face recognition and alignment. It turns out that RPCA can be applied successfully to this problem to exactly recover the face.[1]

See also

Surveys

  • Robust PCA [16]
  • Dynamic RPCA [18]
  • Decomposition into Low-rank plus Additive Matrices [19]
  • Low-rank models[20]

Books, journals and workshops

Books

Journals

Workshops

Sessions

  • Special Session on "Online Algorithms for Static and Dynamic Robust PCA and Compressive Sensing" in conjunction with SSP 2018. (More information: https://ssp2018.org/)

Resources and libraries

Websites

Libraries

The LRS Library (developed by Andrews Sobral) provides a collection of low-rank and sparse decomposition algorithms in MATLAB. The library was designed for moving object detection in videos, but it can be also used for other computer vision / machine learning tasks. Currently the LRSLibrary offers more than 100 algorithms based on matrix and tensor methods.

References

  1. 1.0 1.1 1.2 1.3 Emmanuel J. Candes; Xiaodong Li; Yi Ma; John Wright (2009). "Robust Principal Component Analysis?". Journal of the ACM 58 (3): 1–37. doi:10.1145/1970392.1970395. 
  2. J. Wright; Y. Peng; Y. Ma; A. Ganesh; S. Rao (2009). "Robust Principal Component Analysis: Exact Recovery of Corrupted Low-Rank Matrices by Convex Optimization". Neural Information Processing Systems, NIPS 2009. 
  3. S. Becker; E. Candes, M. Grant (2011). "TFOCS: Flexible First-order Methods for Rank Minimization". Low-rank Matrix Optimization Symposium, SIAM Conference on Optimization. 
  4. G. Tang; A. Nehorai (2011). "Robust principal component analysis based on low-rank and block-sparse matrix decomposition". 2011 45th Annual Conference on Information Sciences and Systems. pp. 1–5. doi:10.1109/CISS.2011.5766144. ISBN 978-1-4244-9846-8. 
  5. B. Wohlberg; R. Chartrand; J. Theiler (2012). "Local principal component pursuit for nonlinear datasets". 2012 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP). pp. 3925–3928. doi:10.1109/ICASSP.2012.6288776. ISBN 978-1-4673-0046-9. 
  6. Z. Lin; M. Chen; L. Wu; Y. Ma (2013). "The Augmented Lagrange Multiplier Method for Exact Recovery of Corrupted Low-Rank Matrices". Journal of Structural Biology 181 (2): 116–27. doi:10.1016/j.jsb.2012.10.010. PMID 23110852. 
  7. X. Yuan; J. Yang (2009). "Sparse and Low-Rank Matrix Decomposition via Alternating Direction Methods". Optimization Online. 
  8. P. Rodríguez; B. Wohlberg (2013). "Fast principal component pursuit via alternating minimization". 2013 IEEE International Conference on Image Processing. pp. 69–73. doi:10.1109/ICIP.2013.6738015. ISBN 978-1-4799-2341-0. 
  9. C. Guyon; T. Bouwmans; E. Zahzah (2012). "Foreground Detection via Robust Low Rank Matrix Decomposition including Spatio-Temporal Constraint". International Workshop on Background Model Challenges, ACCV 2012. 
  10. C. Guyon; T. Bouwmans; E. Zahzah (2012). "Foreground Detection via Robust Low Rank Matrix Factorization including Spatial Constraint with Iterative Reweighted Regression". International Conference on Pattern Recognition, ICPR 2012. 
  11. C. Guyon; T. Bouwmans; E. Zahzah (2012). "Moving Object Detection via Robust Low Rank Matrix Decomposition with IRLS scheme". International Symposium on Visual Computing, ISVC 2012. 
  12. 12.0 12.1 P., Netrapalli; U., Niranjan; S., Sanghavi; A., Anandkumar; P., Jain (2014). "Non-convex robust PCA". Advances in Neural Information Processing Systems 27: 1107–1115. Bibcode2014arXiv1410.7660N. 
  13. 13.0 13.1 Cai, H.; Cai, J.-F.; Wei, K. (2019). "Accelerated alternating projections for robust principal component analysis". The Journal of Machine Learning Research 20 (1): 685–717. Bibcode2017arXiv171105519C. 
  14. 14.0 14.1 Cai, H.; Hamm, K.; Huang, L.; Li, J.; Wang, T. (2021). "Rapid Robust Principal Component Analysis: CUR Accelerated Inexact Low Rank Estimation". IEEE Signal Processing Letters 28: 116–120. doi:10.1109/LSP.2020.3044130. Bibcode2021ISPL...28..116C. 
  15. Cai, H.; Liu, J.; Yin, W. (2021). "Learned Robust PCA: A Scalable Deep Unfolding Approach for High-Dimensional Outlier Detection". Advances in Neural Information Processing Systems 34: 16977–16989. Bibcode2021arXiv211005649C. 
  16. 16.0 16.1 T. Bouwmans; E. Zahzah (2014). "Robust PCA via Principal Component Pursuit: A Review for a Comparative Evaluation in Video Surveillance". Computer Vision and Image Understanding 122: 22–34. doi:10.1016/j.cviu.2013.11.009. 
  17. R. Basri; D. Jacobs. Lambertian reflectance and linear subspaces. 
  18. N. Vaswani; T. Bouwmans; S. Javed; P. Narayanamurthy (2017). "Robust PCA and Robust Subspace Tracking". Preprint 35 (4): 32–55. doi:10.1109/MSP.2018.2826566. Bibcode2018ISPM...35d..32V. 
  19. T. Bouwmans; A. Sobral; S. Javed; S. Jung; E. Zahzahg (2015). "Decomposition into Low-rank plus Additive Matrices for Background/Foreground Separation: A Review for a Comparative Evaluation with a Large-Scale Dataset". Computer Science Review 23: 1–71. doi:10.1016/j.cosrev.2016.11.001. Bibcode2015arXiv151101245B. 
  20. Z. Lin (2016). "A Review on Low-Rank Models in Data Analysis". Big Data and Information Analytics 1 (2): 139–161. doi:10.3934/bdia.2016001. 

External links