Skip to main content
Erschienen in: Knowledge and Information Systems 3/2018

12.06.2017 | Regular Paper

Three iteratively reweighted least squares algorithms for \(L_1\)-norm principal component analysis

verfasst von: Young Woong Park, Diego Klabjan

Erschienen in: Knowledge and Information Systems | Ausgabe 3/2018

Einloggen

Aktivieren Sie unsere intelligente Suche, um passende Fachinhalte oder Patente zu finden.

search-config
loading …

Abstract

Principal component analysis (PCA) is often used to reduce the dimension of data by selecting a few orthonormal vectors that explain most of the variance structure of the data. \(L_1\) PCA uses the \(L_1\) norm to measure error, whereas the conventional PCA uses the \(L_2\) norm. For the \(L_1\) PCA problem minimizing the fitting error of the reconstructed data, we propose three algorithms based on iteratively reweighted least squares. We first develop an exact reweighted algorithm. Next, an approximate version is developed based on eigenpair approximation when the algorithm is near convergent. Finally, the approximate version is extended based on stochastic singular value decomposition. We provide convergence analyses, and compare their performance against benchmark algorithms in the literature. The computational experiment shows that the proposed algorithms consistently perform the best and the scalability is improved as we use eigenpair approximation and stochastic singular value decomposition.

Sie haben noch keine Lizenz? Dann Informieren Sie sich jetzt über unsere Produkte:

Springer Professional "Wirtschaft+Technik"

Online-Abonnement

Mit Springer Professional "Wirtschaft+Technik" erhalten Sie Zugriff auf:

  • über 102.000 Bücher
  • über 537 Zeitschriften

aus folgenden Fachgebieten:

  • Automobil + Motoren
  • Bauwesen + Immobilien
  • Business IT + Informatik
  • Elektrotechnik + Elektronik
  • Energie + Nachhaltigkeit
  • Finance + Banking
  • Management + Führung
  • Marketing + Vertrieb
  • Maschinenbau + Werkstoffe
  • Versicherung + Risiko

Jetzt Wissensvorsprung sichern!

Springer Professional "Wirtschaft"

Online-Abonnement

Mit Springer Professional "Wirtschaft" erhalten Sie Zugriff auf:

  • über 67.000 Bücher
  • über 340 Zeitschriften

aus folgenden Fachgebieten:

  • Bauwesen + Immobilien
  • Business IT + Informatik
  • Finance + Banking
  • Management + Führung
  • Marketing + Vertrieb
  • Versicherung + Risiko




Jetzt Wissensvorsprung sichern!

Springer Professional "Technik"

Online-Abonnement

Mit Springer Professional "Technik" erhalten Sie Zugriff auf:

  • über 67.000 Bücher
  • über 390 Zeitschriften

aus folgenden Fachgebieten:

  • Automobil + Motoren
  • Bauwesen + Immobilien
  • Business IT + Informatik
  • Elektrotechnik + Elektronik
  • Energie + Nachhaltigkeit
  • Maschinenbau + Werkstoffe




 

Jetzt Wissensvorsprung sichern!

Fußnoten
1
Iteratively reweighted least squares (IRLS) is an algorithmic framework to solve weighted least squares where the weights depend on the model parameters. Since the weights change based on the model parameters, iterations are needed until some convergence or a termination criteria are met.
 
Literatur
1.
Zurück zum Zitat Baccini A, Besse P, de Faguerolles A (March 1996) A \(L_1\)-norm PCA and a heuristic approach. In: Proceedings of the international conference on ordinal and symbolic data analysis, pp 359–368 Baccini A, Besse P, de Faguerolles A (March 1996) A \(L_1\)-norm PCA and a heuristic approach. In: Proceedings of the international conference on ordinal and symbolic data analysis, pp 359–368
2.
Zurück zum Zitat Bache K, Lichman M (2013) UCI Machine Learning Repository Bache K, Lichman M (2013) UCI Machine Learning Repository
4.
Zurück zum Zitat Brooks J, Jot S (2012) pcaL1: An implementation in R of three methods for \(L_1\)-norm principal component analysis. Optimization Online Brooks J, Jot S (2012) pcaL1: An implementation in R of three methods for \(L_1\)-norm principal component analysis. Optimization Online
6.
7.
Zurück zum Zitat Croux C, Ruiz-Gazen A (2005) High breakdown estimators for principal components: the projection-pursuit approach revisited. J Multivar Anal 95(1):206–226MathSciNetCrossRefMATH Croux C, Ruiz-Gazen A (2005) High breakdown estimators for principal components: the projection-pursuit approach revisited. J Multivar Anal 95(1):206–226MathSciNetCrossRefMATH
8.
Zurück zum Zitat Daubechies I, DeVore R, Fornasier M, Gntrk CS (2010) Iteratively reweighted least squares minimization for sparse recovery. Commun Pure Appl Math 63(1):1–38MathSciNetCrossRefMATH Daubechies I, DeVore R, Fornasier M, Gntrk CS (2010) Iteratively reweighted least squares minimization for sparse recovery. Commun Pure Appl Math 63(1):1–38MathSciNetCrossRefMATH
9.
10.
Zurück zum Zitat Geladi P, Kowalski BR (1986) Partial least-squares regression: a tutorial. Anal Chim Acta 185:1–17CrossRef Geladi P, Kowalski BR (1986) Partial least-squares regression: a tutorial. Anal Chim Acta 185:1–17CrossRef
11.
Zurück zum Zitat Halko N, Martinsson PG, Tropp JA (2011) Finding structure with randomness: probabilistic algorithms for constructing approximate matrix decompositions. SIAM Rev 53(2):217–288MathSciNetCrossRefMATH Halko N, Martinsson PG, Tropp JA (2011) Finding structure with randomness: probabilistic algorithms for constructing approximate matrix decompositions. SIAM Rev 53(2):217–288MathSciNetCrossRefMATH
12.
Zurück zum Zitat Jolliffe I (2002) Principal component analysis. Springer, New YorkMATH Jolliffe I (2002) Principal component analysis. Springer, New YorkMATH
13.
Zurück zum Zitat Jorgensen M (2006) Iteratively reweighted least squares. Wiley, New YorkCrossRef Jorgensen M (2006) Iteratively reweighted least squares. Wiley, New YorkCrossRef
14.
Zurück zum Zitat Jot S, Brooks P, Visentin A, Park YW (2016) pcaL1: \(L_1\)-norm PCA methods. R package version 1.4.1 Jot S, Brooks P, Visentin A, Park YW (2016) pcaL1: \(L_1\)-norm PCA methods. R package version 1.4.1
15.
Zurück zum Zitat Ke Q, Kanade T (June 2005) Robust \(L_1\) norm factorization in the presence of outliers and missing data by alternative convex programming. In: Proceedings of the 2005 IEEE computer society conference on computer vision and pattern recognition, pp 739–746 Ke Q, Kanade T (June 2005) Robust \(L_1\) norm factorization in the presence of outliers and missing data by alternative convex programming. In: Proceedings of the 2005 IEEE computer society conference on computer vision and pattern recognition, pp 739–746
16.
Zurück zum Zitat Kwak N (2008) Principal component analysis based on \(L_1\)-norm maximization. IEEE Trans Pattern Anal Mach Intell 30(9):1672–1680CrossRef Kwak N (2008) Principal component analysis based on \(L_1\)-norm maximization. IEEE Trans Pattern Anal Mach Intell 30(9):1672–1680CrossRef
17.
Zurück zum Zitat Lax PD (2007) Linear algebra and its applications. Wiley, HobokenMATH Lax PD (2007) Linear algebra and its applications. Wiley, HobokenMATH
18.
Zurück zum Zitat Li G, Chen Z (1985) Projection-pursuit approach to robust dispersion matrices and principal components: primary theory and monte carlo. J Am Stat Assoc 80(391):759–766CrossRefMATH Li G, Chen Z (1985) Projection-pursuit approach to robust dispersion matrices and principal components: primary theory and monte carlo. J Am Stat Assoc 80(391):759–766CrossRefMATH
19.
Zurück zum Zitat Markopoulos PP, Karystinos GN, Pados DA (2014) Optimal algorithms for \(L_1\)-subspace signal processing. IEEE Trans Signal Process 62(19):5046–5058MathSciNetCrossRef Markopoulos PP, Karystinos GN, Pados DA (2014) Optimal algorithms for \(L_1\)-subspace signal processing. IEEE Trans Signal Process 62(19):5046–5058MathSciNetCrossRef
20.
21.
Zurück zum Zitat Nie F, Huang H, Ding C, Luo D, Wang H (June 2011) Robust principal component analysis with non-greedy \(L_1\)-norm maximization. In: Proceeding of 22nd international conference on artificial intelligence, pp 1433–1438 Nie F, Huang H, Ding C, Luo D, Wang H (June 2011) Robust principal component analysis with non-greedy \(L_1\)-norm maximization. In: Proceeding of 22nd international conference on artificial intelligence, pp 1433–1438
22.
Zurück zum Zitat Park YW, Klabjan D (2016) Iteratively reweighted least squares algorithms for \(L_1\)-norm principal component analysis. In 2016 IEEE international conference on data mining, Barcelona Park YW, Klabjan D (2016) Iteratively reweighted least squares algorithms for \(L_1\)-norm principal component analysis. In 2016 IEEE international conference on data mining, Barcelona
23.
Zurück zum Zitat Roweis S (1998) EM Algorithms for PCA and SPCA. In: Advances in neural information processing systems, pp 626–632 Roweis S (1998) EM Algorithms for PCA and SPCA. In: Advances in neural information processing systems, pp 626–632
24.
Zurück zum Zitat R Core Team (2014) R: A language and environment for statistical computing. R Foundation for Statistical Computing, Vienna, Austria R Core Team (2014) R: A language and environment for statistical computing. R Foundation for Statistical Computing, Vienna, Austria
25.
Zurück zum Zitat Shmueli Y, Wolf G, Averbuch A (2012) Updating kernel methods in spectral decomposition by affinity perturbations. Linear Algebra Appl 437:1356–1365MathSciNetCrossRefMATH Shmueli Y, Wolf G, Averbuch A (2012) Updating kernel methods in spectral decomposition by affinity perturbations. Linear Algebra Appl 437:1356–1365MathSciNetCrossRefMATH
26.
Metadaten
Titel
Three iteratively reweighted least squares algorithms for -norm principal component analysis
verfasst von
Young Woong Park
Diego Klabjan
Publikationsdatum
12.06.2017
Verlag
Springer London
Erschienen in
Knowledge and Information Systems / Ausgabe 3/2018
Print ISSN: 0219-1377
Elektronische ISSN: 0219-3116
DOI
https://doi.org/10.1007/s10115-017-1069-6

Weitere Artikel der Ausgabe 3/2018

Knowledge and Information Systems 3/2018 Zur Ausgabe