Skip to main content
Top
Published in: The Journal of Supercomputing 8/2016

01-08-2016

Critical data points-based unsupervised linear dimension reduction technology for science data

Authors: Di Wu, Naixue Xiong, Jinrong He, Chuanhe Huang

Published in: The Journal of Supercomputing | Issue 8/2016

Log in

Activate our intelligent search to find suitable subject content or patents.

search-config
loading …

Abstract

Recent advances in machine learning and data mining have led to powerful methods for the analysis and visualization of high dimensional data. This paper proposes an unsupervised linear dimension reduction algorithm named critical points preserving projection (CPPP). Selecting some key data points to represent the others has become more and more popular owing to its effectiveness and efficiency. Rather than considering all data points equally, the proposed algorithm just preserves both local neighborhood structure and global variance of critical data points. We explore a joint modification of locality preserving projection and principal component analysis to achieve these objectives. Experimental results on the UCI data sets show its good performance on pattern classification.

Dont have a licence yet? Then find out more about our products and how to get one now:

Springer Professional "Wirtschaft"

Online-Abonnement

Mit Springer Professional "Wirtschaft" erhalten Sie Zugriff auf:

  • über 67.000 Bücher
  • über 340 Zeitschriften

aus folgenden Fachgebieten:

  • Bauwesen + Immobilien
  • Business IT + Informatik
  • Finance + Banking
  • Management + Führung
  • Marketing + Vertrieb
  • Versicherung + Risiko




Jetzt Wissensvorsprung sichern!

Springer Professional "Technik"

Online-Abonnement

Mit Springer Professional "Technik" erhalten Sie Zugriff auf:

  • über 67.000 Bücher
  • über 390 Zeitschriften

aus folgenden Fachgebieten:

  • Automobil + Motoren
  • Bauwesen + Immobilien
  • Business IT + Informatik
  • Elektrotechnik + Elektronik
  • Energie + Nachhaltigkeit
  • Maschinenbau + Werkstoffe




 

Jetzt Wissensvorsprung sichern!

Springer Professional "Wirtschaft+Technik"

Online-Abonnement

Mit Springer Professional "Wirtschaft+Technik" erhalten Sie Zugriff auf:

  • über 102.000 Bücher
  • über 537 Zeitschriften

aus folgenden Fachgebieten:

  • Automobil + Motoren
  • Bauwesen + Immobilien
  • Business IT + Informatik
  • Elektrotechnik + Elektronik
  • Energie + Nachhaltigkeit
  • Finance + Banking
  • Management + Führung
  • Marketing + Vertrieb
  • Maschinenbau + Werkstoffe
  • Versicherung + Risiko

Jetzt Wissensvorsprung sichern!

Literature
1.
go back to reference Bellman R (1961) Adaptive control process: a guide tour. Princeton University Press, Princeton, New JerseyCrossRef Bellman R (1961) Adaptive control process: a guide tour. Princeton University Press, Princeton, New JerseyCrossRef
2.
go back to reference Cristianini N, Shawe-Taylor J (2000) An introduction to support vector machines and other kernel-based learning methods. Cambridge University Press, CambridgeCrossRefMATH Cristianini N, Shawe-Taylor J (2000) An introduction to support vector machines and other kernel-based learning methods. Cambridge University Press, CambridgeCrossRefMATH
3.
4.
go back to reference Jolliffe IT (2002) Principal component analysis, 2nd edn. Springer-Verlag, New YorkMATH Jolliffe IT (2002) Principal component analysis, 2nd edn. Springer-Verlag, New YorkMATH
5.
go back to reference Turki T, Roshan U (2014) Weighted maximum variance dimensionality reduction. Pattern recognition. Springer International Publishing, New York Turki T, Roshan U (2014) Weighted maximum variance dimensionality reduction. Pattern recognition. Springer International Publishing, New York
6.
go back to reference Niyogi X (2004) Locality preserving projections. Neural Inform Process Systems 16:153 Niyogi X (2004) Locality preserving projections. Neural Inform Process Systems 16:153
7.
go back to reference He X, Cai D, Yan S et al (2005) Neighborhood preserving embedding. In: Tenth IEEE International Conference on Computer Vision, vol 2, pp 1208–1213 He X, Cai D, Yan S et al (2005) Neighborhood preserving embedding. In: Tenth IEEE International Conference on Computer Vision, vol 2, pp 1208–1213
8.
go back to reference Cai D, He X (2005) Orthogonal locality preserving indexing. In: Proceedings of the 28th annual international ACM SIGIR conference on Research and development in information retrieval. ACM, New York, pp 3–10 Cai D, He X (2005) Orthogonal locality preserving indexing. In: Proceedings of the 28th annual international ACM SIGIR conference on Research and development in information retrieval. ACM, New York, pp 3–10
9.
go back to reference Kokiopoulou E, Saad Y (2007) Orthogonal neighborhood preserving projections: a projection-based dimensionality reduction technique. Pattern Anal Mach Intell IEEE Trans 29(12):2143–2156CrossRef Kokiopoulou E, Saad Y (2007) Orthogonal neighborhood preserving projections: a projection-based dimensionality reduction technique. Pattern Anal Mach Intell IEEE Trans 29(12):2143–2156CrossRef
10.
go back to reference Tenenbaum JB, De Silva V, Langford JC (2000) A global geometric framework for nonlinear dimensionality reduction. Science 290(5500):2319–2323CrossRef Tenenbaum JB, De Silva V, Langford JC (2000) A global geometric framework for nonlinear dimensionality reduction. Science 290(5500):2319–2323CrossRef
11.
go back to reference Boutsidis C, Zouzias A, Mahoney MW et al (2015) Randomized dimensionality reduction for k-means clustering. IEEE Trans Inform Theory 61(2):1045MathSciNetCrossRef Boutsidis C, Zouzias A, Mahoney MW et al (2015) Randomized dimensionality reduction for k-means clustering. IEEE Trans Inform Theory 61(2):1045MathSciNetCrossRef
12.
go back to reference Wang SJ, Yan S, Yang J et al (2014) A general exponential framework for dimensionality reduction. Image Process IEEE Trans 23(2):920–930MathSciNetCrossRef Wang SJ, Yan S, Yang J et al (2014) A general exponential framework for dimensionality reduction. Image Process IEEE Trans 23(2):920–930MathSciNetCrossRef
13.
go back to reference Roweis ST, Saul LK (2000) Nonlinear dimensionality reduction by locally linear embedding. Science 290(5500):2323–2326CrossRef Roweis ST, Saul LK (2000) Nonlinear dimensionality reduction by locally linear embedding. Science 290(5500):2323–2326CrossRef
14.
go back to reference Belkin M, Niyogi P (2001) Laplacian eigenmaps and spectral techniques for embedding and clustering. Adv Neural Inform Process Syst 14:585–591 Belkin M, Niyogi P (2001) Laplacian eigenmaps and spectral techniques for embedding and clustering. Adv Neural Inform Process Syst 14:585–591
15.
go back to reference Zhang Z, Zha H (2004) Principal manifolds and nonlinear dimensionality reduction via tangent space alignment. J Shanghai Univ (English Edition) 8(4):406–424MathSciNetCrossRefMATH Zhang Z, Zha H (2004) Principal manifolds and nonlinear dimensionality reduction via tangent space alignment. J Shanghai Univ (English Edition) 8(4):406–424MathSciNetCrossRefMATH
16.
go back to reference Donoho DL, Grimes C (2003) Hessian eigenmaps: locally linear embedding techniques for high-dimensional data. Proc Natl Acad Sci 100(10):5591–5596MathSciNetCrossRefMATH Donoho DL, Grimes C (2003) Hessian eigenmaps: locally linear embedding techniques for high-dimensional data. Proc Natl Acad Sci 100(10):5591–5596MathSciNetCrossRefMATH
17.
go back to reference Weinberger KQ, Sha F, Saul LK (2004) Learning a kernel matrix for nonlinear dimensionality reduction. In: Proceedings of the twenty-first international conference on Machine learning. ACM, New York, p 106 Weinberger KQ, Sha F, Saul LK (2004) Learning a kernel matrix for nonlinear dimensionality reduction. In: Proceedings of the twenty-first international conference on Machine learning. ACM, New York, p 106
18.
go back to reference Vlassis N, Motomura Y, Kröse B (2002) Supervised dimension reduction of intrinsically low-dimensional data. Neural Comput 14(1):191–215CrossRefMATH Vlassis N, Motomura Y, Kröse B (2002) Supervised dimension reduction of intrinsically low-dimensional data. Neural Comput 14(1):191–215CrossRefMATH
19.
go back to reference Bengio Y, Paiement JF, Vincent P et al (2004) Out-of-sample extensions for lle, isomap, mds, eigenmaps, and spectral clustering. Adv Neural Inform Process Syst 16:177–184 Bengio Y, Paiement JF, Vincent P et al (2004) Out-of-sample extensions for lle, isomap, mds, eigenmaps, and spectral clustering. Adv Neural Inform Process Syst 16:177–184
20.
go back to reference Vlassis N, Motomura Y, Krose B, Fukunaga K (1990) Introduction to statistical pattern recognition. Academic press, San Diego Vlassis N, Motomura Y, Krose B, Fukunaga K (1990) Introduction to statistical pattern recognition. Academic press, San Diego
21.
go back to reference He X, Cai D, Min W (2005) Statistical and computational analysis of locality preserving projection. In: Proceedings of the 22nd international conference on Machine learning. ACM, New York, pp 281–288 He X, Cai D, Min W (2005) Statistical and computational analysis of locality preserving projection. In: Proceedings of the 22nd international conference on Machine learning. ACM, New York, pp 281–288
22.
go back to reference Albert R, Jeong H, Barabási AL (2000) Error and attack tolerance of complex networks. Nature 406(6794):378–382CrossRef Albert R, Jeong H, Barabási AL (2000) Error and attack tolerance of complex networks. Nature 406(6794):378–382CrossRef
23.
go back to reference White S, Smyth P (2003) Algorithms for estimating relative importance in networks. In: Proceedings of the ninth ACM SIGKDD international conference on Knowledge discovery and data mining. ACM, New York, pp 266–275 White S, Smyth P (2003) Algorithms for estimating relative importance in networks. In: Proceedings of the ninth ACM SIGKDD international conference on Knowledge discovery and data mining. ACM, New York, pp 266–275
24.
go back to reference Nieminen J (1974) On the centrality in a graph. Scand J Psychol 15(1):332–336CrossRef Nieminen J (1974) On the centrality in a graph. Scand J Psychol 15(1):332–336CrossRef
25.
go back to reference Freeman LC, Roeder D, Mulholland RR (1980) Centrality in social networks: II experimental results. Soc Netw 2(2):119–141CrossRef Freeman LC, Roeder D, Mulholland RR (1980) Centrality in social networks: II experimental results. Soc Netw 2(2):119–141CrossRef
26.
go back to reference Freeman LC (1977) A set of measures of centrality based on betweenness. Sociometry, pp 35–41 Freeman LC (1977) A set of measures of centrality based on betweenness. Sociometry, pp 35–41
27.
go back to reference Saul LK, Roweis ST (2003) Think globally, fit locally: unsupervised learning of low dimensional manifolds. J Mach Learn Res 4:119–155MathSciNetMATH Saul LK, Roweis ST (2003) Think globally, fit locally: unsupervised learning of low dimensional manifolds. J Mach Learn Res 4:119–155MathSciNetMATH
Metadata
Title
Critical data points-based unsupervised linear dimension reduction technology for science data
Authors
Di Wu
Naixue Xiong
Jinrong He
Chuanhe Huang
Publication date
01-08-2016
Publisher
Springer US
Published in
The Journal of Supercomputing / Issue 8/2016
Print ISSN: 0920-8542
Electronic ISSN: 1573-0484
DOI
https://doi.org/10.1007/s11227-015-1421-0

Other articles of this Issue 8/2016

The Journal of Supercomputing 8/2016 Go to the issue

Premium Partner