Skip to main content
Top
Published in: Soft Computing 6/2015

01-06-2015 | Methodologies and Application

Unsupervised nearest neighbor regression for dimensionality reduction

Author: Oliver Kramer

Published in: Soft Computing | Issue 6/2015

Log in

Activate our intelligent search to find suitable subject content or patents.

search-config
loading …

Abstract

Large numbers of high-dimensional patterns are collected in a variety of disciplines, from astronomy to bioinformatics. In this article, we present an approach to non-linear dimensionality reduction based on fitting nearest neighbor regression to the unsupervised regression framework for learning of low-dimensional manifolds. For each high-dimensional pattern, a low-dimensional latent point is generated. The dimensionality of the induced optimization problem grows with the number of patterns. To cope with the large solution space, an iterative solution construction scheme is proposed. In this paper, we introduce two strategies to embed high-dimensional data. First, the latent sorting approach allows embeddings in a one-dimensional latent space corresponding to a sorting of the high-dimensional patterns. Second, Gaussian embeddings randomly generate candidate positions based on sampling from the Gaussian distribution employing distances on data space as variances. Kernel functions increase the flexibility of the approach by mapping the patterns to feature spaces. We analyze and compare the algorithms experimentally on a set of test functions.

Dont have a licence yet? Then find out more about our products and how to get one now:

Springer Professional "Wirtschaft+Technik"

Online-Abonnement

Mit Springer Professional "Wirtschaft+Technik" erhalten Sie Zugriff auf:

  • über 102.000 Bücher
  • über 537 Zeitschriften

aus folgenden Fachgebieten:

  • Automobil + Motoren
  • Bauwesen + Immobilien
  • Business IT + Informatik
  • Elektrotechnik + Elektronik
  • Energie + Nachhaltigkeit
  • Finance + Banking
  • Management + Führung
  • Marketing + Vertrieb
  • Maschinenbau + Werkstoffe
  • Versicherung + Risiko

Jetzt Wissensvorsprung sichern!

Springer Professional "Wirtschaft"

Online-Abonnement

Mit Springer Professional "Wirtschaft" erhalten Sie Zugriff auf:

  • über 67.000 Bücher
  • über 340 Zeitschriften

aus folgenden Fachgebieten:

  • Bauwesen + Immobilien
  • Business IT + Informatik
  • Finance + Banking
  • Management + Führung
  • Marketing + Vertrieb
  • Versicherung + Risiko




Jetzt Wissensvorsprung sichern!

Springer Professional "Technik"

Online-Abonnement

Mit Springer Professional "Technik" erhalten Sie Zugriff auf:

  • über 67.000 Bücher
  • über 390 Zeitschriften

aus folgenden Fachgebieten:

  • Automobil + Motoren
  • Bauwesen + Immobilien
  • Business IT + Informatik
  • Elektrotechnik + Elektronik
  • Energie + Nachhaltigkeit
  • Maschinenbau + Werkstoffe




 

Jetzt Wissensvorsprung sichern!

Appendix
Available only for authorised users
Literature
go back to reference Baillard A, Bertin E, de Lapparent V, Fouqué P, Arnouts S, Mellier Y, Pelló R, Leborgne J-F, Prugniel P, Markarov D, Makarova L, McCracken HJ, Bijaoui A, Tasca L (2011) Galaxy morphology without classification: self-organizing maps 532(A74):1103.5734 Baillard A, Bertin E, de Lapparent V, Fouqué P, Arnouts S, Mellier Y, Pelló R, Leborgne J-F, Prugniel P, Markarov D, Makarova L, McCracken HJ, Bijaoui A, Tasca L (2011) Galaxy morphology without classification: self-organizing maps 532(A74):1103.5734
go back to reference Bhatia N, Vandana A (2010) Survey of nearest neighbor techniques. Int J Comp Sci Inf Secur 8(2):302–305 Bhatia N, Vandana A (2010) Survey of nearest neighbor techniques. Int J Comp Sci Inf Secur 8(2):302–305
go back to reference Bishop CM (2007) Pattern recognition and machine learning (information science and statistics). Springer, New York Bishop CM (2007) Pattern recognition and machine learning (information science and statistics). Springer, New York
go back to reference Comon P (1994) Independent component analysis, a new concept? Signal Process 36(3):287–314CrossRefMATH Comon P (1994) Independent component analysis, a new concept? Signal Process 36(3):287–314CrossRefMATH
go back to reference Fix E, Hodges J (1951) Discriminatory analysis, nonparametric discrimination: consistency properties. USAF School of Aviation Medicine, Randolph Field, Texas Fix E, Hodges J (1951) Discriminatory analysis, nonparametric discrimination: consistency properties. USAF School of Aviation Medicine, Randolph Field, Texas
go back to reference Friedman JH, Tukey JW (2006) A projection pursuit algorithm for exploratory data analysis. IEEE Trans Comput C–23(9):881–890 Friedman JH, Tukey JW (2006) A projection pursuit algorithm for exploratory data analysis. IEEE Trans Comput C–23(9):881–890
go back to reference Gieseke F, Polsterer KL, Thom A, Zinn P, Bomanns D, Dettmar R-J, Kramer O, Vahrenhold J (2010) Detecting quasars in large-scale astronomical surveys. In: ICMLA, pp 352–357 Gieseke F, Polsterer KL, Thom A, Zinn P, Bomanns D, Dettmar R-J, Kramer O, Vahrenhold J (2010) Detecting quasars in large-scale astronomical surveys. In: ICMLA, pp 352–357
go back to reference Harrison D, Rubinfeld D (1978) Hedonic prices and the demand for clean air. J Environ Econ Manag 5:81–102CrossRefMATH Harrison D, Rubinfeld D (1978) Hedonic prices and the demand for clean air. J Environ Econ Manag 5:81–102CrossRefMATH
go back to reference Hastie T, Tibshirani R, Friedman J (2009) The elements of statistical learning. Springer, BerlinCrossRefMATH Hastie T, Tibshirani R, Friedman J (2009) The elements of statistical learning. Springer, BerlinCrossRefMATH
go back to reference Huber PJ (1985) Projection pursuit. Ann Statist 2(13):435–475 Huber PJ (1985) Projection pursuit. Ann Statist 2(13):435–475
go back to reference Hull J (1994) A database for handwritten text recognition research. IEEE PAMI 5(16):550–554CrossRef Hull J (1994) A database for handwritten text recognition research. IEEE PAMI 5(16):550–554CrossRef
go back to reference Jolliffe I (1986) Principal component analysis. In: Springer series in statistics. Springer, New York Jolliffe I (1986) Principal component analysis. In: Springer series in statistics. Springer, New York
go back to reference Kitchin C (2007) Galaxies in Turmoil—the active and starburst galaxies and the black holes that drive them. Springer, New York Kitchin C (2007) Galaxies in Turmoil—the active and starburst galaxies and the black holes that drive them. Springer, New York
go back to reference Klanke S, Ritter H (2007) Variants of unsupervised kernel regression: general cost functions. Neurocomputing 70(7–9):1289–1303 Klanke S, Ritter H (2007) Variants of unsupervised kernel regression: general cost functions. Neurocomputing 70(7–9):1289–1303
go back to reference Kramer O (2011) Dimensionalty reduction by unsupervised nearest neighbor regression. In: Proceedings of the 10th international conference on machine learning and applications (ICMLA) (IEEE, to appear) Kramer O (2011) Dimensionalty reduction by unsupervised nearest neighbor regression. In: Proceedings of the 10th international conference on machine learning and applications (ICMLA) (IEEE, to appear)
go back to reference Kramer O (2012a) A particle swarm embedding algorithm for nonlinear dimensionality reduction. In: International conference on swarm intelligence (ANTS), pp 1–12 Kramer O (2012a) A particle swarm embedding algorithm for nonlinear dimensionality reduction. In: International conference on swarm intelligence (ANTS), pp 1–12
go back to reference Kramer O (2012b) On unsupervised nearest-neighbor regression and robust loss functions. In: International conference on artificial intelligence (page to appear) Kramer O (2012b) On unsupervised nearest-neighbor regression and robust loss functions. In: International conference on artificial intelligence (page to appear)
go back to reference Kramer O (2012c) Unsupervised nearest neighbors with kernels. In: Advances in artificial intelligence (KI). Lecture notes in artificial intelligence. Springer, Saarbrücken (page to appear) Kramer O (2012c) Unsupervised nearest neighbors with kernels. In: Advances in artificial intelligence (KI). Lecture notes in artificial intelligence. Springer, Saarbrücken (page to appear)
go back to reference Kruskal JB (1964) Nonmetric multidimensional scaling: a numerical method. Psychometrika 29(2) Kruskal JB (1964) Nonmetric multidimensional scaling: a numerical method. Psychometrika 29(2)
go back to reference Lee JA, Verleysen M (2007) Nonlinear dimensionality reduction. Springer, New York Lee JA, Verleysen M (2007) Nonlinear dimensionality reduction. Springer, New York
go back to reference Lee JA, Verleysen M (2009) Quality assessment of dimensionality reduction: rank-based criteria. Neurocomputing 72(7–9):1431–1443CrossRef Lee JA, Verleysen M (2009) Quality assessment of dimensionality reduction: rank-based criteria. Neurocomputing 72(7–9):1431–1443CrossRef
go back to reference Omohundro SM (1989) Five balltree construction algorithms. Technical report. International Computer Science Institute (ICSI), Berkeley Omohundro SM (1989) Five balltree construction algorithms. Technical report. International Computer Science Institute (ICSI), Berkeley
go back to reference Pearson K (1901) On lines and planes of closest fit to systems of points in space. Philos Mag 2(6):559–572CrossRef Pearson K (1901) On lines and planes of closest fit to systems of points in space. Philos Mag 2(6):559–572CrossRef
go back to reference Roweis ST, Saul LK (2000) Nonlinear dimensionality reduction by locally linear embedding. Science 290:2323–2326CrossRef Roweis ST, Saul LK (2000) Nonlinear dimensionality reduction by locally linear embedding. Science 290:2323–2326CrossRef
go back to reference Schölkopf B, Smola AJ (2001) Learning with kernels: support vector machines, regularization, optimization, and beyond. MIT Press, Cambridge Schölkopf B, Smola AJ (2001) Learning with kernels: support vector machines, regularization, optimization, and beyond. MIT Press, Cambridge
go back to reference Suykens JAK, Vandewalle J (1999) Least squares support vector machine classifiers. Neural Process Lett 9(3):293–300CrossRefMathSciNet Suykens JAK, Vandewalle J (1999) Least squares support vector machine classifiers. Neural Process Lett 9(3):293–300CrossRefMathSciNet
go back to reference Tenenbaum JB, Silva VD, Langford JC (2000) A global geometric framework for nonlinear dimensionality reduction. Science 290:2319–2323CrossRef Tenenbaum JB, Silva VD, Langford JC (2000) A global geometric framework for nonlinear dimensionality reduction. Science 290:2319–2323CrossRef
Metadata
Title
Unsupervised nearest neighbor regression for dimensionality reduction
Author
Oliver Kramer
Publication date
01-06-2015
Publisher
Springer Berlin Heidelberg
Published in
Soft Computing / Issue 6/2015
Print ISSN: 1432-7643
Electronic ISSN: 1433-7479
DOI
https://doi.org/10.1007/s00500-014-1354-1

Other articles of this Issue 6/2015

Soft Computing 6/2015 Go to the issue

Premium Partner