Skip to main content

2013 | OriginalPaper | Buchkapitel

3. Graph-Based Visualisation of High Dimensional Data

verfasst von : Ágnes Vathy-Fogarassy, János Abonyi

Erschienen in: Graph-Based Clustering and Data Visualization Algorithms

Verlag: Springer London

Aktivieren Sie unsere intelligente Suche, um passende Fachinhalte oder Patente zu finden.

search-config
loading …

Abstract

In this chapter we give an overview of classical dimensionality reduction and graph based visualisation methods that are able to uncover hidden structure of high dimensional data and visualise it in a low-dimensional vector space.

Sie haben noch keine Lizenz? Dann Informieren Sie sich jetzt über unsere Produkte:

Springer Professional "Wirtschaft+Technik"

Online-Abonnement

Mit Springer Professional "Wirtschaft+Technik" erhalten Sie Zugriff auf:

  • über 102.000 Bücher
  • über 537 Zeitschriften

aus folgenden Fachgebieten:

  • Automobil + Motoren
  • Bauwesen + Immobilien
  • Business IT + Informatik
  • Elektrotechnik + Elektronik
  • Energie + Nachhaltigkeit
  • Finance + Banking
  • Management + Führung
  • Marketing + Vertrieb
  • Maschinenbau + Werkstoffe
  • Versicherung + Risiko

Jetzt Wissensvorsprung sichern!

Springer Professional "Technik"

Online-Abonnement

Mit Springer Professional "Technik" erhalten Sie Zugriff auf:

  • über 67.000 Bücher
  • über 390 Zeitschriften

aus folgenden Fachgebieten:

  • Automobil + Motoren
  • Bauwesen + Immobilien
  • Business IT + Informatik
  • Elektrotechnik + Elektronik
  • Energie + Nachhaltigkeit
  • Maschinenbau + Werkstoffe




 

Jetzt Wissensvorsprung sichern!

Springer Professional "Wirtschaft"

Online-Abonnement

Mit Springer Professional "Wirtschaft" erhalten Sie Zugriff auf:

  • über 67.000 Bücher
  • über 340 Zeitschriften

aus folgenden Fachgebieten:

  • Bauwesen + Immobilien
  • Business IT + Informatik
  • Finance + Banking
  • Management + Führung
  • Marketing + Vertrieb
  • Versicherung + Risiko




Jetzt Wissensvorsprung sichern!

Literatur
1.
Zurück zum Zitat Tukey, J.: Exploratory Data Analysis. Addison-Wesley, New York (1977)MATH Tukey, J.: Exploratory Data Analysis. Addison-Wesley, New York (1977)MATH
2.
Zurück zum Zitat McQueen, J.: Some methods for classification and analysis of multivariate observations. Proceedings of fifth Berkeley symposium on mathematical statistics and probability, pp. 281–297 (1967) McQueen, J.: Some methods for classification and analysis of multivariate observations. Proceedings of fifth Berkeley symposium on mathematical statistics and probability, pp. 281–297 (1967)
3.
Zurück zum Zitat Martinetz, T.M., Shulten, K.J.: A neural-gas network learns topologies. In: Kohonen, T., Mäkisara, K., Simula, O., Kangas, J. (eds.) Artificial Neural Networks, pp. 397–402. Elsevier Science Publishers B.V, North-Holland (1991) Martinetz, T.M., Shulten, K.J.: A neural-gas network learns topologies. In: Kohonen, T., Mäkisara, K., Simula, O., Kangas, J. (eds.) Artificial Neural Networks, pp. 397–402. Elsevier Science Publishers B.V, North-Holland (1991)
4.
Zurück zum Zitat Johannes, M., Brase, J.C., Fröhlich, H., Gade, S., Gehrmann, M., Fälth, M., Sültmann, H., Beißbarth, T.: Integration of pathway knowledge into a reweighted recursive feature elimination approach for risk stratification of cancer patients. Bioinformatics 26(17), 2136–2144 (2010)CrossRef Johannes, M., Brase, J.C., Fröhlich, H., Gade, S., Gehrmann, M., Fälth, M., Sültmann, H., Beißbarth, T.: Integration of pathway knowledge into a reweighted recursive feature elimination approach for risk stratification of cancer patients. Bioinformatics 26(17), 2136–2144 (2010)CrossRef
5.
Zurück zum Zitat Lai, C., Reinders, M.J.T., Wessels, L.: Random subspace method for multivariate feature selection. Pattern Recognit. Lett. 27(10), 1067–1076 (2006)CrossRef Lai, C., Reinders, M.J.T., Wessels, L.: Random subspace method for multivariate feature selection. Pattern Recognit. Lett. 27(10), 1067–1076 (2006)CrossRef
6.
Zurück zum Zitat Nguyen, M.H., de la Torre, F.: Optimal feature selection for support vector machines. Pattern Recognit. 43(3), 584–591 (2010)MATHCrossRef Nguyen, M.H., de la Torre, F.: Optimal feature selection for support vector machines. Pattern Recognit. 43(3), 584–591 (2010)MATHCrossRef
7.
Zurück zum Zitat Rong, J., Li, G., Chen, Y.P.P.: Acoustic feature selection for automatic emotion recognition from speech. Inf. Process. Manag. 45(3), 315–328 (2009)CrossRef Rong, J., Li, G., Chen, Y.P.P.: Acoustic feature selection for automatic emotion recognition from speech. Inf. Process. Manag. 45(3), 315–328 (2009)CrossRef
8.
Zurück zum Zitat Tsang, I.W., Kocsor, A., Kwok, J.T.: Efficient kernel feature extraction for massive data sets. Proceedings of the 12th ACM SIGKDD international conference on Knowledge discovery and data mining, pp. 724–729 (2006) Tsang, I.W., Kocsor, A., Kwok, J.T.: Efficient kernel feature extraction for massive data sets. Proceedings of the 12th ACM SIGKDD international conference on Knowledge discovery and data mining, pp. 724–729 (2006)
9.
Zurück zum Zitat Wang, J., Zhang, B., Wang, S., Qi, M., Kong, J.: An adaptively weighted sub-pattern locality preserving projection for face recognition. J. Netw. Comput. Appl. 332(3), 323–332 (2010)CrossRef Wang, J., Zhang, B., Wang, S., Qi, M., Kong, J.: An adaptively weighted sub-pattern locality preserving projection for face recognition. J. Netw. Comput. Appl. 332(3), 323–332 (2010)CrossRef
10.
Zurück zum Zitat Blum, A.L., Langley, P.: Selection of relevant features and examples in machine learning. Artif. Intell. 97(12), 245271 (1997)MathSciNet Blum, A.L., Langley, P.: Selection of relevant features and examples in machine learning. Artif. Intell. 97(12), 245271 (1997)MathSciNet
11.
Zurück zum Zitat Guyon, I., Elisseeff, A.: An introduction to variable and feature selection. J. Mach. Learn. Res. 3, 1157–1182 (2003)MATH Guyon, I., Elisseeff, A.: An introduction to variable and feature selection. J. Mach. Learn. Res. 3, 1157–1182 (2003)MATH
12.
Zurück zum Zitat Jain, A., Zongker, D.: Feature selection: Evaluation, application, and small sample performance. IEEE Trans. Pattern Anal. Mach. Intell. 192, 153–158 (1997)CrossRef Jain, A., Zongker, D.: Feature selection: Evaluation, application, and small sample performance. IEEE Trans. Pattern Anal. Mach. Intell. 192, 153–158 (1997)CrossRef
13.
Zurück zum Zitat Weston, J., et al.: Feature selection for SVMs. In: Leen, T.K., Dietterich, T.G., Tresp, V. (eds.) Advances in Neural Information Processing Systems, vol. 13, pp. 668–674. The MIT Press, Cambride (2001) Weston, J., et al.: Feature selection for SVMs. In: Leen, T.K., Dietterich, T.G., Tresp, V. (eds.) Advances in Neural Information Processing Systems, vol. 13, pp. 668–674. The MIT Press, Cambride (2001)
14.
Zurück zum Zitat Narendra, P., Fukunaga, K.: A branch and bound algorithm for feature subset selection. IEEE Trans. Comput. C–26(9), 917–922 (1977)CrossRef Narendra, P., Fukunaga, K.: A branch and bound algorithm for feature subset selection. IEEE Trans. Comput. C–26(9), 917–922 (1977)CrossRef
15.
Zurück zum Zitat Pudil, P., Novovičová, J., Kittler, J.: Floating search methods in feature selection. Pattern Recognit. Lett. 15(1), 1119–1125 (1994)CrossRef Pudil, P., Novovičová, J., Kittler, J.: Floating search methods in feature selection. Pattern Recognit. Lett. 15(1), 1119–1125 (1994)CrossRef
16.
Zurück zum Zitat Hotelling, H.: Analysis of a complex of statistical variables into principal components. J. Educ. Psychol. 24, 417–441 (1933)CrossRef Hotelling, H.: Analysis of a complex of statistical variables into principal components. J. Educ. Psychol. 24, 417–441 (1933)CrossRef
17.
Zurück zum Zitat Jolliffe, T.: Principal Component Analysis. Springer, New York (1996) Jolliffe, T.: Principal Component Analysis. Springer, New York (1996)
18.
Zurück zum Zitat Sammon, J.W.: A non-linear mapping for data structure analysis. IEEE Trans. Comput. 18(5), 401–409 (1969)CrossRef Sammon, J.W.: A non-linear mapping for data structure analysis. IEEE Trans. Comput. 18(5), 401–409 (1969)CrossRef
19.
Zurück zum Zitat Tenenbaum, J.B., Silva, V., Langford, J.C.: A global geometric framework for nonlinear dimensionality reduction. Science 290, 2319–2323 (2000)CrossRef Tenenbaum, J.B., Silva, V., Langford, J.C.: A global geometric framework for nonlinear dimensionality reduction. Science 290, 2319–2323 (2000)CrossRef
20.
Zurück zum Zitat Comon, P.: Independent component analysis: a new concept? Signal Process. 36(3), 287–317 (1994)MATHCrossRef Comon, P.: Independent component analysis: a new concept? Signal Process. 36(3), 287–317 (1994)MATHCrossRef
21.
Zurück zum Zitat Fisher, R.A.: The use of multiple measurements in taxonomic problems. Ann. Eugen. 7, 179–188 (1936)CrossRef Fisher, R.A.: The use of multiple measurements in taxonomic problems. Ann. Eugen. 7, 179–188 (1936)CrossRef
24.
Zurück zum Zitat Roweis, S.T., Saul, L.K.: Nonlinear dimensionality reduction by locally linear embedding. Science 290, 2323–2326 (2000)CrossRef Roweis, S.T., Saul, L.K.: Nonlinear dimensionality reduction by locally linear embedding. Science 290, 2323–2326 (2000)CrossRef
25.
Zurück zum Zitat Saul, L.K., Roweis, S.T.: Think globally, fit locally: unsupervised learning of low dimensional manifolds. J. Mach. Learn. Res. 4, 119–155 (2003)MathSciNet Saul, L.K., Roweis, S.T.: Think globally, fit locally: unsupervised learning of low dimensional manifolds. J. Mach. Learn. Res. 4, 119–155 (2003)MathSciNet
26.
Zurück zum Zitat Belkin, M., Niyogi, P.: Laplacian eigenmaps for dimensionality reduction and data representation. Neural Comput. 15(6), 1373–1396 (2003)MATHCrossRef Belkin, M., Niyogi, P.: Laplacian eigenmaps for dimensionality reduction and data representation. Neural Comput. 15(6), 1373–1396 (2003)MATHCrossRef
27.
Zurück zum Zitat Borg, I.: Modern multidimensional scaling: theory and applications. Springer, New York (1977) Borg, I.: Modern multidimensional scaling: theory and applications. Springer, New York (1977)
28.
Zurück zum Zitat Kruskal, J.B., Carroll, J.D.: Geometrical models and badness-of-fit functions. In: Krishnaiah, R. (ed.) Multivariate Analysis II, vol. 2, pp. 639–671. Academic Press Pachuri, New York (1969) Kruskal, J.B., Carroll, J.D.: Geometrical models and badness-of-fit functions. In: Krishnaiah, R. (ed.) Multivariate Analysis II, vol. 2, pp. 639–671. Academic Press Pachuri, New York (1969)
29.
Zurück zum Zitat Kaski, S., Nikkilä, J., Oja, M., Venna, J., Törönen, J., Castrén, E.: Trustworthiness and metrics in visualizing similarity of gene expression. BMC Bioinformatics 4, 48 (2003)CrossRef Kaski, S., Nikkilä, J., Oja, M., Venna, J., Törönen, J., Castrén, E.: Trustworthiness and metrics in visualizing similarity of gene expression. BMC Bioinformatics 4, 48 (2003)CrossRef
30.
Zurück zum Zitat Venna, J., Kaski, S.: Local multidimensional scaling with controlled tradeoff between trustworthiness and continuity, In: Proceedings of the workshop on self-organizing maps, pp. 695–702 (2005) Venna, J., Kaski, S.: Local multidimensional scaling with controlled tradeoff between trustworthiness and continuity, In: Proceedings of the workshop on self-organizing maps, pp. 695–702 (2005)
31.
Zurück zum Zitat Venna, J., Kaski, S.: Local multidimensional scaling. Neural Netw. 19(6), 889–899 (2006)MATHCrossRef Venna, J., Kaski, S.: Local multidimensional scaling. Neural Netw. 19(6), 889–899 (2006)MATHCrossRef
32.
Zurück zum Zitat Kiviluoto, K.: Topology preservation in self-organizing maps. Proceedings of IEEE international conference on neural networks, pp. 294–299 (1996) Kiviluoto, K.: Topology preservation in self-organizing maps. Proceedings of IEEE international conference on neural networks, pp. 294–299 (1996)
33.
Zurück zum Zitat Bauer, H.U., Pawelzik, K.R.: Quantifying the neighborhood preservation of selforganizing feature maps. IEEE Trans. Neural Netw. 3(4), 570–579 (1992)CrossRef Bauer, H.U., Pawelzik, K.R.: Quantifying the neighborhood preservation of selforganizing feature maps. IEEE Trans. Neural Netw. 3(4), 570–579 (1992)CrossRef
34.
Zurück zum Zitat Duda, R.O., Hart, P.E., Stork, D.: Pattern classification. Wiley, New York (2000) Duda, R.O., Hart, P.E., Stork, D.: Pattern classification. Wiley, New York (2000)
35.
Zurück zum Zitat Mika, S., Schölkopf, B., Smola, A.J., Müller, K.-R., Scholz, M., Rätsch, G.: Kernel PCA and de-noising in feature spaces. In: Advances in neural information processing systems, vol. 11, Cambridge, USA (1999) Mika, S., Schölkopf, B., Smola, A.J., Müller, K.-R., Scholz, M., Rätsch, G.: Kernel PCA and de-noising in feature spaces. In: Advances in neural information processing systems, vol. 11, Cambridge, USA (1999)
36.
Zurück zum Zitat Schölkopf, B., Smola, A.J., Müller, K.-R.: Nonlinear component analysis as a kernel eigenvalue problem. Neural Comput. 10(5), 1299–1319 (1998)CrossRef Schölkopf, B., Smola, A.J., Müller, K.-R.: Nonlinear component analysis as a kernel eigenvalue problem. Neural Comput. 10(5), 1299–1319 (1998)CrossRef
37.
Zurück zum Zitat Mao, J., Jain, A.K.: Artifical neural networks for feature extraction and multivariate data projection. IEEE Trans. Neural Netw 6(2), 629–637 (1995) Mao, J., Jain, A.K.: Artifical neural networks for feature extraction and multivariate data projection. IEEE Trans. Neural Netw 6(2), 629–637 (1995)
38.
Zurück zum Zitat Pal, N.R., Eluri, V.K.: Two efficient connectionist schemes for structure preserving dimensionality reduction. IEEE Trans. Neural Netw. 9, 1143–1153 (1998)CrossRef Pal, N.R., Eluri, V.K.: Two efficient connectionist schemes for structure preserving dimensionality reduction. IEEE Trans. Neural Netw. 9, 1143–1153 (1998)CrossRef
39.
Zurück zum Zitat Young, G., Householder, A.S.: Discussion of a set of points in terms of their mutual distances. Psychometrika 3(1), 19–22 (1938)MATHCrossRef Young, G., Householder, A.S.: Discussion of a set of points in terms of their mutual distances. Psychometrika 3(1), 19–22 (1938)MATHCrossRef
40.
Zurück zum Zitat Naud, A.: Neural and statistical methods for the visualization of multidimensional data. Technical Science Katedra Metod Komputerowych Uniwersytet Mikoaja Kopernika w Toruniu (2001) Naud, A.: Neural and statistical methods for the visualization of multidimensional data. Technical Science Katedra Metod Komputerowych Uniwersytet Mikoaja Kopernika w Toruniu (2001)
41.
Zurück zum Zitat Kruskal, J.B.: Multidimensional scaling by optimizing goodness-of-fit to a nonmetric hypothesis. Psychometrika 29, 1–29 (1964)MathSciNetMATHCrossRef Kruskal, J.B.: Multidimensional scaling by optimizing goodness-of-fit to a nonmetric hypothesis. Psychometrika 29, 1–29 (1964)MathSciNetMATHCrossRef
43.
Zurück zum Zitat UC Irvine Machine Learning Repository www.ics.uci.edu/ mlearn/ Cited 15 Oct 2012 UC Irvine Machine Learning Repository www.ics.uci.edu/ mlearn/ Cited 15 Oct 2012
44.
Zurück zum Zitat Haykin, S.: Neural Networks: A Comprehensive Foundation. Prentice Hall, Upper Saddle River (1999)MATH Haykin, S.: Neural Networks: A Comprehensive Foundation. Prentice Hall, Upper Saddle River (1999)MATH
45.
Zurück zum Zitat Ultsch, A.: Self-organization neural networks for visualization and classification. In: Opitz, O., Lausen, B., Klar, R. (eds.) Information and Classification, pp. 307–313. Springer, Berlin (1993)CrossRef Ultsch, A.: Self-organization neural networks for visualization and classification. In: Opitz, O., Lausen, B., Klar, R. (eds.) Information and Classification, pp. 307–313. Springer, Berlin (1993)CrossRef
46.
Zurück zum Zitat Blackmore, J., Miikkulainen, R.: Incremental grid growing: encoding high-dimensional structure into a two-dimensional feature map. In: Proceedong on IEEE international conference on neural networks, vol. 1, pp. 450–455 (1993) Blackmore, J., Miikkulainen, R.: Incremental grid growing: encoding high-dimensional structure into a two-dimensional feature map. In: Proceedong on IEEE international conference on neural networks, vol. 1, pp. 450–455 (1993)
47.
Zurück zum Zitat Merkl, D., He, S.H., Dittenbach, M., Rauber, A.: Adaptive hierarchical incremental grid growing: an architecture for high-dimensional data visualization. In: Proceeding of the workshop on SOM, Advances in SOM, pp. 293–298 (2003) Merkl, D., He, S.H., Dittenbach, M., Rauber, A.: Adaptive hierarchical incremental grid growing: an architecture for high-dimensional data visualization. In: Proceeding of the workshop on SOM, Advances in SOM, pp. 293–298 (2003)
48.
Zurück zum Zitat Lee, J.A., Lendasse, A., Donckers, N., Verleysen, M.: A robust nonlinear projection method. Proceedings of ESANN’2000, 8th european symposium on artificial, neural networks, pp. 13–20 (2000) Lee, J.A., Lendasse, A., Donckers, N., Verleysen, M.: A robust nonlinear projection method. Proceedings of ESANN’2000, 8th european symposium on artificial, neural networks, pp. 13–20 (2000)
49.
Zurück zum Zitat Estévez, P.A., Figueroa, C.J.: Online data visualization using the neural gas network. Neural Netw. 19, 923–934 (2006)MATHCrossRef Estévez, P.A., Figueroa, C.J.: Online data visualization using the neural gas network. Neural Netw. 19, 923–934 (2006)MATHCrossRef
50.
Zurück zum Zitat Estévez, P.A., Chong, A.M., Held, C.M., Perez, C.A.: Nonlinear projection using geodesic distances and the neural gas network. Lect. Notes Comput. Sci. 4131, 464–473 (2006)CrossRef Estévez, P.A., Chong, A.M., Held, C.M., Perez, C.A.: Nonlinear projection using geodesic distances and the neural gas network. Lect. Notes Comput. Sci. 4131, 464–473 (2006)CrossRef
51.
Zurück zum Zitat Wu, Y., Chan, K.L.: An extended isomap algorithm for learning multi-class manifold. Proceeding of IEEE international conference on machine learning and, cybernetics (ICMLC2004), vol. 6, pp. 3429–3433 (2004) Wu, Y., Chan, K.L.: An extended isomap algorithm for learning multi-class manifold. Proceeding of IEEE international conference on machine learning and, cybernetics (ICMLC2004), vol. 6, pp. 3429–3433 (2004)
52.
Zurück zum Zitat Lee, J.A., Verleysen, M.: Nonlinear projection with the isotop method. Proceedings of ICANN’2002, international conference on artificial, neural networks, pp. 933–938 (2002) Lee, J.A., Verleysen, M.: Nonlinear projection with the isotop method. Proceedings of ICANN’2002, international conference on artificial, neural networks, pp. 933–938 (2002)
53.
Zurück zum Zitat Lee, J.A., Archambeau, C., Verleysen, M.: Locally linear embedding versus isotop. In: ESANN’2003 proceedings: European symposium on artificial neural networks Bruges (Belgium), pp. 527–534 (2003) Lee, J.A., Archambeau, C., Verleysen, M.: Locally linear embedding versus isotop. In: ESANN’2003 proceedings: European symposium on artificial neural networks Bruges (Belgium), pp. 527–534 (2003)
54.
Zurück zum Zitat Demartines, P., Herault, J.: Curvilinear component analysis: a self-organizing neural network for nonlinear mapping of data sets. IEEE Trans. Neural Netw. 8, 148–154 (1997)CrossRef Demartines, P., Herault, J.: Curvilinear component analysis: a self-organizing neural network for nonlinear mapping of data sets. IEEE Trans. Neural Netw. 8, 148–154 (1997)CrossRef
55.
Zurück zum Zitat Lee, J.A., Lendasse, A., Verleysen, M.: Curvilinear distance analysis versus isomap. Proceedings of ESANN’2002, 10th European symposium on artificial, neural networks, pp. 185–192 (2000) Lee, J.A., Lendasse, A., Verleysen, M.: Curvilinear distance analysis versus isomap. Proceedings of ESANN’2002, 10th European symposium on artificial, neural networks, pp. 185–192 (2000)
56.
Zurück zum Zitat Lee, J.A., Lendasse, A., Verleysen, M.: Nonlinear projection with curvilinear distances: isomap versus curvilinear distance analysis. Neurocomputing 57, 49–76 (2004)CrossRef Lee, J.A., Lendasse, A., Verleysen, M.: Nonlinear projection with curvilinear distances: isomap versus curvilinear distance analysis. Neurocomputing 57, 49–76 (2004)CrossRef
57.
Zurück zum Zitat Vathy-Fogarassy, A., Kiss, A., Abonyi, J.: Topology representing network map—a new tool for visualization of high-dimensional data. Trans. Comput. Sci. I. 4750, 61–84 (2008) (Springer) Vathy-Fogarassy, A., Kiss, A., Abonyi, J.: Topology representing network map—a new tool for visualization of high-dimensional data. Trans. Comput. Sci. I. 4750, 61–84 (2008) (Springer)
58.
Zurück zum Zitat Vathy-Fogarassy, A., Abonyi, J.: Local and global mappings of topology representing networks. Inf. Sci. 179, 3791–3803 (2009)MathSciNetCrossRef Vathy-Fogarassy, A., Abonyi, J.: Local and global mappings of topology representing networks. Inf. Sci. 179, 3791–3803 (2009)MathSciNetCrossRef
60.
Zurück zum Zitat Martinetz, T.M., Shulten, K.J.: Topology representing networks. Neural Netw. 7(3), 507–522 (1994)CrossRef Martinetz, T.M., Shulten, K.J.: Topology representing networks. Neural Netw. 7(3), 507–522 (1994)CrossRef
Metadaten
Titel
Graph-Based Visualisation of High Dimensional Data
verfasst von
Ágnes Vathy-Fogarassy
János Abonyi
Copyright-Jahr
2013
Verlag
Springer London
DOI
https://doi.org/10.1007/978-1-4471-5158-6_3