Skip to main content
Top
Published in: Neural Processing Letters 4/2023

01-12-2022

Active Learning by Extreme Learning Machine with Considering Exploration and Exploitation Simultaneously

Authors: Yan Gu, Hualong Yu, Xibei Yang, Shang Gao

Published in: Neural Processing Letters | Issue 4/2023

Log in

Activate our intelligent search to find suitable subject content or patents.

search-config
loading …

Abstract

As an important machine learning paradigm, active learning has been widely applied to scenarios in which it is easy to acquire a large number of instances but labeling them is expensive and/or time-consuming. In such scenario, active learning can significantly reduce the cost of labeling the instances. Extreme learning machine (ELM) is a popular supervised learning model that has the structure of a single-hidden-layer feed-forward network, and has such merits as low computational cost, high training speed, and high generalization ability. Previous studies have shown that the integration of active learning with the ELM can yield effective and efficient results. However, the currently used method of integration considers only the capability for exploitation neglecting that for exploration, further increasing the risk of the results falling into local optima in context of a cold start. To address this problem, we propose an improved algorithm called the AL-SNN-ELM in this paper. It contains two sub-procedures for a sequential query: The exploration strategy, which uses the shared nearest neighbor (SNN) clustering algorithm, takes charge of exploring the sample space to query representative instances, and the exploitation strategy is responsible for transforming the actual outputs of the ELM into posterior probabilities to query uncertain instances. That is to say, the exploration sub-procedure helps roughly locate the decision boundary for sound classification by observing the global distribution of the data, while the exploitation sub-procedure subtly tunes this decision boundary by observing the distribution of local instances surrounding it. In addition, to reduce the time-complexity of active learning, online-sequential extreme learning machine is also adopted to replace the traditional ELM. The results of experiments on 20 UCI benchmark datasets and two real-world datasets show that the proposed AL-SNN-ELM algorithm can yield a significant improvement in performance in comparison with the traditional AL-ELM algorithm, indicating that it is useful to consider the exploration and exploitation simultaneously in the framework of active learning.

Dont have a licence yet? Then find out more about our products and how to get one now:

Springer Professional "Wirtschaft+Technik"

Online-Abonnement

Mit Springer Professional "Wirtschaft+Technik" erhalten Sie Zugriff auf:

  • über 102.000 Bücher
  • über 537 Zeitschriften

aus folgenden Fachgebieten:

  • Automobil + Motoren
  • Bauwesen + Immobilien
  • Business IT + Informatik
  • Elektrotechnik + Elektronik
  • Energie + Nachhaltigkeit
  • Finance + Banking
  • Management + Führung
  • Marketing + Vertrieb
  • Maschinenbau + Werkstoffe
  • Versicherung + Risiko

Jetzt Wissensvorsprung sichern!

Springer Professional "Technik"

Online-Abonnement

Mit Springer Professional "Technik" erhalten Sie Zugriff auf:

  • über 67.000 Bücher
  • über 390 Zeitschriften

aus folgenden Fachgebieten:

  • Automobil + Motoren
  • Bauwesen + Immobilien
  • Business IT + Informatik
  • Elektrotechnik + Elektronik
  • Energie + Nachhaltigkeit
  • Maschinenbau + Werkstoffe




 

Jetzt Wissensvorsprung sichern!

Springer Professional "Wirtschaft"

Online-Abonnement

Mit Springer Professional "Wirtschaft" erhalten Sie Zugriff auf:

  • über 67.000 Bücher
  • über 340 Zeitschriften

aus folgenden Fachgebieten:

  • Bauwesen + Immobilien
  • Business IT + Informatik
  • Finance + Banking
  • Management + Führung
  • Marketing + Vertrieb
  • Versicherung + Risiko




Jetzt Wissensvorsprung sichern!

Literature
1.
go back to reference Chakraborty S, Balasubramanian V, Panchanathan S (2015) Adaptive batch mode active learning. IEEE Trans Neural Netw Learn Syst 26(8):1747–1760MathSciNetCrossRef Chakraborty S, Balasubramanian V, Panchanathan S (2015) Adaptive batch mode active learning. IEEE Trans Neural Netw Learn Syst 26(8):1747–1760MathSciNetCrossRef
2.
go back to reference Hazarika BB, Gupta D (2021) Density weighted twin support vector machines for binary class imbalance learning. Neural Process Lett 54(2):1091–1130CrossRef Hazarika BB, Gupta D (2021) Density weighted twin support vector machines for binary class imbalance learning. Neural Process Lett 54(2):1091–1130CrossRef
3.
go back to reference Hazarika BB, Gupta D (2020) Density-weighted support vector machines for binary class imbalance learning. Neural Comput Appl 33(9):4243–4261CrossRef Hazarika BB, Gupta D (2020) Density-weighted support vector machines for binary class imbalance learning. Neural Comput Appl 33(9):4243–4261CrossRef
4.
go back to reference Du B, Wang Z, Zhang L et al (2017) Exploring representativeness and informativeness for active learning. IEEE Trans Cybern 47(1):14–26CrossRef Du B, Wang Z, Zhang L et al (2017) Exploring representativeness and informativeness for active learning. IEEE Trans Cybern 47(1):14–26CrossRef
5.
go back to reference Settles B (2011) From theories to queries: active learning in practice. In: JMLR workshop and conference proceedings, vol 16, pp 1–18 Settles B (2011) From theories to queries: active learning in practice. In: JMLR workshop and conference proceedings, vol 16, pp 1–18
6.
go back to reference Yang Y, Loog M (2018) A variance maximization criterion for active learning. Pattern Recognit 78:358–370CrossRef Yang Y, Loog M (2018) A variance maximization criterion for active learning. Pattern Recognit 78:358–370CrossRef
7.
go back to reference Konyushkova K, Sznitman R, Fua P (2015) Introducing geometry in active learning for image segmentation. In: 2015 IEEE international conference on computer vision (ICCV), Santiago, Chile, pp 2974–2982 Konyushkova K, Sznitman R, Fua P (2015) Introducing geometry in active learning for image segmentation. In: 2015 IEEE international conference on computer vision (ICCV), Santiago, Chile, pp 2974–2982
8.
go back to reference Liu B, Ferrari V (2017) Active learning for human pose estimation. In: 2017 IEEE international conference on computer vision (ICCV), Venice, Italy, pp 4363–4372 Liu B, Ferrari V (2017) Active learning for human pose estimation. In: 2017 IEEE international conference on computer vision (ICCV), Venice, Italy, pp 4363–4372
9.
go back to reference She Q, Chen K, Luo Z et al (2020) Double-criteria active learning for multiclass brain–computer interfaces. Comput Intell Neurosci 2020:1–13CrossRef She Q, Chen K, Luo Z et al (2020) Double-criteria active learning for multiclass brain–computer interfaces. Comput Intell Neurosci 2020:1–13CrossRef
10.
go back to reference Malhotra K, Bansal S, Ganapathy S (2019) Active learning methods for low resource end-to-end speech recognition. In: Interspeech, Graz, Austria, pp 2215–2219 Malhotra K, Bansal S, Ganapathy S (2019) Active learning methods for low resource end-to-end speech recognition. In: Interspeech, Graz, Austria, pp 2215–2219
11.
go back to reference Han X, Kwoh CK, Kim J (2016) Clustering based active learning for biomedical named entity recognition. In: 2016 International joint conference on neural networks (IJCNN), Vancouver, BC, Canada, pp 1253–1260 Han X, Kwoh CK, Kim J (2016) Clustering based active learning for biomedical named entity recognition. In: 2016 International joint conference on neural networks (IJCNN), Vancouver, BC, Canada, pp 1253–1260
12.
go back to reference Flores CA, Figueroa RL, Pezoa JE (2021) Active learning for biomedical text classification based on automatically generated regular expressions. IEEE Access 9:38767–38777CrossRef Flores CA, Figueroa RL, Pezoa JE (2021) Active learning for biomedical text classification based on automatically generated regular expressions. IEEE Access 9:38767–38777CrossRef
14.
go back to reference Lughofer E, Pratama M (2018) Online active learning in data stream regression using uncertainty sampling based on evolving generalized fuzzy models. IEEE Trans Fuzzy Syst 26(1):292–309CrossRef Lughofer E, Pratama M (2018) Online active learning in data stream regression using uncertainty sampling based on evolving generalized fuzzy models. IEEE Trans Fuzzy Syst 26(1):292–309CrossRef
15.
go back to reference Wang G, Hwang JN, Rose C, Wallace F (2019) Uncertainty-based active learning via sparse modeling for image classification. IEEE Trans Image Process 28(1):316–329MathSciNetCrossRefMATH Wang G, Hwang JN, Rose C, Wallace F (2019) Uncertainty-based active learning via sparse modeling for image classification. IEEE Trans Image Process 28(1):316–329MathSciNetCrossRefMATH
16.
go back to reference Wang R, Kwong S, Chen D (2012) Inconsistency-based active learning for support vector machines. Pattern Recognit 45(10):3751–3767CrossRef Wang R, Kwong S, Chen D (2012) Inconsistency-based active learning for support vector machines. Pattern Recognit 45(10):3751–3767CrossRef
17.
go back to reference Yu G, Yang Y, Wang X et al (2020) Adversarial active learning for the identification of medical concepts and annotation inconsistency. J Biomed Inform 108:103481CrossRef Yu G, Yang Y, Wang X et al (2020) Adversarial active learning for the identification of medical concepts and annotation inconsistency. J Biomed Inform 108:103481CrossRef
18.
go back to reference Smith JS, Nebgen B, Lubbers N et al (2018) Less is more: sampling chemical space with active learning. J Chem Phys 148(24):241733CrossRef Smith JS, Nebgen B, Lubbers N et al (2018) Less is more: sampling chemical space with active learning. J Chem Phys 148(24):241733CrossRef
19.
go back to reference Settles B, Craven M (2008) An analysis of active learning strategies for sequence labeling tasks. In: Proceedings of the conference on empirical methods in natural language processing (EMNLP), Honolulu, Hawaii, USA, pp 1070–1079 Settles B, Craven M (2008) An analysis of active learning strategies for sequence labeling tasks. In: Proceedings of the conference on empirical methods in natural language processing (EMNLP), Honolulu, Hawaii, USA, pp 1070–1079
20.
go back to reference Mingkun Li, Sethi IK (2006) Confidence-based active learning. IEEE Trans Pattern Anal Mach Intell 28(8):1251–1261CrossRef Mingkun Li, Sethi IK (2006) Confidence-based active learning. IEEE Trans Pattern Anal Mach Intell 28(8):1251–1261CrossRef
21.
go back to reference Roy N, McCallum A (2001) Toward optimal active learning through monte carlo estimation of error reduction. In: Proceedings of the international conference on machine learning (ICML), Williamstown, MA, USA, vol 2, pp 441–448 Roy N, McCallum A (2001) Toward optimal active learning through monte carlo estimation of error reduction. In: Proceedings of the international conference on machine learning (ICML), Williamstown, MA, USA, vol 2, pp 441–448
22.
go back to reference Ling C, Lu Z, Zhu X (2019) Efficient methods by active learning kriging coupled with variance reduction based sampling methods for time-dependent failure probability. Reliab Eng Syst Saf 188:23–35CrossRef Ling C, Lu Z, Zhu X (2019) Efficient methods by active learning kriging coupled with variance reduction based sampling methods for time-dependent failure probability. Reliab Eng Syst Saf 188:23–35CrossRef
23.
go back to reference Yang Y, Loog M (2018) A variance maximization criterion for active learning. Pattern Recognit 78:358–370CrossRef Yang Y, Loog M (2018) A variance maximization criterion for active learning. Pattern Recognit 78:358–370CrossRef
24.
go back to reference Huang SJ, Jin R, Zhou ZH (2014) Active learning by querying informative and representative examples. IEEE Trans Pattern Anal Mach Intell 36(10):1936–1949CrossRef Huang SJ, Jin R, Zhou ZH (2014) Active learning by querying informative and representative examples. IEEE Trans Pattern Anal Mach Intell 36(10):1936–1949CrossRef
25.
go back to reference Yang Y, Ma Z, Nie F et al (2014) Multi-class active learning by uncertainty sampling with diversity maximization. Int J Comput Vis 113(2):113–127MathSciNetCrossRef Yang Y, Ma Z, Nie F et al (2014) Multi-class active learning by uncertainty sampling with diversity maximization. Int J Comput Vis 113(2):113–127MathSciNetCrossRef
26.
go back to reference Wang M, Min F, Zhang ZH, Wu YX (2017) Active learning through density clustering. Expert Syst Appl 85:305–317CrossRef Wang M, Min F, Zhang ZH, Wu YX (2017) Active learning through density clustering. Expert Syst Appl 85:305–317CrossRef
27.
go back to reference He D, Yu H, Wang G, Li J (2021) A two-stage clustering-based cold-start method for active learning. Intell Data Anal 25(5):1169–1185CrossRef He D, Yu H, Wang G, Li J (2021) A two-stage clustering-based cold-start method for active learning. Intell Data Anal 25(5):1169–1185CrossRef
28.
go back to reference Reitmaier T, Calma A, Sick B (2015) Transductive active learning: a new semi-supervised learning approach based on iteratively refined generative models to capture structure in data. Inf Sci 293:275–298CrossRef Reitmaier T, Calma A, Sick B (2015) Transductive active learning: a new semi-supervised learning approach based on iteratively refined generative models to capture structure in data. Inf Sci 293:275–298CrossRef
29.
go back to reference Yu K, Bi J, Tresp V (2006) Active learning via transductive experimental design. In: Proceedings of the 23rd international conference on machine learning (ICML), Pittsburgh, Pennsylvania, pp 1081–1088 Yu K, Bi J, Tresp V (2006) Active learning via transductive experimental design. In: Proceedings of the 23rd international conference on machine learning (ICML), Pittsburgh, Pennsylvania, pp 1081–1088
30.
go back to reference Yang Y, Yin X, Zhao Y et al (2021) Batch mode active learning based on multi-set clustering. IEEE Access 9:51452–51463CrossRef Yang Y, Yin X, Zhao Y et al (2021) Batch mode active learning based on multi-set clustering. IEEE Access 9:51452–51463CrossRef
31.
go back to reference Chen DW, Jin YH (2020) An active learning algorithm based on Shannon entropy for constraint-based clustering. IEEE Access 8:171447–171456CrossRef Chen DW, Jin YH (2020) An active learning algorithm based on Shannon entropy for constraint-based clustering. IEEE Access 8:171447–171456CrossRef
32.
go back to reference Huang GB, Zhu QY, Siew CK (2006) Extreme learning machine: theory and applications. Neurocomputing 70(1–3):489–501CrossRef Huang GB, Zhu QY, Siew CK (2006) Extreme learning machine: theory and applications. Neurocomputing 70(1–3):489–501CrossRef
33.
go back to reference Huang GB, Wang DH, Lan Y (2011) Extreme learning machines: a survey. Int J Mach Learn Cybern 2(2):107–122CrossRef Huang GB, Wang DH, Lan Y (2011) Extreme learning machines: a survey. Int J Mach Learn Cybern 2(2):107–122CrossRef
34.
go back to reference Huang GB, Zhou HM, Ding XJ, Zhang R (2012) Extreme learning machine for regression and multiclass classification. IEEE Trans Syst Man Cybern B 42(2):513–529CrossRef Huang GB, Zhou HM, Ding XJ, Zhang R (2012) Extreme learning machine for regression and multiclass classification. IEEE Trans Syst Man Cybern B 42(2):513–529CrossRef
35.
go back to reference Borah P, Gupta D (2020) Unconstrained convex minimization based implicit lagrangian twin extreme learning machine for classification (ULTELMC). Appl Intell 50(4):1327–1344CrossRef Borah P, Gupta D (2020) Unconstrained convex minimization based implicit lagrangian twin extreme learning machine for classification (ULTELMC). Appl Intell 50(4):1327–1344CrossRef
36.
go back to reference Hazarika BB, Gupta D, Berlin M (2020) A Coiflet LDMR and coiflet OB-elm for river suspended sediment load prediction. Int J Environ Sci Technol 18(9):2675–2692CrossRef Hazarika BB, Gupta D, Berlin M (2020) A Coiflet LDMR and coiflet OB-elm for river suspended sediment load prediction. Int J Environ Sci Technol 18(9):2675–2692CrossRef
37.
go back to reference Yu H, Sun C, Yang W et al (2015) AL-ELM: one uncertainty-based active learning algorithm using extreme learning machine. Neurocomputing 166:140–150CrossRef Yu H, Sun C, Yang W et al (2015) AL-ELM: one uncertainty-based active learning algorithm using extreme learning machine. Neurocomputing 166:140–150CrossRef
38.
go back to reference Yu H, Yang X, Zheng S, Sun C (2019) Active learning from imbalanced data: a solution of online weighted extreme learning machine. IEEE Trans Neural Netw Learn Syst 30(4):1088–1103CrossRef Yu H, Yang X, Zheng S, Sun C (2019) Active learning from imbalanced data: a solution of online weighted extreme learning machine. IEEE Trans Neural Netw Learn Syst 30(4):1088–1103CrossRef
39.
go back to reference Qin J, Wang C, Zou Q et al (2021) Active learning with extreme learning machine for online imbalanced multiclass classification. Knowl Based Syst 231:107385CrossRef Qin J, Wang C, Zou Q et al (2021) Active learning with extreme learning machine for online imbalanced multiclass classification. Knowl Based Syst 231:107385CrossRef
40.
go back to reference Yoon J, Hwang SJ (2017) Combined group and exclusive sparsity for deep neural networks. In: Proceedings of international conference on machine learning (ICML), Sydney, NSW, Australia, vol 70, pp 3958–3966 Yoon J, Hwang SJ (2017) Combined group and exclusive sparsity for deep neural networks. In: Proceedings of international conference on machine learning (ICML), Sydney, NSW, Australia, vol 70, pp 3958–3966
41.
go back to reference Kumar V, Pujari AK, Padmanabhan V, Kagita VR (2019) Group preserving label embedding for multi-label classification. Pattern Recognit 90:23–34CrossRef Kumar V, Pujari AK, Padmanabhan V, Kagita VR (2019) Group preserving label embedding for multi-label classification. Pattern Recognit 90:23–34CrossRef
42.
go back to reference Ertöz L, Steinbach M, Kumar V (2003) Finding clusters of different sizes, shapes, and densities in noisy, high dimensional data. In: Proceedings of the 2003 SIAM international conference on data mining (SDM), San Francisco, CA, USA, pp 47–58 Ertöz L, Steinbach M, Kumar V (2003) Finding clusters of different sizes, shapes, and densities in noisy, high dimensional data. In: Proceedings of the 2003 SIAM international conference on data mining (SDM), San Francisco, CA, USA, pp 47–58
43.
go back to reference Liang NY, Huang GB, Saratchandran P, Sundararajan N (2006) A fast and accurate online sequential learning algorithm for feedforward networks. IEEE Trans Neural Netw Learn Syst 17(6):1411–1423CrossRef Liang NY, Huang GB, Saratchandran P, Sundararajan N (2006) A fast and accurate online sequential learning algorithm for feedforward networks. IEEE Trans Neural Netw Learn Syst 17(6):1411–1423CrossRef
44.
go back to reference Wang Z, Du B, Tu W et al (2021) Incorporating distribution matching into uncertainty for multiple kernel active learning. IEEE Trans Knowl Data Eng 33(1):128–142CrossRef Wang Z, Du B, Tu W et al (2021) Incorporating distribution matching into uncertainty for multiple kernel active learning. IEEE Trans Knowl Data Eng 33(1):128–142CrossRef
45.
go back to reference Jarvis RA, Patrick EA (1973) Clustering using a similarity measure based on shared near neighbors. IEEE Trans Comput C –22(11):1025–1034CrossRef Jarvis RA, Patrick EA (1973) Clustering using a similarity measure based on shared near neighbors. IEEE Trans Comput C –22(11):1025–1034CrossRef
46.
go back to reference Wang WT, Wu YL, Tang CY, Hor MK (2015) Adaptive density-based spatial clustering of applications with noise (DBSCAN) according to Data. In: 2015 International conference on machine learning and cybernetics (ICMLC), GuangDong, China, vol 1, pp 445–451 Wang WT, Wu YL, Tang CY, Hor MK (2015) Adaptive density-based spatial clustering of applications with noise (DBSCAN) according to Data. In: 2015 International conference on machine learning and cybernetics (ICMLC), GuangDong, China, vol 1, pp 445–451
47.
go back to reference Sawant K (2014) Adaptive methods for determining dbscan parameters. Int J Innov Sci Eng Technol 1(4):329–334 Sawant K (2014) Adaptive methods for determining dbscan parameters. Int J Innov Sci Eng Technol 1(4):329–334
48.
go back to reference Blake C, Keogh E, Merz CJ (1998) UCI repository of machine learning databases, Department of Information and Computer Science, University of California, Technical Report 213, Irvine, CA Blake C, Keogh E, Merz CJ (1998) UCI repository of machine learning databases, Department of Information and Computer Science, University of California, Technical Report 213, Irvine, CA
51.
go back to reference Xu Z, Yu K, Tresp V et al (2003) Representative sampling for text classification using support vector machines. In: European conference on information retrieval (ECIR), Berlin, Heidelberg, pp 393–407 Xu Z, Yu K, Tresp V et al (2003) Representative sampling for text classification using support vector machines. In: European conference on information retrieval (ECIR), Berlin, Heidelberg, pp 393–407
52.
go back to reference Zhang X, Delpha C, Diallo D (2020) Incipient fault detection and estimation based on Jensen–Shannon divergence in a data-driven approach. Signal Process 169:107410CrossRef Zhang X, Delpha C, Diallo D (2020) Incipient fault detection and estimation based on Jensen–Shannon divergence in a data-driven approach. Signal Process 169:107410CrossRef
53.
go back to reference Demšar J (2006) Statistical comparisons of classifiers over multiple data sets. J Mach Learn Res 7:1–30MathSciNetMATH Demšar J (2006) Statistical comparisons of classifiers over multiple data sets. J Mach Learn Res 7:1–30MathSciNetMATH
54.
go back to reference Garcia S, Herrera F (2008) An extension on "statistical comparisons of classifiers over multiple data sets" for all pairwise comparisons. J Mach Learn Res 9:12MATH Garcia S, Herrera F (2008) An extension on "statistical comparisons of classifiers over multiple data sets" for all pairwise comparisons. J Mach Learn Res 9:12MATH
55.
go back to reference García S, Fernández A, Luengo J, Herrera F (2010) Advanced nonparametric tests for multiple comparisons in the design of experiments in computational intelligence and data mining: experimental analysis of power. Inf Sci 180(10):2044–2064CrossRef García S, Fernández A, Luengo J, Herrera F (2010) Advanced nonparametric tests for multiple comparisons in the design of experiments in computational intelligence and data mining: experimental analysis of power. Inf Sci 180(10):2044–2064CrossRef
Metadata
Title
Active Learning by Extreme Learning Machine with Considering Exploration and Exploitation Simultaneously
Authors
Yan Gu
Hualong Yu
Xibei Yang
Shang Gao
Publication date
01-12-2022
Publisher
Springer US
Published in
Neural Processing Letters / Issue 4/2023
Print ISSN: 1370-4621
Electronic ISSN: 1573-773X
DOI
https://doi.org/10.1007/s11063-022-11089-w

Other articles of this Issue 4/2023

Neural Processing Letters 4/2023 Go to the issue