Skip to main content
Erschienen in: Neural Computing and Applications 24/2022

29.07.2022 | Original Article

P + FELU: Flexible and trainable fast exponential linear unit for deep learning architectures

verfasst von: Kemal Adem

Erschienen in: Neural Computing and Applications | Ausgabe 24/2022

Einloggen

Aktivieren Sie unsere intelligente Suche, um passende Fachinhalte oder Patente zu finden.

search-config
loading …

Abstract

Activation functions have an important role in obtaining the most appropriate output by processing the information coming into the network in deep learning architectures. Deep learning architectures are widely used in areas such as image processing applications, time series, and disease classification, generally in line with the analysis of large and complex data. Choosing the appropriate architecture and activation function is an important factor in achieving successful learning and classification performance. There are many studies to improve the performance of deep learning architectures and to overcome the disappearing gradient and negative region problems in activation functions. A flexible and trainable fast exponential linear unit (P + FELU) activation function is proposed to overcome existing problems. With the proposed P + FELU activation function, a higher success rate and faster calculation time can be achieved by incorporating the advantages of fast exponentially linear unit (FELU), exponential linear unit (ELU), and rectified linear unit (RELU) activation functions. Performance evaluations of the proposed P + FELU activation function were made on MNIST, CIFAR-10, and CIFAR-100 benchmark datasets. Experimental evaluations have shown that the proposed activation function outperforms the ReLU, ELU, SELU, MPELU, TReLU, and FELU activation functions and effectively improves the noise robustness of the network. Experimental results show that this activation function with “flexible and trainable” properties can effectively prevent vanishing gradient and make multilayer perceptron neural networks deeper.

Sie haben noch keine Lizenz? Dann Informieren Sie sich jetzt über unsere Produkte:

Springer Professional "Wirtschaft"

Online-Abonnement

Mit Springer Professional "Wirtschaft" erhalten Sie Zugriff auf:

  • über 67.000 Bücher
  • über 340 Zeitschriften

aus folgenden Fachgebieten:

  • Bauwesen + Immobilien
  • Business IT + Informatik
  • Finance + Banking
  • Management + Führung
  • Marketing + Vertrieb
  • Versicherung + Risiko




Jetzt Wissensvorsprung sichern!

Springer Professional "Technik"

Online-Abonnement

Mit Springer Professional "Technik" erhalten Sie Zugriff auf:

  • über 67.000 Bücher
  • über 390 Zeitschriften

aus folgenden Fachgebieten:

  • Automobil + Motoren
  • Bauwesen + Immobilien
  • Business IT + Informatik
  • Elektrotechnik + Elektronik
  • Energie + Nachhaltigkeit
  • Maschinenbau + Werkstoffe




 

Jetzt Wissensvorsprung sichern!

Springer Professional "Wirtschaft+Technik"

Online-Abonnement

Mit Springer Professional "Wirtschaft+Technik" erhalten Sie Zugriff auf:

  • über 102.000 Bücher
  • über 537 Zeitschriften

aus folgenden Fachgebieten:

  • Automobil + Motoren
  • Bauwesen + Immobilien
  • Business IT + Informatik
  • Elektrotechnik + Elektronik
  • Energie + Nachhaltigkeit
  • Finance + Banking
  • Management + Führung
  • Marketing + Vertrieb
  • Maschinenbau + Werkstoffe
  • Versicherung + Risiko

Jetzt Wissensvorsprung sichern!

Literatur
1.
Zurück zum Zitat Adem K (2018) Exudate detection for diabetic retinopathy with circular Hough transformation and convolutional neural networks. Expert Syst Appl 114:289–295CrossRef Adem K (2018) Exudate detection for diabetic retinopathy with circular Hough transformation and convolutional neural networks. Expert Syst Appl 114:289–295CrossRef
2.
Zurück zum Zitat Adem K, Közkurt C (2019) Defect detection of seals in multilayer aseptic packages using deep learning. Turk J Electr Eng Comput Sci 27(6):4220–4230CrossRef Adem K, Közkurt C (2019) Defect detection of seals in multilayer aseptic packages using deep learning. Turk J Electr Eng Comput Sci 27(6):4220–4230CrossRef
3.
Zurück zum Zitat Bawa VS, Kumar V (2019) Linearized sigmoidal activation: A novel activation function with tractable non-linear characteristics to boost representation capability. Expert Syst Appl 120:346–356CrossRef Bawa VS, Kumar V (2019) Linearized sigmoidal activation: A novel activation function with tractable non-linear characteristics to boost representation capability. Expert Syst Appl 120:346–356CrossRef
4.
Zurück zum Zitat Clevert, D. A., Unterthiner, T., & Hochreiter, S. (2015). Fast and accurate deep network learning by exponential linear units (Elus). Clevert, D. A., Unterthiner, T., & Hochreiter, S. (2015). Fast and accurate deep network learning by exponential linear units (Elus). 
5.
Zurück zum Zitat Gao H, Xu K, Cao M, Xiao J, Xu Q, Yin Y (2021) The deep features and attention mechanism-based method to dish healthcare under social IoT systems: an empirical study with a hand-deep local-global net. IEEE Transact Comput Soc Syst 9(1):336–347CrossRef Gao H, Xu K, Cao M, Xiao J, Xu Q, Yin Y (2021) The deep features and attention mechanism-based method to dish healthcare under social IoT systems: an empirical study with a hand-deep local-global net. IEEE Transact Comput Soc Syst 9(1):336–347CrossRef
6.
Zurück zum Zitat Glorot, X., & Bengio, Y. (2010, March). Understanding the difficulty of training deep feedforward neural networks. In Proceedings of the thirteenth int conf on art intell stat JMLR Workshop and Conference Proceedings: 249–256) Glorot, X., & Bengio, Y. (2010, March). Understanding the difficulty of training deep feedforward neural networks. In Proceedings of the thirteenth int conf on art intell stat JMLR Workshop and Conference Proceedings: 249–256)
7.
Zurück zum Zitat Godfrey LB (2019) An evaluation of parametric activation functions for deep learning. In IEEE Int Conf Syst, Man Cybern (SMC) IEEE, pp 3006–3011 Godfrey LB (2019) An evaluation of parametric activation functions for deep learning. In IEEE Int Conf Syst, Man Cybern (SMC) IEEE, pp 3006–3011
8.
Zurück zum Zitat Godin F, Degrave J, Dambre J, De Neve W (2018) Dual rectified linear units (DReLUs): A replacement for tanh activation functions in quasi-recurrent neural networks. Pattern Recogn Lett 116:8–14CrossRef Godin F, Degrave J, Dambre J, De Neve W (2018) Dual rectified linear units (DReLUs): A replacement for tanh activation functions in quasi-recurrent neural networks. Pattern Recogn Lett 116:8–14CrossRef
9.
Zurück zum Zitat Gupta, S., & Dinesh, D. A. (2017). Resource usage prediction of cloud workloads using deep bidirectional long short term memory networks. In 2017 IEEE international conference on advanced networks and telecommunications systems (ANTS): 1–6 IEEE. Gupta, S., & Dinesh, D. A. (2017). Resource usage prediction of cloud workloads using deep bidirectional long short term memory networks. In 2017 IEEE international conference on advanced networks and telecommunications systems (ANTS): 1–6 IEEE.
10.
Zurück zum Zitat He, K., Zhang, X., Ren, S., & Sun, J. (2015). Delving deep into rectifiers: Surpassing human-level performance on imagenet classification. In Proceedings of the IEEE international conference on computer vision: 1026–1034 He, K., Zhang, X., Ren, S., & Sun, J. (2015). Delving deep into rectifiers: Surpassing human-level performance on imagenet classification. In Proceedings of the IEEE international conference on computer vision: 1026–1034
11.
Zurück zum Zitat Huizhen ZHAO, Fuxian L, Longyue L (2018) A novel softplus linear unit for deep CNN. J Harbin Inst Technol 50(4):117–123 Huizhen ZHAO, Fuxian L, Longyue L (2018) A novel softplus linear unit for deep CNN. J Harbin Inst Technol 50(4):117–123
12.
Zurück zum Zitat Kiliçarslan S, Celik M (2021) RSigELU: A nonlinear activation function for deep neural networks. Expert Syst Appl 174:114805CrossRef Kiliçarslan S, Celik M (2021) RSigELU: A nonlinear activation function for deep neural networks. Expert Syst Appl 174:114805CrossRef
13.
Zurück zum Zitat Kingma, D. P., & Ba, J. (2014). Adam: A method for stochastic optimization. arXiv preprint . Kingma, D. P., & Ba, J. (2014). Adam: A method for stochastic optimization. arXiv preprint .
14.
Zurück zum Zitat Kiseľák J, Lu Y, Švihra J, Szépe P, Stehlík M (2021) “SPOCU”: scaled polynomial constant unit activation function. Neural Comput Appl 33(8):3385–3401CrossRef Kiseľák J, Lu Y, Švihra J, Szépe P, Stehlík M (2021) “SPOCU”: scaled polynomial constant unit activation function. Neural Comput Appl 33(8):3385–3401CrossRef
15.
Zurück zum Zitat LeCun Y, Bengio Y, Hinton G (2015) Deep learning. Nature 521(7553):436–444CrossRef LeCun Y, Bengio Y, Hinton G (2015) Deep learning. Nature 521(7553):436–444CrossRef
16.
Zurück zum Zitat LeCun, Y., Boser, B., Denker, J., Henderson, D., Howard, R., Hubbard, W., & Jackel, L. (1989). Handwritten digit recognition with a back-propagation network. Advances in neural information processing systems 2 LeCun, Y., Boser, B., Denker, J., Henderson, D., Howard, R., Hubbard, W., & Jackel, L. (1989). Handwritten digit recognition with a back-propagation network. Advances in neural information processing systems 2
17.
Zurück zum Zitat Lee, J., Shridhar, K., Hayashi, H., Iwana, B. K., Kang, S., & Uchida, S. (2019). Probact: A probabilistic activation function for deep neural networks. arXiv preprint 5, 13 Lee, J., Shridhar, K., Hayashi, H., Iwana, B. K., Kang, S., & Uchida, S. (2019). Probact: A probabilistic activation function for deep neural networks. arXiv preprint  5, 13
18.
Zurück zum Zitat Li Y, Fan C, Li Y, Wu Q, Ming Y (2018) Improving deep neural network with multiple parametric exponential linear units. Neurocomputing 301:11–24CrossRef Li Y, Fan C, Li Y, Wu Q, Ming Y (2018) Improving deep neural network with multiple parametric exponential linear units. Neurocomputing 301:11–24CrossRef
19.
Zurück zum Zitat Livieris IE, Pintelas E, Pintelas P (2020) A CNN–LSTM model for gold price time-series forecasting. Neural Comput Appl 32(23):17351–17360CrossRef Livieris IE, Pintelas E, Pintelas P (2020) A CNN–LSTM model for gold price time-series forecasting. Neural Comput Appl 32(23):17351–17360CrossRef
20.
Zurück zum Zitat Maas, A. L., Hannun, A. Y., & Ng, A. Y. (2013, June). Rectifier nonlinearities improve neural network acoustic models. In Proc Icml 30(1): 3 Maas, A. L., Hannun, A. Y., & Ng, A. Y. (2013, June). Rectifier nonlinearities improve neural network acoustic models. In Proc Icml 30(1): 3
21.
Zurück zum Zitat Nair, V., & Hinton, G. E. (2010, January). Rectified linear units improve restricted boltzmann machines. In Proceedings of the 27th international conference on machine learning (Icml): 807–814 Nair, V., & Hinton, G. E. (2010, January). Rectified linear units improve restricted boltzmann machines. In Proceedings of the 27th international conference on machine learning (Icml): 807–814
22.
Zurück zum Zitat Ozguven MM, Adem K (2019) Automatic detection and classification of leaf spot disease in sugar beet using deep learning algorithms. Physica A 535:122537CrossRef Ozguven MM, Adem K (2019) Automatic detection and classification of leaf spot disease in sugar beet using deep learning algorithms. Physica A 535:122537CrossRef
23.
Zurück zum Zitat Pacal I, Karaboga D (2021) A Robust Real-Time Deep Learning Based Automatic Polyp Detection System. Comput Biol Med 134:104519CrossRef Pacal I, Karaboga D (2021) A Robust Real-Time Deep Learning Based Automatic Polyp Detection System. Comput Biol Med 134:104519CrossRef
24.
Zurück zum Zitat Qiumei Z, Dan T, Fenghua W (2019) Improved convolutional neural network based on fast exponentially linear unit activation function. Ieee Access 7:151359–151367CrossRef Qiumei Z, Dan T, Fenghua W (2019) Improved convolutional neural network based on fast exponentially linear unit activation function. Ieee Access 7:151359–151367CrossRef
25.
Zurück zum Zitat Ramachandran, P., Zoph, B., & Le, Q. V. (2017). Searching for activation functions. arXiv preprint . Ramachandran, P., Zoph, B., & Le, Q. V. (2017). Searching for activation functions. arXiv preprint .
26.
Zurück zum Zitat Simonyan, K., & Zisserman, A. (2014). Very deep convolutional networks for large-scale image recognition. arXiv preprint . Simonyan, K., & Zisserman, A. (2014). Very deep convolutional networks for large-scale image recognition. arXiv preprint .
27.
Zurück zum Zitat Trottier, L., Giguere, P., & Chaib-Draa, B. (2017, December). Parametric exponential linear unit for deep convolutional neural networks. In 2017 16th IEEE International Conference on Machine Learning and Applications (ICMLA): 207–214 IEEE Trottier, L., Giguere, P., & Chaib-Draa, B. (2017, December). Parametric exponential linear unit for deep convolutional neural networks. In 2017 16th IEEE International Conference on Machine Learning and Applications (ICMLA): 207–214 IEEE
28.
Zurück zum Zitat Wang X, Qin Y, Wang Y, Xiang S, Chen H (2019) ReLTanh: An activation function with vanishing gradient resistance for SAE-based DNNs and its application to rotating machinery fault diagnosis. Neurocomputing 363:88–98CrossRef Wang X, Qin Y, Wang Y, Xiang S, Chen H (2019) ReLTanh: An activation function with vanishing gradient resistance for SAE-based DNNs and its application to rotating machinery fault diagnosis. Neurocomputing 363:88–98CrossRef
29.
Zurück zum Zitat Wang Y, Li Y, Song Y, Rong X (2020) The influence of the activation function in a convolution neural network model of facial expression recognition. Appl Sci 10(5):1897CrossRef Wang Y, Li Y, Song Y, Rong X (2020) The influence of the activation function in a convolution neural network model of facial expression recognition. Appl Sci 10(5):1897CrossRef
30.
Zurück zum Zitat Xiao J, Xu H, Gao H, Bian M, Li Y (2021) A weakly supervised semantic segmentation network by aggregating seed cues: the multi-object proposal generation perspective. ACM Transact Multimidia Comput Communicat Appl 17(1s):1–19CrossRef Xiao J, Xu H, Gao H, Bian M, Li Y (2021) A weakly supervised semantic segmentation network by aggregating seed cues: the multi-object proposal generation perspective. ACM Transact Multimidia Comput Communicat Appl 17(1s):1–19CrossRef
31.
Zurück zum Zitat Zhang T, Yang J, Song WA, Song CF (2019) Research on improved activation function TReLU. J Chinese Comput Syst 40(1):58–63MathSciNet Zhang T, Yang J, Song WA, Song CF (2019) Research on improved activation function TReLU. J Chinese Comput Syst 40(1):58–63MathSciNet
32.
Zurück zum Zitat Zhou Y, Li D, Huo S, Kung SY (2021) Shape autotuning activation function. Expert Syst Appl 171:114534CrossRef Zhou Y, Li D, Huo S, Kung SY (2021) Shape autotuning activation function. Expert Syst Appl 171:114534CrossRef
33.
Zurück zum Zitat Zhu H, Zeng H, Liu J, Zhang X (2021) Logish: A new nonlinear nonmonotonic activation function for convolutional neural network. Neurocomputing 458:490–499CrossRef Zhu H, Zeng H, Liu J, Zhang X (2021) Logish: A new nonlinear nonmonotonic activation function for convolutional neural network. Neurocomputing 458:490–499CrossRef
Metadaten
Titel
P + FELU: Flexible and trainable fast exponential linear unit for deep learning architectures
verfasst von
Kemal Adem
Publikationsdatum
29.07.2022
Verlag
Springer London
Erschienen in
Neural Computing and Applications / Ausgabe 24/2022
Print ISSN: 0941-0643
Elektronische ISSN: 1433-3058
DOI
https://doi.org/10.1007/s00521-022-07625-3

Weitere Artikel der Ausgabe 24/2022

Neural Computing and Applications 24/2022 Zur Ausgabe

Premium Partner