Skip to main content
Top
Published in: Arabian Journal for Science and Engineering 9/2021

22-05-2021 | Research Article-Computer Engineering and Computer Science

An Enhanced Gated Recurrent Unit with Auto-Encoder for Solving Text Classification Problems

Authors: Muhammad Zulqarnain, Rozaida Ghazali, Yana Mazwin Mohmad Hassim, Muhammad Aamir

Published in: Arabian Journal for Science and Engineering | Issue 9/2021

Log in

Activate our intelligent search to find suitable subject content or patents.

search-config
loading …

Abstract

Classification has become an important task for automatically categorizing documents based on their respective group. The purpose of classification is to assign the pre-specified group or class to an instance based on the observed features related to that instance. For accurate text classification, feature selection techniques are normally used to identify important features and to remove irrelevant, undesired and noisy features for minimizing the dimensionality of feature space. Therefore, in this research, a new model namely Encoder Simplified GRU (ES-GRU) is proposed to reduce dimension of data using an auto-encoder (AE). Gated Recurrent Unit (GRU) is a deep learning algorithm that contains update gate and reset gate, which is considered as one of the most efficient text classification technique, specifically on sequential datasets. Accordingly, the reset gate is replaced with an update gate in order to reduce the redundancy and complexity in the standard GRU. The proposed model has been evaluated on five benchmark text datasets and compared with six baseline well-known text classification approaches, which includes standard GRU, AE, Long Short-Term Memory, Convolutional Neural Network, Support Vector Machine, and Naïve Bayes. Based on various types of performance evaluation parameters, a considerable amount of improvement has been observed in the performance of the proposed model as compared to state-of-the-art approaches.

Dont have a licence yet? Then find out more about our products and how to get one now:

Springer Professional "Wirtschaft+Technik"

Online-Abonnement

Mit Springer Professional "Wirtschaft+Technik" erhalten Sie Zugriff auf:

  • über 102.000 Bücher
  • über 537 Zeitschriften

aus folgenden Fachgebieten:

  • Automobil + Motoren
  • Bauwesen + Immobilien
  • Business IT + Informatik
  • Elektrotechnik + Elektronik
  • Energie + Nachhaltigkeit
  • Finance + Banking
  • Management + Führung
  • Marketing + Vertrieb
  • Maschinenbau + Werkstoffe
  • Versicherung + Risiko

Jetzt Wissensvorsprung sichern!

Springer Professional "Technik"

Online-Abonnement

Mit Springer Professional "Technik" erhalten Sie Zugriff auf:

  • über 67.000 Bücher
  • über 390 Zeitschriften

aus folgenden Fachgebieten:

  • Automobil + Motoren
  • Bauwesen + Immobilien
  • Business IT + Informatik
  • Elektrotechnik + Elektronik
  • Energie + Nachhaltigkeit
  • Maschinenbau + Werkstoffe




 

Jetzt Wissensvorsprung sichern!

Literature
1.
go back to reference Wang Z.; and Qu, Z. “Research on web text classification algorithm based on improved CNN and SVM,” IEEE, pp. 1958–1961, 2017. Wang Z.; and Qu, Z. “Research on web text classification algorithm based on improved CNN and SVM,” IEEE, pp. 1958–1961, 2017.
2.
go back to reference Sharif, W.; Samsudin, N.A; M. M. Deris, M.M and M. Aamir, “Improved relative discriminative criterion feature ranking technique for text classification.” Int. J. Artif. Intell., 15(2), pp. 61–78, 2017. Sharif, W.; Samsudin, N.A; M. M. Deris, M.M and M. Aamir, “Improved relative discriminative criterion feature ranking technique for text classification.” Int. J. Artif. Intell., 15(2), pp. 61–78, 2017.
4.
go back to reference Zulqarnain, M.; Ishak, S.A.; Ghazali, R.; Nawi, N.M.: An improved deep learning approach based on variant two-state gated recurrent unit and word embeddings for sentiment classification. Int. J. Adv. Comput. Sci. Appl. 11(1), 594–603 (2020) Zulqarnain, M.; Ishak, S.A.; Ghazali, R.; Nawi, N.M.: An improved deep learning approach based on variant two-state gated recurrent unit and word embeddings for sentiment classification. Int. J. Adv. Comput. Sci. Appl. 11(1), 594–603 (2020)
9.
go back to reference Sharif, W.; Tun, U.; Onn, H.; Tri, I.; Yanto, R.: An optimised support vector machine with ringed seal search algorithm for efficient text classification. J. Eng. Sci. Technol. 14(3), 1601–1613 (2019) Sharif, W.; Tun, U.; Onn, H.; Tri, I.; Yanto, R.: An optimised support vector machine with ringed seal search algorithm for efficient text classification. J. Eng. Sci. Technol. 14(3), 1601–1613 (2019)
13.
go back to reference Berger, M.J.: “Large scale multi-label text classification with semantic word vectors.” Tech. Rep., pp. 1–8, 2014. Berger, M.J.: “Large scale multi-label text classification with semantic word vectors.” Tech. Rep., pp. 1–8, 2014.
14.
go back to reference Yeh, C.K.; Wu, W.C.; Ko, W.J.; Wang, Y.C.F.: “Learning deep latent spaces for multi-label classification.” 31st AAAI Conf Artif. Intell. AAAI 2017, 2838–2844 (2017) Yeh, C.K.; Wu, W.C.; Ko, W.J.; Wang, Y.C.F.: “Learning deep latent spaces for multi-label classification.” 31st AAAI Conf Artif. Intell. AAAI 2017, 2838–2844 (2017)
21.
go back to reference Zhou, G.: Minimal gated unit for recurrent neural networks. ICML 7, 153–163 (2016) Zhou, G.: Minimal gated unit for recurrent neural networks. ICML 7, 153–163 (2016)
24.
go back to reference Conneau, A.; Schwenk, H.; Barrault, L. and Lecun, Y.: “Very deep convolutional networks for text classification,” arXiv, pp. 1–10, 2017. Conneau, A.; Schwenk, H.; Barrault, L. and Lecun, Y.: “Very deep convolutional networks for text classification,” arXiv, pp. 1–10, 2017.
26.
go back to reference Vincent, P.: “A Neural Probabilistic Language Model.” Neural Probabilistic Lang Model. J. Mach. Learn. Res. 3, 1137–1155 (2003)MATH Vincent, P.: “A Neural Probabilistic Language Model.” Neural Probabilistic Lang Model. J. Mach. Learn. Res. 3, 1137–1155 (2003)MATH
27.
go back to reference Socher, R.; Huval, B.; Manning, C.D. and Ng, A.Y.: “Semantic compositionality through recursive matrix-vector spaces.” Proc. 2012 Jt. Conf. Empir. methods Nat. Lang. Process. Comput. Nat. Lang. Learn., pp. 1201–1211, 2012. Socher, R.; Huval, B.; Manning, C.D. and Ng, A.Y.: “Semantic compositionality through recursive matrix-vector spaces.” Proc. 2012 Jt. Conf. Empir. methods Nat. Lang. Process. Comput. Nat. Lang. Learn., pp. 1201–1211, 2012.
28.
go back to reference Joulin, A.; Grave, E.; Bojanowski, P. and Mikolov, T.: “Bag of tricks for efficient text classification.” 2016 1511.09249v1. Joulin, A.; Grave, E.; Bojanowski, P. and Mikolov, T.: “Bag of tricks for efficient text classification.” 2016 1511.09249v1.
29.
30.
go back to reference Zhang, X.; Zhao, J. and Lecun, Y.: “Character-level convolutional networks for text classification.” Adv. Neural Inf. Process. Syst., vol. 2015 pp. 649–657, 2015. Zhang, X.; Zhao, J. and Lecun, Y.: “Character-level convolutional networks for text classification.” Adv. Neural Inf. Process. Syst., vol. 2015 pp. 649–657, 2015.
32.
go back to reference Zhou, P.; Qi, Z.; Zheng, S.; Xu, J.; Bao, H. and Xu, B.: “Text classification improved by integrating bidirectional LSTM with two-dimensional max pooling.” COLING 2016-26th Int. Conf. Comput. Linguist. Proc. COLING 2016 Tech. Pap., 2(1), pp. 3485–3495, 2016. Zhou, P.; Qi, Z.; Zheng, S.; Xu, J.; Bao, H. and Xu, B.: “Text classification improved by integrating bidirectional LSTM with two-dimensional max pooling.” COLING 2016-26th Int. Conf. Comput. Linguist. Proc. COLING 2016 Tech. Pap., 2(1), pp. 3485–3495, 2016.
33.
go back to reference Wei, J. and Zou, K.: “EDA: Easy data augmentation techniques for boosting performance on text classification tasks.” EMNLP-IJCNLP 2019-2019 Conf. Empir. Methods Nat. Lang. Process. 9th Int. Jt. Conf. Nat. Lang. Process. Proc. Conf., pp. 6382–6388, 2020 https://doi.org/10.18653/v1/d19-1670. Wei, J. and Zou, K.: “EDA: Easy data augmentation techniques for boosting performance on text classification tasks.” EMNLP-IJCNLP 2019-2019 Conf. Empir. Methods Nat. Lang. Process. 9th Int. Jt. Conf. Nat. Lang. Process. Proc. Conf., pp. 6382–6388, 2020 https://​doi.​org/​10.​18653/​v1/​d19-1670.
35.
go back to reference Liao, R. et al.: “Reviving and improving recurrent back-propagation,” arXiv, pp. 3082–3091, 2018. Liao, R. et al.: “Reviving and improving recurrent back-propagation,” arXiv, pp. 3082–3091, 2018.
38.
go back to reference Wang, Y.; Wang, H.; Zhang, X.; Chaspari, T.; Choe, Y. and Lu, M.: “An Attention-aware Bidirectional Multi-residual Recurrent Neural Network (Abmrnn): A Study about Better Short-term Text Classification,” ICASSP, IEEE Int. Conf. Acoust. Speech Signal Process. Proc., 2019, pp. 3582–3586, 2019 https://doi.org/10.1109/ICASSP.2019.8682565. Wang, Y.; Wang, H.; Zhang, X.; Chaspari, T.; Choe, Y. and Lu, M.: “An Attention-aware Bidirectional Multi-residual Recurrent Neural Network (Abmrnn): A Study about Better Short-term Text Classification,” ICASSP, IEEE Int. Conf. Acoust. Speech Signal Process. Proc., 2019, pp. 3582–3586, 2019 https://​doi.​org/​10.​1109/​ICASSP.​2019.​8682565.
41.
go back to reference Sundermeyer, M.; Ney, H. and Schlüter, R.: “From feedforward to recurrent LSTM neural networks for language modeling.” IEEE/ACM Trans. Audio, Speech, Lang. Process, 23(3), pp. 517–529, 2015. Sundermeyer, M.; Ney, H. and Schlüter, R.: “From feedforward to recurrent LSTM neural networks for language modeling.” IEEE/ACM Trans. Audio, Speech, Lang. Process, 23(3), pp. 517–529, 2015.
42.
go back to reference Lee, H.: “For modeling sentences and documents.” Proc. 15th Annu. Conf. North Am. Chapter Assoc. Comput., pp. 1512–1521, 2015. Lee, H.: “For modeling sentences and documents.” Proc. 15th Annu. Conf. North Am. Chapter Assoc. Comput., pp. 1512–1521, 2015.
43.
go back to reference Pascanu, R.; Tour, D.; Mikolov, T.; Tour, D.: On the difficulty of training recurrent neural networks. Conf. ICLR 2, 1310–1318 (2013) Pascanu, R.; Tour, D.; Mikolov, T.; Tour, D.: On the difficulty of training recurrent neural networks. Conf. ICLR 2, 1310–1318 (2013)
46.
go back to reference Metwally, A.A.; Yu, P.S.; Reiman, D.; Dai, Y.; Finn, P.W.; Perkins, D.L.: Utilizing longitudinal microbiome taxonomic profiles to predict food allergy via long short-term memory networks. PLoS Comput. Biol. 15(2), 1–16 (2019).CrossRef Metwally, A.A.; Yu, P.S.; Reiman, D.; Dai, Y.; Finn, P.W.; Perkins, D.L.: Utilizing longitudinal microbiome taxonomic profiles to predict food allergy via long short-term memory networks. PLoS Comput. Biol. 15(2), 1–16 (2019).CrossRef
47.
go back to reference Pennington, J.; Socher, R. and Manning, C.D.: “GloVe : Global vectors for word representation.” Proc. Conf. Empir. Methods Nat. Lang. Process., pp. 1532–1543, 2014. Pennington, J.; Socher, R. and Manning, C.D.: “GloVe : Global vectors for word representation.” Proc. Conf. Empir. Methods Nat. Lang. Process., pp. 1532–1543, 2014.
48.
go back to reference Hinton, G.: Dropout: a simple way to prevent neural networks from overfitting. J. Machine Learn. Res. 15(1), 1929–1958 (2014)MathSciNetMATH Hinton, G.: Dropout: a simple way to prevent neural networks from overfitting. J. Machine Learn. Res. 15(1), 1929–1958 (2014)MathSciNetMATH
49.
go back to reference Chu, A.; Stratos, K. and Gimpel, K.: “Unsupervised label refinement improves dataless text classification,” arXiv, 2020. Chu, A.; Stratos, K. and Gimpel, K.: “Unsupervised label refinement improves dataless text classification,” arXiv, 2020.
50.
51.
go back to reference Yogatama, D.; Dyer, C.; Ling, W.; Blunsom, P.: “Generative and discriminative text classification with recurrent neural networks.” arXiv, no. May, pp. 1–9, 2017. Yogatama, D.; Dyer, C.; Ling, W.; Blunsom, P.: “Generative and discriminative text classification with recurrent neural networks.” arXiv, no. May, pp. 1–9, 2017.
57.
go back to reference Camacho-Collados, J. and Pilehvar, M. T.: “on the role of text preprocessing in neural network architectures: An evaluation study on text categorization and sentiment analysis,” arXiv Prepr. arXiv1707.01780, pp. 40–46, 2018, https://doi.org/10.18653/v1/w18-5406. Camacho-Collados, J. and Pilehvar, M. T.: “on the role of text preprocessing in neural network architectures: An evaluation study on text categorization and sentiment analysis,” arXiv Prepr. arXiv1707.01780, pp. 40–46, 2018, https://​doi.​org/​10.​18653/​v1/​w18-5406.
Metadata
Title
An Enhanced Gated Recurrent Unit with Auto-Encoder for Solving Text Classification Problems
Authors
Muhammad Zulqarnain
Rozaida Ghazali
Yana Mazwin Mohmad Hassim
Muhammad Aamir
Publication date
22-05-2021
Publisher
Springer Berlin Heidelberg
Published in
Arabian Journal for Science and Engineering / Issue 9/2021
Print ISSN: 2193-567X
Electronic ISSN: 2191-4281
DOI
https://doi.org/10.1007/s13369-021-05691-8

Other articles of this Issue 9/2021

Arabian Journal for Science and Engineering 9/2021 Go to the issue

Research Article-Computer Engineering and Computer Science

Classification of Marine Plankton Based on Few-shot Learning

Research Article-Computer Engineering and Computer Science

Arabic Sentiment Analysis Using Deep Learning and Ensemble Methods

Premium Partners