Skip to main content
Top
Published in: Cognitive Computation 4/2017

27-05-2017

Effective Use of Evaluation Measures for the Validation of Best Classifier in Urdu Sentiment Analysis

Authors: Neelam Mukhtar, Mohammad Abid Khan, Nadia Chiragh

Published in: Cognitive Computation | Issue 4/2017

Log in

Activate our intelligent search to find suitable subject content or patents.

search-config
loading …

Abstract

Sentiment analysis (SA) can help in decision making, drawing conclusion, or recommending appropriate solution for different business, political, or other problems. At the same time reliable ways are also required to verify the results that are achieved after SA. In the frame of biologically inspired approaches for machine learning, getting reliable result is challenging but important. Properly verified and validated results are always appreciated and preferred by the research community. The strategy of achieving reliable result is adopted in this research by using three standard evaluation measures. First, SA of Urdu is performed. After collection and annotation of data, five classifiers, i.e., PART, Naives Bayes mutinomial Text, Lib SVM (support vector machine), decision tree (J48), and k nearest neighbor (KNN, IBK) are employed using Weka. After using 10-fold cross-validation, three top most classifiers, i.e., Lib SVM, J48, and IBK are selected on the basis of high accuracy, precision, recall, and F-measure. Further, IBK resulted as the best classifier among the three. For verification of this result, labels of the sentences (positive, negative, or neutral) are predicted by using training and test data, followed by the application of the three standard evaluation measures, i.e., McNemar’s test, kappa statistic, and root mean squared error. IBK performs much better than the other two classifiers. To make this result more reliable, a number of steps are taken including the use of three evaluation measures for getting a confirmed and validated result which is the main contribution of this research. It is concluded with confidence that IBK is the best classifier in this case.

Dont have a licence yet? Then find out more about our products and how to get one now:

Springer Professional "Wirtschaft+Technik"

Online-Abonnement

Mit Springer Professional "Wirtschaft+Technik" erhalten Sie Zugriff auf:

  • über 102.000 Bücher
  • über 537 Zeitschriften

aus folgenden Fachgebieten:

  • Automobil + Motoren
  • Bauwesen + Immobilien
  • Business IT + Informatik
  • Elektrotechnik + Elektronik
  • Energie + Nachhaltigkeit
  • Finance + Banking
  • Management + Führung
  • Marketing + Vertrieb
  • Maschinenbau + Werkstoffe
  • Versicherung + Risiko

Jetzt Wissensvorsprung sichern!

Springer Professional "Technik"

Online-Abonnement

Mit Springer Professional "Technik" erhalten Sie Zugriff auf:

  • über 67.000 Bücher
  • über 390 Zeitschriften

aus folgenden Fachgebieten:

  • Automobil + Motoren
  • Bauwesen + Immobilien
  • Business IT + Informatik
  • Elektrotechnik + Elektronik
  • Energie + Nachhaltigkeit
  • Maschinenbau + Werkstoffe




 

Jetzt Wissensvorsprung sichern!

Springer Professional "Wirtschaft"

Online-Abonnement

Mit Springer Professional "Wirtschaft" erhalten Sie Zugriff auf:

  • über 67.000 Bücher
  • über 340 Zeitschriften

aus folgenden Fachgebieten:

  • Bauwesen + Immobilien
  • Business IT + Informatik
  • Finance + Banking
  • Management + Führung
  • Marketing + Vertrieb
  • Versicherung + Risiko




Jetzt Wissensvorsprung sichern!

Literature
1.
go back to reference Cambria E, Schuller B, Xia Y, Havasi C. New avenues in opinion mining and sentiment analysis. IEEE Intell Syst. 2013;28(2):15–21.CrossRef Cambria E, Schuller B, Xia Y, Havasi C. New avenues in opinion mining and sentiment analysis. IEEE Intell Syst. 2013;28(2):15–21.CrossRef
2.
go back to reference Palogiannidi E, Kolovou A, Christopoulou F, Kokkinos F, Iosif E, Malandrakis N, et al., editors. Tweester at SemEval-2016 Task 4: Sentiment analysis in Twitter using semantic- affective model adaptation. 10th International Workshop on Semantic Evaluation (SemEval 2016) 2016; San Diego, US. Palogiannidi E, Kolovou A, Christopoulou F, Kokkinos F, Iosif E, Malandrakis N, et al., editors. Tweester at SemEval-2016 Task 4: Sentiment analysis in Twitter using semantic- affective model adaptation. 10th International Workshop on Semantic Evaluation (SemEval 2016) 2016; San Diego, US.
3.
go back to reference Cambria E. Affective computing and sentiment analysis. IEEE Intell Syst. 2016;31:102–7.CrossRef Cambria E. Affective computing and sentiment analysis. IEEE Intell Syst. 2016;31:102–7.CrossRef
4.
go back to reference Ofek N, Rokach L, Cambria E, Hussain A, Shabtai A. Unsupervised commonsense knowledge enrichment for domain-specific sentiment analysis. Cogn Comput. 2016;8(3):467–77.CrossRef Ofek N, Rokach L, Cambria E, Hussain A, Shabtai A. Unsupervised commonsense knowledge enrichment for domain-specific sentiment analysis. Cogn Comput. 2016;8(3):467–77.CrossRef
5.
go back to reference Oneto L, Bisio F, Cambria E, Anguita D. Statistical learning theory and ELM for big social data analysis. IEEE Comput Intell Mag. 2016;11(3):45–55.CrossRef Oneto L, Bisio F, Cambria E, Anguita D. Statistical learning theory and ELM for big social data analysis. IEEE Comput Intell Mag. 2016;11(3):45–55.CrossRef
6.
go back to reference Bautin M, Vijayarenu L, Skiena S, editors. International Sentiment Analysis for News and Blog. Second International Conference on Weblogs and Social Media Seattle, WA; 2008. Bautin M, Vijayarenu L, Skiena S, editors. International Sentiment Analysis for News and Blog. Second International Conference on Weblogs and Social Media Seattle, WA; 2008.
7.
go back to reference Cambria E, Poria S, Bajpai R, Schuller B, editors. SenticNet 4: A Semantic Resource for Sentiment Analysis Based on Conceptual Primitives. Proceedings of COLING 2016, the 26th International Conference on Computational Linguistics: Technical Papers; 2016; Japan. Cambria E, Poria S, Bajpai R, Schuller B, editors. SenticNet 4: A Semantic Resource for Sentiment Analysis Based on Conceptual Primitives. Proceedings of COLING 2016, the 26th International Conference on Computational Linguistics: Technical Papers; 2016; Japan.
8.
go back to reference Appela O, Chiclana F, Cartera J, Fujitab H. A hybrid approach to the sentiment analysis problem at the sentence level. Spec Issue New Avenues Knowl Bases Nat Lang Process Knowl-Based Syst. 2016;108:110–24. Appela O, Chiclana F, Cartera J, Fujitab H. A hybrid approach to the sentiment analysis problem at the sentence level. Spec Issue New Avenues Knowl Bases Nat Lang Process Knowl-Based Syst. 2016;108:110–24.
9.
go back to reference Minhas S, Hussain A. From spin to swindle: identifying falsification in financial text. Cogn Comput. 2016;8:729–45.CrossRef Minhas S, Hussain A. From spin to swindle: identifying falsification in financial text. Cogn Comput. 2016;8:729–45.CrossRef
10.
go back to reference Khan FH, Qamar U, Bashir S. Multi-objective model selection (MOMS)-based semi-supervised framework for sentiment analysis. Cogn Comput. 2016;8(4):614–28.CrossRef Khan FH, Qamar U, Bashir S. Multi-objective model selection (MOMS)-based semi-supervised framework for sentiment analysis. Cogn Comput. 2016;8(4):614–28.CrossRef
11.
go back to reference Dashtipour K, Poria S, Hussain A, Cambria E, Hawalah AYA, Gelbukh A, et al. Multilingual sentiment analysis: state of the art and independent comparison of techniques. Cogn Comput. 2016;8:757–71.CrossRef Dashtipour K, Poria S, Hussain A, Cambria E, Hawalah AYA, Gelbukh A, et al. Multilingual sentiment analysis: state of the art and independent comparison of techniques. Cogn Comput. 2016;8:757–71.CrossRef
12.
go back to reference Bilal M, Israr H, Shahid M, Khan A. Sentiment classification of Roman-Urdu opinions using Naı¨ve Bayesian, decision tree and KNN classification techniques. J King Saud Univ Comput Inf Sci. 2015; Bilal M, Israr H, Shahid M, Khan A. Sentiment classification of Roman-Urdu opinions using Naı¨ve Bayesian, decision tree and KNN classification techniques. J King Saud Univ Comput Inf Sci. 2015;
13.
go back to reference Syed AZ, Muhammad A, Enríquez AMM, editors. Lexicon Based Sentiment Analysis of Urdu Text Using SentiUnits. Proceedings of the 9th Mexican international conference of artificial intelligence, MICAI; 2010; Berlin Heidelberg. Springer. Syed AZ, Muhammad A, Enríquez AMM, editors. Lexicon Based Sentiment Analysis of Urdu Text Using SentiUnits. Proceedings of the 9th Mexican international conference of artificial intelligence, MICAI; 2010; Berlin Heidelberg. Springer.
14.
go back to reference Syed AZ, Muhammad A, Enríquez AMM. Adjectival phrases as the sentiment carriers in Urdu. J Am Sci. 2011;7(3):644–52. Syed AZ, Muhammad A, Enríquez AMM. Adjectival phrases as the sentiment carriers in Urdu. J Am Sci. 2011;7(3):644–52.
15.
go back to reference Syed AZ, Muhammad A, Enríquez AMM. Associating targets with SentiUnits: a step forward in sentiment analysis of Urdu text. Artif Intell Rev Springer. 2014;41(4):535–61.CrossRef Syed AZ, Muhammad A, Enríquez AMM. Associating targets with SentiUnits: a step forward in sentiment analysis of Urdu text. Artif Intell Rev Springer. 2014;41(4):535–61.CrossRef
16.
go back to reference Daud M, Khan R, Duad A. Roman Urdu opinion mining system (RUOMiS). CSEIJ. 2014;4(6):1–9.CrossRef Daud M, Khan R, Duad A. Roman Urdu opinion mining system (RUOMiS). CSEIJ. 2014;4(6):1–9.CrossRef
17.
go back to reference Dietterich TG. Approximate statistical tests for comparing supervised classification learning algorithms. Neural Comput. 1998;10:1895–923.CrossRefPubMed Dietterich TG. Approximate statistical tests for comparing supervised classification learning algorithms. Neural Comput. 1998;10:1895–923.CrossRefPubMed
18.
go back to reference Bouckaert RR, Frank E, editors. Evaluating the replicability of significance tests for comparing learning algorithms. 8th Pacific-Asia Conference; 2004. Bouckaert RR, Frank E, editors. Evaluating the replicability of significance tests for comparing learning algorithms. 8th Pacific-Asia Conference; 2004.
19.
go back to reference Demsar J. Statistical comparisons of classifiers over multiple data sets. J Mach Learn Res. 2006:1–6. Demsar J. Statistical comparisons of classifiers over multiple data sets. J Mach Learn Res. 2006:1–6.
20.
go back to reference Bostanci B, Bostanci E, editors. An evaluation of classification algorithms using Mc Nemar’s test. Seventh International Conference on Bio-Inspired Computing: Theories and Applications; 2013; New Delhi. Advances in Intelligent Systems and Computing, Springer. Bostanci B, Bostanci E, editors. An evaluation of classification algorithms using Mc Nemar’s test. Seventh International Conference on Bio-Inspired Computing: Theories and Applications; 2013; New Delhi. Advances in Intelligent Systems and Computing, Springer.
22.
go back to reference Vieira S, Kaymak U, Sousa J, editors. Cohen’s kappa coefficient as a performance measure for feature selection. IEEE International Conference on Fuzzy Systems (FUZZ) 2010; Piscataway. Vieira S, Kaymak U, Sousa J, editors. Cohen’s kappa coefficient as a performance measure for feature selection. IEEE International Conference on Fuzzy Systems (FUZZ) 2010; Piscataway.
23.
go back to reference Ben-David A. Comparison of classification accuracy using Cohen's weighted kappa. Expert Syst Appl. 2008;34(2):825–32. Ben-David A. Comparison of classification accuracy using Cohen's weighted kappa. Expert Syst Appl. 2008;34(2):825–32.
24.
go back to reference Petrakos M, Benediktsson J. The effect of classifier agreement on the accuracy of the combined classifier in decision level fusion. IEEE Trans Geosci Remote Sens. 2001;39(11):2539–46.CrossRef Petrakos M, Benediktsson J. The effect of classifier agreement on the accuracy of the combined classifier in decision level fusion. IEEE Trans Geosci Remote Sens. 2001;39(11):2539–46.CrossRef
25.
go back to reference Caruana R, Niculescu-Mizil A, editors. An empirical comparison of supervised learning algorithms. 23rd International Conference on Machine learning; 2006; New York. ACM. Caruana R, Niculescu-Mizil A, editors. An empirical comparison of supervised learning algorithms. 23rd International Conference on Machine learning; 2006; New York. ACM.
26.
go back to reference Tushkanova O, editor. Comparative analysis of the numerical measures for mining associative and causal relationships in big data Creativity in intelligent technologies and data science, First conference Proceedings, CIT &DS 2015; Russia. Tushkanova O, editor. Comparative analysis of the numerical measures for mining associative and causal relationships in big data Creativity in intelligent technologies and data science, First conference Proceedings, CIT &DS 2015; Russia.
28.
go back to reference Siegel S, John Castellan N. Nonparametric statistics for the behavioral sciences. Second ed: McGraw-Hill; 1988. Siegel S, John Castellan N. Nonparametric statistics for the behavioral sciences. Second ed: McGraw-Hill; 1988.
29.
go back to reference McHugh M. Interrater reliability: the kappa statistic. Biochem Med. 2012;22:276–82.CrossRef McHugh M. Interrater reliability: the kappa statistic. Biochem Med. 2012;22:276–82.CrossRef
30.
go back to reference Viera AJ, Garrett JM. Understanding inter observer agreement: the kappa statistic. Family Med. 2005;37(5):360–3. Viera AJ, Garrett JM. Understanding inter observer agreement: the kappa statistic. Family Med. 2005;37(5):360–3.
31.
go back to reference Silva C, Ribeiro B, editors. The importance of stop word removal on recall values in text categorization. Neural Netw, 2003 Proceedings of the International Joint Conference; 2003. IEEE. Silva C, Ribeiro B, editors. The importance of stop word removal on recall values in text categorization. Neural Netw, 2003 Proceedings of the International Joint Conference; 2003. IEEE.
32.
go back to reference Sun X, Yang Z, editors. Generalized McNemar's test for homogeneity of the marginal distributions. SAS Global Forum. Cary: SAS Institute; 2008. Sun X, Yang Z, editors. Generalized McNemar's test for homogeneity of the marginal distributions. SAS Global Forum. Cary: SAS Institute; 2008.
33.
go back to reference McNemar Q. Note on the sampling error of the difference between correlated proportions or percentages. Psychometrika. 1947;17:153–7.CrossRef McNemar Q. Note on the sampling error of the difference between correlated proportions or percentages. Psychometrika. 1947;17:153–7.CrossRef
34.
go back to reference Witten IH, Frank E, Hall MA, editors. Data mining: practical machine learning tools and techniques; 2011. Witten IH, Frank E, Hall MA, editors. Data mining: practical machine learning tools and techniques; 2011.
35.
go back to reference Japkowicz N, Shah M, editors. Evaluating learning algorithms: a classification perspective. Cambridge: Cambridge University Press; 2011. Japkowicz N, Shah M, editors. Evaluating learning algorithms: a classification perspective. Cambridge: Cambridge University Press; 2011.
Metadata
Title
Effective Use of Evaluation Measures for the Validation of Best Classifier in Urdu Sentiment Analysis
Authors
Neelam Mukhtar
Mohammad Abid Khan
Nadia Chiragh
Publication date
27-05-2017
Publisher
Springer US
Published in
Cognitive Computation / Issue 4/2017
Print ISSN: 1866-9956
Electronic ISSN: 1866-9964
DOI
https://doi.org/10.1007/s12559-017-9481-5

Other articles of this Issue 4/2017

Cognitive Computation 4/2017 Go to the issue

Premium Partner