Skip to main content
Advertisement
Browse Subject Areas
?

Click through the PLOS taxonomy to find articles in your field.

For more information about PLOS Subject Areas, click here.

  • Loading metrics

Question classification based on Bloom’s taxonomy cognitive domain using modified TF-IDF and word2vec

  • Manal Mohammed ,

    Contributed equally to this work with: Manal Mohammed, Nazlia Omar

    Roles Conceptualization, Data curation, Formal analysis, Investigation, Methodology, Project administration, Software, Validation, Writing – original draft

    manal.altamimi@outlook.com

    Affiliations CAIT, Faculty of Information Science and Technology, Universiti Kebangsaan Malaysia, Bangi, Selangor, Malaysia, Management Information Systems department, Faculty of Administrative Science, Hadhramout University, AL-Mukalla, Yemen

  • Nazlia Omar

    Contributed equally to this work with: Manal Mohammed, Nazlia Omar

    Roles Funding acquisition, Supervision, Writing – review & editing

    Affiliation CAIT, Faculty of Information Science and Technology, Universiti Kebangsaan Malaysia, Bangi, Selangor, Malaysia

Abstract

The assessment of examination questions is crucial in educational institutes since examination is one of the most common methods to evaluate students’ achievement in specific course. Therefore, there is a crucial need to construct a balanced and high-quality exam, which satisfies different cognitive levels. Thus, many lecturers rely on Bloom’s taxonomy cognitive domain, which is a popular framework developed for the purpose of assessing students’ intellectual abilities and skills. Several works have been proposed to automatically handle the classification of questions in accordance with Bloom’s taxonomy. Most of these works classify questions according to specific domain. As a result, there is a lack of technique of classifying questions that belong to the multi-domain areas. The aim of this paper is to present a classification model to classify exam questions based on Bloom’s taxonomy that belong to several areas. This study proposes a method for classifying questions automatically, by extracting two features, TFPOS-IDF and word2vec. The purpose of the first feature was to calculate the term frequency-inverse document frequency based on part of speech, in order to assign a suitable weight for essential words in the question. The second feature, pre-trained word2vec, was used to boost the classification process. Then, the combination of these features was fed into three different classifiers; K-Nearest Neighbour, Logistic Regression, and Support Vector Machine, in order to classify the questions. The experiments used two datasets. The first dataset contained 141 questions, while the other dataset contained 600 questions. The classification result for the first dataset achieved an average of 71.1%, 82.3% and 83.7% weighted F1-measure respectively. The classification result for the second dataset achieved an average of 85.4%, 89.4% and 89.7% weighted F1-measure respectively. The finding from this study showed that the proposed method is significant in classifying questions from multiple domains based on Bloom’s taxonomy.

1. Introduction

Examination assessment plays a fundamental role in evaluating how students’ proficiency in the course content [1]. Writing high quality and balanced exam questions that satisfy different levels of cognition is not an easy task [2]. Constructing the examination in a comprehensive manner must also take into consideration the difficulty levels, that match the standard objectives and outcomes of the course in a standard way such as Bloom’s taxonomy [1,3]. Bloom’s taxonomy involves three domains, i.e. cognitive, affective, and psychomotor domains. The cognitive domain, covers different thinking skills starting from the most straightforward to the most complex one [4,5] as shown in Fig 1. The key to writing learning outcomes by Bloom’s taxonomy is verbs [4].

Recently, researchers [59] have shown an increased interest in automating evaluating examination based on Bloom’s taxonomy cognitive domain. Various techniques have been used along with different features such as lexical features, and syntactic features, while few of them used semantic features. On the other hand, most previous works handled classifying questions from a specific domain, where there is a lack of techniques on classifying questions based on Bloom’s taxonomy over the multi-domain area [10,11]. Therefore, this study aims to build a question classification model based on Bloom’s taxonomy cognitive domain. Hence, considerable room for further improvement still remains, particularly in open domain area.

In general, questions classification is unlike documents classification, since the questions are written in short forms. Documents classification aids users to get and extract useful information easily due to the extensive available information. Whereas, short text suffers from the lack of gained information and sparsity [12,13]. Thus, it is not suitable to use the pure statistical method in order to perform question classification such as N-gram and TF-IDF, since these techniques need a massive amount of data in order to get high accuracy [5]. In addition, removing stop-words during the text classification pre-processing stage of the document is a common step to reduce insignificant words. Nevertheless, some of these stop-words such as what, when, where, and how are valuable in the process of question classification [14].

Another main issue in classifying questions based on Bloom’s cognitive is assigning a suitable weight for keywords that determine the level of the question, especially for the words that might appear in more than one taxonomy, such as the word ‘define’ that belongs to the Knowledge and Comprehension levels. To handle this issue, [15,16] proposed assigning weight to the words from the perspective of experts. Using this method may lead to inconsistencies due to the variety of background knowledge of each expert. This resulted in the poor performance of the classification process.

Furthermore, there are still other methods to classifying questions based on Bloom’s cognitive level which might not have been investigated yet such as using word embedding, e.g., word2vec which have shown good results in sentiment analysis and question classification in the question answering system [1719]. Thus, it might be useful to gain benefit from the use of word embedding with a combination of the modified feature TF-IDF, to enhance the classification process.

2. Related work

The issue of classifying exam questions based on Bloom’s taxonomy has received a considerable critical attention in recent years. In order to handle this task, researchers use different techniques and features. For example, [15] used a pre-defined list of keywords on their online test system, in which the verbs on the questions checked against the keyword list. Their dataset consists of 228 questions written in English. This method can only classify questions that belong to the knowledge level. A rule-based approach with natural language processing techniques was used by [16] to return significant words in a question, then assign weights to these words which are defined previously by experts. Their dataset consists of 100 computer programming questions written in English where the dataset is split into 70% training set and 30% test set. Similarly, [2] used a rule-based approach with N-gram, to enhance the classification result and produce 86% average F1. The dataset that is used in this work consists of 135 questions in computer programming subject written in English language which is split into 100 questions as a training set and 35 questions as a testing set. Likewise, [8] used rules to classify questions from three subjects; Object-oriented programming, software engineering and algorithm and data structure. The dataset contains 53 questions written in English. This work [8] used WordNet with cosine similarity to define question pattern. The accuracy of the outcome from the classification process reached 71%.

However, other researchers [5,6,2024] used machine learning techniques to classify Bloom’s taxonomy into cognitive levels. A neural network is used by [22] with different feature sets: whole feature, document frequency, and category frequency-document frequency to classify questions by training the model with a scaled conjugate gradient learning algorithm. The result shows that document frequency reduction is more effective with regard to the classification and coverage time compared to other feature reductions. Their dataset consists of 274 questions, divided into 192 training set and 82 testing set.

TF-IDF feature is extracted extensively in many works, [23] performed pre-processing for the dataset which contains 272 questions collected from different websites. Then TF-IDF was calculated and fed into Linear SVM, which produced satisfactory results in term of accuracy and precision, but not in F-measure and recall. This work was extended in [20] by increasing the size of the dataset to 600 questions and testing the effect of classification with and without stop words using the same TF-IDF feature. As a result, removing stop words does not make a big impact on the result. Similarly, [6] used the same dataset prepared by [20], but with enhanced TF-IDF which is multiplied with impact factor. Then the classification process handled by three classifiers NB, KNN, and SVM where SVM performance superior other classifiers. In addition, an ensemble technique is used by [5,24] to classify questions into Bloom’s taxonomy. [5] applied voting algorithm on a combination of three classifiers i.e. SVM, KNN and NB with chi-square, mutual information & odd ratio on questions from programming subject. [24] used ensemble to handle keywords overlapping issue by combining four classifiers namely rule-based, SVM, KNN and NB using majority voting algorithm and WordNet similarity. The dataset that is used in this work contains 100 programming questions which divided into 60 questions training set and 40 questions test set. The results show that the ensemble returns the best performance compared to each individual classifier by achieving 95% F-measure.

One of the most popular weighting methods is TF-IDF, which stands for Term Frequency- Inverse Document Frequency. It assigns scores to the importance of a word in a document, based on the lexical and morphological properties of the text. It is extensively used in many studies [2529], and in question classification [3,6,20,23,3032] However, some researchers used TF-IDF as it is, while others proposed some enhancement to the way TF-IDF is calculated in order to improve it. This is because classical TF-IDF does not capture some useful information such as the impact of word distribution between various classes [6].

A Massive Open Online Courses (MOOC) search engine proposed by [26], used a modified TF-IDF, that assigned the important words in the query to specific weights according to their parts of speech. This study changed the way of calculating the term ‘frequency’ by assigning a higher weight for the most important word, which is a noun or verb. Then in the second priority, the adjective and adverb are assigned to the weight that is less than a noun and verb but higher than other parts of speech which are assigned to the lowest weight. This method produces a significant result. Other researchers [3335] also use the syntactic feature of the parts of speech to assign a suitable weight to the terms. Another example of improving the way of calculating the TF-IDF is proposed by [36] who introduced an impact factor that is multiplied by the traditional TF-IDF. The purpose of the impact factor equation is to calculate the weights of the class distribution. This enhancement performs better than the classical method. Similarly, [6] used the impact factor to enhance the result of TF-IDF to classify question based on Bloom’s taxonomy.

A comprehensive review in question classification by [37] mentioned that most of the works in question classification used semantic and syntactic features rather than pure statistical features such as bag-of-word and n-gram. Moreover, [37] stated that semantic features have been significantly used in question classification in question answering system and information retrieval and produced considerable accuracy, while there is a lack of extracting semantic features in educational environments.

On the other hand, many Natural Language Processing (NLP) studies with deep learning models have included learning word vector representation. The word vectors are represented in a dense form known as word embedding, in which the words that are semantically and syntactically related are close to each other in the embedding space [13,3840]. Word embedding has been used efficiently in many NLP tasks [41].

Word2vec is one of the common word embeddings, where vectors representation is learnt via neural-network language model. It was developed based on Google by [38]. It required extensive training text as input and established vocabulary, then the model learned word vectors representation in such a way that the words that shared the same context are close to each other in the vector space. These word vectors can be used as features in various NLP tasks. As [17] mentioned, the pre-trained vectors can be considered as universal feature extractors. Many NLP tasks had benefited by word2vec, e.g., sentiment analysis, machine translation, and paraphrase detection [38]. [17] proposed a model that classified questions in question answering system, and sentiment analysis using Convolutional Neural Network (CNN) that was trained on the top of pre-trained word2vec, which produced excellent results. Similarly, [18] used CNN for Arabic sentiment analysis with Arabic pre-trained word2vec, which produced a significant result and outperformed previously existed methods.

The work [42] proposed a method that classified texts via SVM classifier, with the use of a semantic feature based on word2vec weighted by TF-IDF. In this study, several experiments compared if TF-IDF was better than combining it with word2vec as well as whether with or without stop words. Three steps were performed to produce the vector representation of word2vec with TF-IDF. The first step was summing the word2vec vectors to produce a single vector. The second step was the summation of multiplication of word2vec and TF-IDF. The third step was to concatenate the outcome of first and second step. For example, if the size of the vector of the first step was 200, and the vector size of the second step was 2000, then after appending them together the final representation of the vector would be 2200 which represented a document. The result showed that the best performance was achieved by word2vec without stop words being combined with TF-IDF without stop words.

Another example of using a combination of word2vec and TF-IDF was [36]. This work used three classifiers; SVM, KNN and Radial Basis Function (RBF) network, in order to classify Chinese documents. They performed enhancement to the TF-IDF by multiplying with impact factor, the improved feature called ITF-IDF. Then the ITF-IDF was combined with word2vec by applying the following equation: (1) where t is a term in document d, and J is the number of words in document d. For example, if the size of word2vec vectors is 200, then the size of the vector representation of the document can be obtained by applying Eq (1) is 200. The result of this study showed that the combination of both features outperformed the using of each feature individually with all classifiers.

The work of [32] classified questions based on a Persian question answering system. This study proposed three feature extraction methods; one of them was word2vec with weighted TF-IDF. The TF-IDF was raised to the power of TF-IDF factor and then multiplied by word2vec. The study stated that tuning the value of TF-IDF factor helped enhanced the accuracy significantly. SVM and Neural Network used as classification algorithms, in which they produced results that showed the effectiveness of the vector representation.

3. Method

The process of the proposed model is shown in Fig 2, and it involves five stages. The details of these stages will be discussed in detail in this section.

3.1 Question collection

Since the aim of this study was to build a model to classify questions based on Bloom’s Taxonomy from different domains, two open domain datasets had been used to ensure the stability of the proposed method. The first dataset were collected from several websites, books, and previous work [2]; the total number of questions in this dataset was 141 open-ended questions. The second dataset was introduced by Yahya et al. (2012), and it consisted of 600 open-ended questions. Both datasets were labeled and annotated into six classes. Table 1 shows the distribution of questions at each level in both datasets. Table 2 shows a sample of questions from both datasets.

3.2 Pre-processing

Performing pre-processing for unstructured data is highly recommended in any framework which uses machine learning [43] to reduce the unnecessary, duplicated, irrelevant, and noisy data. Therefore, pre-processing is implemented before the feature extraction phase and the classification process. However, in the pre-processing stage, the questions had to pass through several steps: normalization, tokenization, part-of-speech tagging, and stemming in order to refine and prepare them for the next stage.

The first step was the normalization process in which the unwanted data such as punctuation marks, numbers, and non-English characters will be eliminated. Then, all the words in the question will be converted into lowercase. In addition, stop-words are discarded from the question. In this step, not all stop-words will be eliminated from the question. The reason behind keeping some of the stop-words is that some questions contain important words, which can be considered as stop words in the default stop-words list, while these words have a significant impact in determining the level of the question. For example, the question ‘Define transferable skill’ belongs to Knowledge level, while the question ‘In your own words, how would you define transferable skills’ belongs to the Comprehension level. In case of eliminating the stop-words from the latter example, the question will be ‘words, define transferable skills’ where the following words will be deleted ‘your, own, how, you’, which means the question will look like Knowledge level question; these words can be effectively used in determining the class of question during classification process.

After performing normalization for the question, the tokenization will take place by splitting the question into individual words based on the whitespaces, known as tokens. Then the Stanford Part-of-Speech (POS) tagger (version 3.9.1) [44] tags the question words. The word will be labelled with a syntactic class such as noun (NN), verb (VB), adjective (JJ), and adverb (RB). Finally, the last step in pre-processing is stemming, the process of deleting the prefix and suffix in order to return the word into its original root. For example, waits, waiting, and waited after stemming will be wait. However, in order to perform stemming one of the most popular stemmers is used, which is the Porter stemmer in NLTK toolkit [45]. Unfortunately, the stemmer sometimes suffers from a problem of failure to return the correct root of some words e.g., recalls, recalling and recalled after stemming will be recal instead of recall. Although, in this case, it does not matter since similar terms are stored at the same index, which means during search and retrieval, the desirable word will be retrieved correctly. Moreover, POS tagging and the stemming will be used only with the TFPOS-IDF feature, and not word2vec as it will be discussed later.

3.3 Feature extraction

After pre-processing have been being applied, now the questions are ready for features extraction phase. Feature extraction is a process of transforming the raw input data into a meaningful set of features [46], in such a way that it can be understandable to the machine learning classifiers. Two features will be extracted. The aim of the first feature is to develop a weighting method to set the priority of words based on modified TF-IDF with Part-of-Speech (POS). Whereas the second feature is used to enhance the classification process, by extracting the semantic feature based on pre-trained word2vec. These two features will be combined to produce the vectors that will be used in the classification process.

3.3.1 Modified TF-IDF (term weighting method TFPOS-IDF).

The Term Frequency (TF) is known as a local term weight, whereas the Inverse Document Frequency (IDF) known as a global term weight, is calculated using the following formulas: (2) (3) (4) where c(t,d) indicates the occurrence of term t appears in document d, and the denominator indicates the total number of terms in document d, D is the total number of documents in the dataset, and dt is the number of the documents a term t appeared in. Many researchers [2628, 36] tries to improve the performance of TF-IDF by proposing some modification to the original equation of TF-IDF. In the case of Bloom’s taxonomy verb plays an important role to determine the level of the question. Relying on the pre-defined list does not always guarantee a good performance, especially for words that appear in more than one level. Therefore, the proposed feature Term Frequency (TF)-Inverse Document Frequency (IDF) based on Part-Of-Speech (POS) is introduced known as TFPOS-IDF. The idea beyond this approach is inspired from [26]. In order to determine the suitable rank for word weight according to Bloom’s taxonomy, several cases have been examined. The experiments have been run for fifteen times using both dataset and three classifiers. Table 3 demonstrates the average of weighted F1-measure using KNN, LR, and SVM with the collected dataset. Table 4 displays the average of weighted F1-measure using KNN, LR, and SVM with the Yahya et al.'s (2012) dataset.

thumbnail
Table 3. Weighted F1-measure of different weight cases with collected dataset using KNN, LR and SVM.

https://doi.org/10.1371/journal.pone.0230442.t003

thumbnail
Table 4. Weighted F1-measure of different weight cases with Yahya et al. (2012) dataset using KNN, LR and SVM.

https://doi.org/10.1371/journal.pone.0230442.t004

where MD, WH, VB, NN, JJ, and RB stand for the model, wh-question word, verb, noun, adjective, and adverb, respectively. From the results shown in Table 3, the best result achieved with the collected dataset when verb assigned to w1, noun and adjective assigned to w2, and other POS assigned to w3. Similar experiments had been implemented using Yahya et al.'s (2012) dataset. Table 4 shows the average of weighted F1-measure using KNN, LR, and SVM with Yahya et al.'s (2012) dataset.

The findings stated from Tables 3 and 4, confirm that verbs are the most important words to determine the level of the question according to Bloom’s taxonomy. In addition, assigning a higher weight for wh-question words, or model along with verbs does not help to improve the performance of classification since these words appear in all levels. Furthermore, nouns often carry significant information in question and document classification [10]. That is obviously clear by comparing the results of Case 1 versus Case 2, Case 6 versus Case 7, and Case 11 versus Case 12. Giving noun some priority over other POS improves the performance of the classification. Lastly, it can be observed that the best result is obtained in Case 3. Which means adjective also carries a significant information in question classification based on Bloom’s taxonomy, since it improves the classification outcome.

Thus, in the case of question classification based on Bloom’s taxonomy cognitive domain, the highest priority will be given to verb by assigning w1. Moreover, from the experiments, the best result with both datasets and both classifiers are obtained when the noun and adjective are assigned to w2. Therefore, noun and adjective will be in the second rank, in which their weight will be less than the verbs but more significant than other POS. Lastly, other POS such as preposition, and pronoun will be assigned to the lowest weights. The method for assigning a weight for the term t is illustrated by Eq 5 (5) where w1>w2>w3> 0, assume that the weights values for w1,w2, and w3 are w1 = 5, w2 = 3 and w3 = 1, as in [26] study. Now the new equation that represents TFPOS, i.e. the term frequency based on POS is: (6)

finally, the TFPOS-IDF(t,d) is calculated for term t in document d by the following equation: (7)

The output of this phase is the sparse matrix, i.e. high-dimensional vectors. After calculating TFPOS-IDF the feature vectors will be normalized to prevent the numerical complexities during the calculation. L2 norm is used for normalization, which is the most popular norm. The representation of the L2 norm is demonstrated by the following equation: (8)

Now the feature vector representation is already normalized in which the sum of square values of a row is equal to one. For example, the values of each term in question “Recall the main components of the flowchart” after pre-processing and applying TF-IDF and TFPOS-IDF is shown in Table 5. The higher value indicates that it is a more relevant and important term in this document. For example, the TF-IDF’s value for recall is 0.774, but for the modified feature TFPOS-IDF is 0.878 since this word is a verb, it has a higher value, which denotes that recall appears more in this document. Whereas other words main, components and flowchart have a higher value in TF-IDF comparing to the modified feature TFPOS-IDF which reduces the values of these words, since it is not as important as verb.

thumbnail
Table 5. Example of weighting method using classical TF-IDF and modified TFPOS-IDF.

https://doi.org/10.1371/journal.pone.0230442.t005

3.3.2 Word2vec.

The first proposed feature TFPOS-IDF focuses on giving verbs higher priority over other words. Focuses on the verbs might not be enough. Therefore, it is worthwhile to try extracting another feature, which is the semantic feature word embedding; to the best of our knowledge has not been applied yet in the question classification based on Bloom’s taxonomy. Since the context of the question might carry an important information, this may help in determining the level of question.

This study used pre-trained word2vec, which trains approximately one hundred billion tokens using Google News dataset. The word vector for each word consists of floating-point values, and the meanings of these feature values are hard to explain, except it denotes that the words with similar meaning have a similar vector representation. For example, the top five words that have a similar vector to the word ‘Examine’ are ‘Scrutinize, Investigate, Determine, Assess, Evaluate’.

Fig 3 shows an example of converting question into vector representation. The tokenized question that was produced after the pre-processing step, will be converted into the matrix of floating-point values with the size of 300 dimensions. Gensim library is used to retrieve the word vectors from the word2vec. Lastly, if the word does not exist on the word2vec, then vectors will be initialized by zeros.

thumbnail
Fig 3. Example of converting question into a word vector.

https://doi.org/10.1371/journal.pone.0230442.g003

As mentioned previously, the stemming process is performed only in the TFPOS-IDF step. The question word in this step is not stemmed. The reason behind keeping the word as it is without stemming, is that the pre-trained word2vec, i.e. (Google News) is not pre-processed before the process of learning and establishing the word vectors, and it builds the relationships between these vectors. For example, there is no word vector for the stemmed word ‘recal’. But on the other hand, there are word vectors for ‘recall’, ‘recalling’, and ‘recalls’ which are not totally equal but approximately have similar vectors. The strength of these features lies in producing similar word vectors for the words that have similar meanings, in low dimensional vectors representation. After retrieving the words vector for each question in the corpus, it is a time to combine them with the improved feature TFPOS-IDF.

3.3.3 Combination of word2vec and TFPOS-IDF (W2VTFPOS-IDF).

The aim of this phase is to produce a question with high-quality feature vectors representation by combining the two proposed features, discussed before in the previous two sections. [32] used word2vec with TF-IDF to represent feature vector for question classification in the question answering system, by multiplying the term value of TF-IDF with the word vector of that term. Then after that, all the terms vector of the question is summed together to produce a single vector. This study uses the same approach as demonstrated in Eq (9). (9) where the term t belongs to the document d. The output of this phase is a dense vector, i.e. low-dimensional vector. The size of the vector is 300 dimensions since it is based on the pre-trained word2vec with the dimensions of 300. The advantage of representing question vector in the dense representation (low dimensional) over sparse representation (high dimensional) lies in decreasing the complexity of the classification process [36]. Fig 4 demonstrate an example of producing question vectors by combining word2vec with TFPOS-IDF.

Now the questions vectors W2V-TFPOSIDF are ready to be fed into the classification algorithms, which will be implemented in the next section.

3.4 Classification

Three of the most common supervised machine learning classification algorithms are used, which are the and K-Nearest Neighbour (KNN), Logistic Regression (LR), and Support Vector Machine (SVM).

3.4.1 K-Nearest Neighbours (KNN).

K-Nearest Neighbours (KNN) is one of the simplest algorithms in machine learning. KNN performs well with low-dimensional vectors as shown in[36,47]. This algorithm is dominated by three crucial components [48]; (K) which is the number of nearest neighbours, the labelled training set, and how to estimate the distance among the points.

The way K-NN behave is by looking to the K nearest neighbours among the labelled training questions in which the similarity between a test question and K training questions is calculated to predict the class of the test question. The class or the category of the test question is assigned based on the highest score (majority class) of the K training questions. For example, for the Question X, if the algorithm found five nearest neighbours to X in which one belongs to the Knowledge class, one belongs to the Analysis class, two belongs to the Application class, and one belongs to the Evaluation class, then the question X is assigned to the Application class.

3.4.2 Logistic Regression (LR).

Logistic Regression classifier is based on logistic function, also known as the sigmoid function, which has been used in several text classification applications [49,50] and shows significant results. The idea behind it is to find the relationship among features and specific output. Logistics regression algorithm applied one versus all approach for the multiclass classification task. For example, to predict the question class, six binary classification problem are considered i.e. whether the class is knowledge, comprehension and so on with all classes. Following this, the Maximum Likelihood Estimation is used to assign the predicted class and it is implemented by the following equation [51]: (10)

3.4.3 Support Vector Machine (SVM).

Support Vector Machine (SVM) was proposed by [52] at AT&T Bell Laboratories. It is widely used in text classification problems [53]. The idea behind SVM is to find an appropriate hyperplane that best split-up the two sets of data from each other, by maximizing the margin between the hyperplane and the set of data points nearest to it. In-text classification SVM with linear kernel perform well [54]. Therefore, this study will use it.

Given N data points from questions in the training set , where xi is the i-th input features (i.e. pattern), and yi is the i-th output class label. The SVM linear kernel equation can be written as the following [55]: (11) where n is the number of support vectors, xi is the pattern of i-th question, yi is the class label of the i-th question, and θ(x,xi) is linear kernel function, which can be written as: (12)

3.5 Evaluation metrics

To measure the effectiveness of the proposed model, the weighted recall, precision and F1-mueasre will be calculated. In order to define these metrics, the terms True Positive (TP), False Positive (FP) and False Negative (FN) that must be introduced. TP is the occurrence of instances a classifier correctly classified to the suitable class. FP is the occurrence of instances incorrectly classified. FN is the occurrence of instances that have not been classified. The value of recall, precision, and F1-measure varies from 0 to 1. The closer value to zero of recall, precision, and F1-measure implies the poor performance. Whereas, the closer value to 1 denotes a good performance.

Recall metric measures the perfection of classifiers by computing this formula (13)

Precision metric measures the fineness of classifiers by calculating this equation (14)

Finally, the F-measures or also known as F-score, is calculated by (15) When the value of β = 1 that represents F1-measure, where the equation will be: (16)

Since the training data and test data are randomly selected, and usually imbalanced, the weighted recall, precision, and F1-measure will be calculated. Scikit-learn defined the term weighted “Calculate metrics for each label, and find their average, weighted by support (the number of true instances for each label). This alters ‘macro’ to account for label imbalance; it can result in an F-score that is not between precision and recall..

The equations for calculating the weighted precision, weighted recall and weighted F1-measure are: (17) (18) (19)

4. Experiment setting

In order to evaluate the effectiveness of the proposed model, several experiments had been conducted, with three features and three classifiers. The first feature was the traditional statistical feature TF-IDF. The second feature was the modified feature TFPOS-IDF. The purpose of mentioning the traditional feature TF-IDF was to compare it with the improved feature TFPOS-IDF, in order to check whether the improvement is significant or not. The third feature W2V-TFPOSIDF is word2vec, which was weighted by the improved feature TFPOS-IDF. All these features were fed into three classifiers KNN, LR, and SVM with a linear kernel, which were implemented using Scikit-learn library in Python with the default settings. The default value of K in the KNN classifier is five, and the default value of the penalty is l2 in LR. The default setting of the parameter C in SVM is 1. However, the experiments were run 15 times; during each run, the training and testing set were randomly picked up. The dataset was divided into 80% training set and 20% test set, for the first dataset, the size of the training set was 112 and 29 for the testing set. The size of the training set in the second dataset is 480, and the testing set is 120.

4.1 Result of KNN

This section discusses the experiments result that has been obtained from the collected dataset and Yahya et al. (2012) dataset using the KNN classifier in terms of weighted recall, precision, and F1-measure for each Bloom class. The experiments measured the performance of KNN with three features TF-IDF, TFPOS-IDF, and W2V-TFPOSIDF.

It was observed from the result demonstrated in Table 6 that the modified feature TFPOS-IDF performs better than the traditional feature TF-IDF from all aspects. In addition, the proposed features W2V-TFPOSIDF outperform the improved feature by 5.2% in terms of F1-measure. On the other hand, the score of recall, precision, and F1-measure for the application class with all features are low compared to other classes. This is due to the small number of instances which belongs to the application level in the collected dataset. Therefore, the learning algorithm KNN misclassified the questions that belong to the application level.

thumbnail
Table 6. Results of using KNN with TF-IDF, TFPOS-IDF, W2V-TFPOSIDF for the collected dataset.

https://doi.org/10.1371/journal.pone.0230442.t006

Based on the results demonstrated in Table 7, it can be observed that the classifier learnt more about Bloom’s cognitive levels using the modified feature TFPOS-IDF, since the performance has increased from all aspects. The classification results of using the proposed feature W2V-TFPOSIDF outperforms the modified feature TFPOS-IDF. The result with this dataset is better than the obtained result from the collected dataset, since the size of Yahya et al. (2012) dataset is larger than the collected dataset. Moreover, the number of the questions in Yahya et al. (2012) dataset is distributed evenly over all classes, unlike the collected dataset.

thumbnail
Table 7. Results of using KNN with TF-IDF, TFPOS-IDF, W2V-TFPOSIDF for the Yahya et al. (2012) dataset.

https://doi.org/10.1371/journal.pone.0230442.t007

4.2 Result of LR

This section discusses the results of the experiment that have been obtained with the LR classifier in terms of the weighted recall, precision, and F1-measure for each Bloom class. Table 8 demonstrates the results of the collected dataset. Table 9 shows the results of Yahya et al. (2012) dataset.

thumbnail
Table 8. Results of LR with TF-IDF, TFPOS-IDF, W2V-TFPOSIDF for the collected dataset.

https://doi.org/10.1371/journal.pone.0230442.t008

thumbnail
Table 9. Results of LR TF-IDF, TFPOS-IDF, W2V-TFPOSIDF for Yahya et al. (2012) dataset.

https://doi.org/10.1371/journal.pone.0230442.t009

Tables 8 and 9 show the performance of the LR classifier with both datasets in which the W2V-TFPOSIDF feature produces a better result than TFPOS-IDF and TF-IDF.

4.3 Result of SVM

This section discusses the results of the experiment that have been obtained with the SVM classifier in the terms of the weighted recall, precision, and F1-measure for each Bloom class using both dataset; collected dataset, and Yahya et al. (2012) dataset.

As shown in Table 10 the SVM classifier with the modified feature TFPOS-IDF outperforms the traditional feature TF-IDF in terms of all the evaluation metrics recall, precision and F1-measure, as well as W2V-TFPOSIDF which outperform the two other features.

thumbnail
Table 10. Results of SVM with TF-IDF, TFPOS-IDF, W2V-TFPOSIDF for the collected dataset.

https://doi.org/10.1371/journal.pone.0230442.t010

The detailed results of using SVM classifier with Yahya et al. (2012) dataset have been shown in Table 11, where TFPOS-IDF outperforms TF-IDF, and W2V-TFPOSIDF Superior both features.

thumbnail
Table 11. Results of SVM TF-IDF, TFPOS-IDF, W2V-TFPOSIDF for Yahya et al. (2012) dataset.

https://doi.org/10.1371/journal.pone.0230442.t011

Based on the results from Tables (6–11), it can be concluded that LR and SVM approximately produce similar results, and achieve better performance compared to KNN in classifying questions from different domains in accordance with Bloom’s taxonomy, using both datasets.

5. Statistical test

The statistical test, the t-test can be used to decide whether two samples of data are significantly different from each other. In order to check the significance of the first modified feature TFPOS-IDF and the second proposed feature W2V-TFPOSIDF, the weighted F1-measure values of each run were used. This section discusses the results of the significant test in two phases, using both the datasets and the three classifiers KNN, LR and SVM. In the first phase, the first modified feature TFPOS-IDF was checked against the traditional feature TF-IDF, to check if the proposed improvement was significant or otherwise. The second phase, the TFPOS-IDF was checked against the W2V-TFPOSIDF to check the significance of adding TFPOS-IDF to the word2vec.

A two-sample t-test of the null hypothesis was used. The null hypothesis meant that the value of the hypothesized mean difference was equal to zero, which assumed that there was no difference between the two samples. The value of alpha P-value was equal to 0.05. The idea behind t-test was to check if the obtained value of alpha was less than the assumed one (i.e. P-value = 0.05), if so then the null hypothesis would be rejected, which means there was a significant difference between the two samples. Otherwise, if the P-value was greater than or equal to the assumed one, then the null hypothesis was accepted, which meant that there was no significant difference between the two samples. The alpha values of the t-test are shown in Table 12.

As previously mentioned, if the P-value of the t-test is less than 0.05, then there is a significant difference between the two samples. Therefore, from the results shown in Table 11, it is clear that in all cases, the modified feature TFPOS-IDF is statistically significant compared to the classical one TF-IDF. In addition, the proposed feature W2V-TFPOSIDF is statistically significant compared to the modified feature TFPOS-IDF.

6. Conclusion

In conclusion, this paper proposed a method to classify open-domain questions based on Bloom’s taxonomy cognitive domain. In order to check the robustness of the proposed model, two datasets with three classifiers have been examined. In addition, the statistical test shows that there is a significant improvement between TF-IDF versus TFPOS-IDF and TFPOS-IDF versus W2V-TFPOSIDF. This method can be useful for educators and lecturers to analyze the exam question to fulfill the requirement for the different levels of education such as Bachelor’s degree or Master’s level. For example, the exam paper at a Master’s level usually contains the higher thinking order questions more than the lower thinking order. In addition, it can be effectively used throughout different kinds of applications such as automatic test generation systems, intelligent tutoring systems and even more in serious game such as in [56]. Nevertheless, further study could improve, enhance or extend this work by expanding the dataset through adding different kinds of questions. The dataset used in this study only consists of open-ended questions, in which there are neither multiple choices nor true or false questions. Most of the previous studies are conducted with an English dataset where there is a lack of studies in other languages. Since there is no public benchmark dataset that contains real exams questions labeled based on Bloom’s taxonomy, providing one would be beneficial to relevant researchers in this area.

Supporting information

References

  1. 1. Swart AJ. Evaluation of final examination papers in engineering: A case study using Bloom’s taxonomy. IEEE Trans Educ. 2010;
  2. 2. Haris SS, Omar N. Bloom’s taxonomy question categorization using rules and N-gram approach. J Theor Appl Inf Technol. 2015;76: 401–407.
  3. 3. Osman A, Yahya AA. Classification of exam questions using linguistically-motivated features: A case study based on bloom’s taxonomy. The Sixth International Arab Conference on Quality Assurance in Higher Education (IACQA’2016). 2016.
  4. 4. Kennedy D. Writing and using learning outcomes: a practical guide. University College Cork. Cork; 2006.
  5. 5. Abduljabbar D, Omar N. Exam Questions Classification Based on Bloom’s Taxonomy Cognitive Level Using Classifiers Combination. J Theor Appl Inf Technol. 2015;78: 447–455.
  6. 6. Mohammed M, Omar N. Question Classification Based on Bloom’s Taxonomy Using Enhanced TF-IDF. Int J Adv Sci Eng Inf Technol. 2018;8: 1679–1685.
  7. 7. Yahya AA. Swarm intelligence-based Approach for Educational Data Classification. J King Saud Univ—Comput Inf Sci. 2017;
  8. 8. Jayakodi K, Bandara M, Meedeniya D. An Automatic Classifier for Exam Questions with WordNet and Cosine Similarity. Moratuwa Engineering Research Conference (MERCon), 2016 IEEE. 2016. pp. 12–17.
  9. 9. Kusuma SF, Suahaan D, Yuhana UL. Automatic Indonesia’s Questions Classification Based On Bloom ‘s Taxonomy Using Natural Language Processing. nformation Technology Systems and Innovation (ICITSI), 2015 International Conference on IEEE. 2015. pp. 0–5.
  10. 10. Sangodiah A, Ahmad R, Fatimah WAN, Ahmad WAN. Taxonomy Based Features In Question Classification Using Support Vector Machine. 2017;95: 2814–2823.
  11. 11. Hussein A. A Query Expansion Method For Classifying of Examination Questions Into Bloom’s Taxonomy Cognitive Level. Master Dissertation. Universiti Kebangsaan Malaysia. 2017.
  12. 12. Yang L, Li C, Ding Q, Li L. Combining lexical and semantic features for short text classification. Procedia Comput Sci. 2013;22: 78–86.
  13. 13. Wang P, Xu B, Xu J, Tian G, Liu CL, Hao H. Semantic expansion using word embedding clustering and convolutional neural network for improving short text classification. Neurocomputing. 2016;174: 806–814.
  14. 14. Sangodiah A, Ahmad R, Fatimah W, Ahmad W. A Review in Feature Extraction Approach in Question Classification Using Support Vector Machine. 2014 IEEE Int Conf Control Syst Comput Eng. 2014; 536–541. 10.1109/ICCSCE.2014.7072776
  15. 15. Chang WC, Chung MS. Automatic applying Bloom’s taxonomy to classify and analysis the cognition level of english question items. 2009 Jt Conf Pervasive Comput JCPC 2009. 2009; 727–733. 10.1109/JCPC.2009.5420087
  16. 16. Omar N, Haris SS, Hassan R, Arshad H, Rahmat M, Zainal NFA, et al. Automated Analysis of Exam Questions According to Bloom’s Taxonomy. Procedia—Soc Behav Sci. 2012;59: 297–303.
  17. 17. Kim Y. Convolutional Neural Networks for Sentence Classification. 2014;
  18. 18. Dahou A, Xiong S, Zhou J, Haddoud MH. Word Embeddings and Convolutional Neural Network for Arabic Sentiment Classification. 2016; 2418–2427.
  19. 19. Hutchison D, Mitchell JC. Theory and Practice of Natural Computing. 1973;3: 58–70.
  20. 20. Yahya AA, Toukal Z, Osman A. Bloom’s Taxonomy—Based Classification for Item Bank Questions Using Support Vector Machines Bloom’s Taxonomy—Based Classification for Item Bank. Modern Advances in Intelligent Systems and Tools, SCI 431. Springer-Verlag Berlin Heidelberg 2012; 2012. pp. 135–140. https://doi.org/10.1007/978-3-642-30732-4
  21. 21. Sangodiah A, Ahmad R, Ahmad WFW. Integration of machine learning approach in item bank test system. 2016 3rd Int Conf Comput Inf Sci ICCOINS 2016—Proc. 2016; 164–168. 10.1109/ICCOINS.2016.7783208
  22. 22. Yusof N, Hui CJ. Determination of Bloom’s cognitive level of question items using artificial neural network. Proc 2010 10th Int Conf Intell Syst Des Appl ISDA’10. 2010; 866–870. 10.1109/ISDA.2010.5687152
  23. 23. Yahya AA, Osama A. Automatic Classification of Questions into Bloom’s Cognitive Levels Using Support Vector Machines. The International Arab Conference on Information Technology, Naif Arab University for Security Science (NAUSS). 2011. pp. 1–6.
  24. 24. Osadi KA, Fernando MGNAS, Welgama W V. Ensemble Classifier based Approach for Classification of Examination Questions into Bloom ‘ s Taxonomy Cognitive Levels. 2017;162: 1–6.
  25. 25. Ramos J. Using TF-IDF to Determine Word Relevance in Document Queries. Proc first Instr Conf Mach Learn. 2003; 1–4.
  26. 26. Xu R. POS Weighted TF-IDF Algorithm and its Application for an MOOC Search Engine. Audio, Language and Image Processing (ICALIP), 2014 International Conference on IEEE. 2014. pp. 868–873.
  27. 27. Domeniconi G, Moro G, Pasolini R, Sartori C. A Study on Term Weighting for Text Categorization: A Novel Supervised Variant of tf.idf. Proc 4th Int Conf Data Manag Technol Appl. 2015; 26–37. 10.5220/0005511900260037
  28. 28. Chen K, Zhang Z, Long J, Zhang H. Turning from TF-IDF to TF-IGM for term weighting in text classification. Expert Syst Appl. 2016;66: 1339–1351.
  29. 29. White HD. Bag of works retrieval: TF*IDF weighting of works co-cited with a seed. Int J Digit Libr. 2017; 1–11.
  30. 30. Yahya AA, Osman A, Taleb A, Alattab AA. Analyzing the Cognitive Level of Classroom Questions Using Machine Learning Techniques. Procedia—Soc Behav Sci. 2013;97: 587–595.
  31. 31. Yahya A, Osman A, Alattab A. Educational data mining: A case study of teacher’s classroom questions. Intelligent Systems Design and Applications (ISDA), 2013 13th International Conference on IEEE. 2013. pp. 92–97.
  32. 32. Minai A, Razzaghnoori M, Sajedi H, Jazani IK. Question classification in Persian using word vectors and frequencies. Cogn Syst Res. 2018;47: 16–27.
  33. 33. Lioma C, Blanco R. Part of Speech Based Term Weighting for Information Retrieval. 31th European Conference on IR Research, ECIR. Toulouse, France: springer; 2009. pp. 412–423.
  34. 34. Jovanovska J, Zdravkova K. Information Retrieval with Reinforced Word Classes. 2017;
  35. 35. Lioma C, Blanco R. Part of Speech Based Term Weighting for Information Retrieval. 2017;
  36. 36. Zhu L, Wang G, Zou X. A Study of Chinese Document Representation and Classification with Word2vec. Computational Intelligence and Design (ISCID), 2016 9th International Symposium. IEEE; 2016. pp. 298–302. 10.1109/ISCID.2016.74
  37. 37. Sangodiah A, Muniandy M, Heng LE. Question classification using statistical approach: A complete review. J Theor Appl Inf Technol. 2015;71: 386–395.
  38. 38. Mikolov T, Corrado G, Chen K, Dean J. Efficient Estimation ofWord Representations in Vector Space. arXiv Prepr arXiv13013781. 2013; 1–12.
  39. 39. Wang P, Xu J, Xu B, Liu C, Zhang H, Wang F, et al. Semantic Clustering and Convolutional Neural Network for Short Text Categorization. 2015; 352–357.
  40. 40. Jang B, Kim I, Kim JW. Word2vec convolutional neural networks for classification of news articles and tweets. PLoS One. 2019;14: e0220976. pmid:31437181
  41. 41. Mikolov T, Sutskever I, Chen K, Corrado G, Dean J. Distributed Representations of Words and Phrases and their Compositionality. Adv Neural Inf Process Syst. 2013;cs.CL: 3111–3119.
  42. 42. Lilleberg J. Support Vector Machines and Word2vec for Text Classification with Semantic Features. 2015; 136–140.
  43. 43. Abduljabbar D. Exam Questions Classification Based on Bloom’s Taxonomy Cognitive Level Using Classifiers Combination. 2015.
  44. 44. Tian Y, Lo D. a Comparative Study on the Effectiveness of Part-of-Speech Tagging Techniques on Bug Report. 2015 IEEE 22nd Int Conf Softw Anal Evol Reengineering. 2015; 570–574.
  45. 45. Hingu D, Shah D, Udmale SS. Automatic text summarization of Wikipedia articles. Proc—2015 Int Conf Commun Inf Comput Technol ICCICT 2015. 2015; 15–18. 10.1109/ICCICT.2015.7045732
  46. 46. Scherer R, Rao R. Non-Manual Control Devices. Handb Res Pers Auton Technol Disabil Informatics. 2011; 233–250.
  47. 47. Wang J-Y, Kuo M-F, Han J-C, Shih C-C, Chen C-H, Lee P-C, et al. A Telecom-Domain Online Customer Service Assistant Based on Question Answering with Word Embedding and Intent Classification. 2017; 17–20.
  48. 48. Awad M, Khanna R. Efficient learning machines: Theories, concepts, and applications for engineers and system designers. Effic Learn Mach Theor Concepts, Appl Eng Syst Des. 2015; 1–248.
  49. 49. Pranckevičius T, Marcinkevičius V. Comparison of Naive Bayes, Random Forest, Decision Tree, Support Vector Machines, and Logistic Regression Classifiers for Text Reviews Classification. Balt J Mod Comput. 2017;5: 221–232.
  50. 50. Alsmadi I, Hoon GK. Term weighting scheme for short-text classification: Twitter corpuses. Neural Comput Appl. 2019;31: 3819–3831.
  51. 51. Connor BO. Logistic regression classifiers Naive Bayes: bag of words. 2017;
  52. 52. Cortes C, Vapnik V. Support Vector Networks. Mach Learn. 1995;20: 273–297.
  53. 53. Al-Saffar A, Awang S, Tao H, Omar N, Al-Saiagh W, Al-bared M. Malay sentiment analysis based on combined classification approaches and Senti-lexicon algorithm. PLoS One. 2018;13: 1–18. pmid:29684036
  54. 54. Pota M, Esposito M, De Pietro G. A forward-selection algorithm for SVM-based question classification in cognitive systems. Intelligent Interactive Multimedia Systems and Services. Cham: Springer; 2016.
  55. 55. Nirob SMH, Nayeem MK, Islam MS. Question classification using support vector machine with hybrid feature extraction method. 2017 20th Int Conf Comput Inf Technol. 2017; 1–6. 10.1109/ICCITECHN.2017.8281790
  56. 56. Rasim, Z.R. Langi A, Rosmansyah Y, Munir. Generation Quiz with Genetic Algorithm Based on Bloom’s Taxonomy Classification in Serious Game Based Virtual Environments. 2016 Int Conf ICT Smart Soc Surabaya, 20–21 July 2016 ISBN 978-1-5090-1620-4. 2016; 20–21.