Skip to main content
Advertisement
Browse Subject Areas
?

Click through the PLOS taxonomy to find articles in your field.

For more information about PLOS Subject Areas, click here.

  • Loading metrics

Feature Selection Using Information Gain for Improved Structural-Based Alert Correlation

Abstract

Grouping and clustering alerts for intrusion detection based on the similarity of features is referred to as structurally base alert correlation and can discover a list of attack steps. Previous researchers selected different features and data sources manually based on their knowledge and experience, which lead to the less accurate identification of attack steps and inconsistent performance of clustering accuracy. Furthermore, the existing alert correlation systems deal with a huge amount of data that contains null values, incomplete information, and irrelevant features causing the analysis of the alerts to be tedious, time-consuming and error-prone. Therefore, this paper focuses on selecting accurate and significant features of alerts that are appropriate to represent the attack steps, thus, enhancing the structural-based alert correlation model. A two-tier feature selection method is proposed to obtain the significant features. The first tier aims at ranking the subset of features based on high information gain entropy in decreasing order. The‏ second tier extends additional features with a better discriminative ability than the initially ranked features. Performance analysis results show the significance of the selected features in terms of the clustering accuracy using 2000 DARPA intrusion detection scenario-specific dataset.

1. Introduction

Intrusion detection systems (IDS) work as the "second line of defense" for computer and network systems [1]. Intrusion detection systems employ one intuitive rule that intrusive patterns are noticeable and they are unusual to regular communication. Generally, an IDS is categorized to be a host-based or network-based system depending on its monitoring capability. Host-based IDSs focus on monitoring the individual host/computer with regard to the internal activities and statuses. It cannot detect intrusion across the network [2]. Network-based IDSs (NIDSs) detect intrusions over the network by examining the packets arriving into the network. Since attempted intrusions can occur via the network, an NIDS needs to monitor different actions triggered on multiple hosts in order to build adequate evidence. In this case, NIDSs are made to handle huge data packets and communications in large networks as compared to host-based IDSs. A NIDS collects and analyses network information to check if there are actions violating security strategies [3]. The NIDS triggers alert to the network operator to take action against the suspicious activities. However, even a single NIDS generates a huge amount of alerts that overwhelms the operators [4] [5]. Most of the generated alerts have irrelevant features which result in slow training and testing correlation processes, higher resource consumption, lower accuracy and higher performance costs [6]. Furthermore, inappropriate features lead to the less accurate discovery of the attack steps. The pattern of attack steps taken by the attacker is discovered when a similar pattern of alerts are recognized and grouped. Therefore, this paper aims to identify appropriate features to achieve high accuracy in the identification of the attack steps. After that, the list of the attack steps from the alerts patterns can be determined accurately by clustering the most significant features of the alerts. The selected features are evaluated in terms of clustering accuracy. This paper proposes a 2-tier feature selection method, namely, feature ranking (first tier) and additional feature (second tier). The feature ranking tier ranks the features based on high information gain entropy while the additional feature tier provides extended additional features with better discriminative ability.

This paper is organized as follows: Section 2 provides an overview of some related research and presents the necessary background information regarding feature selection in alert correlation. Section 3 presents an overview of the proposed feature selection and discusses the experimental results. Finally, in Section 4 the paper is concluded.

2. Related Work

Feature selection has been widely applied in many domains, such as text categorization [7], genomic analysis [8], intrusion detection [9][10] and bioinformatics [11][12].

In a complex classification domain, such as intrusion detection, features may contain a false correlation that hinders the learning task to be processed [13]. Some features may be irrelevant and others may be redundant [14]. These extra features can increase computational time and can have an impact on the system accuracy [13]. Therefore, selecting important features from input data leads to the simplification of a problem, and faster and more accurate detection rates [9]. For this reason, alert correlation researchers have tried to select the relevant features of alerts. However, the relevant features were manually selected with different researchers selecting different features based on their knowledge and experience. For example, in alert clustering, seven different features are selected by Tjhai et al. [15] and Man et al. [16]. Tjhai et al. [15] proposed a framework that contains four phases: feature extraction, alarm aggregation, cluster analysis and classification. In the phase of feature extraction, seven alerts features (attributes) such as the number of alerts, number of signatures, port number, protocol, priority, time interval and the number of events are evaluated and chosen to represent the value of each input vector in the last phase. Man et al. [16] proposed ISODATA algorithm or the purpose of solving flood and duplicated alarms of IDS effectively. The essence of their algorithm is to generate an initial class as "seed", and then iterate clustering automatically according to some discriminate rule. DARPA 1999 is used to test their algorithm. Considering that there were a lot of useless information in the original alert, they selected a part of the attributes as their main characteristic attributes in their aggregation algorithm and represented them as a tuple containing the attributes: alert id, alert type, SrcIP, DestIP, SrcPort, DestPort, and Time. Mohamed et al. [17] extracted three attributes from the alerts (destination IP, signature type or id and timestamp) and applied these attributes to the MD5 hash function. The MD5 generates a unique hash value that is used for the initial clustering process. Meanwhile, Siraj [18] proposed a novel hybrid clustering model based on Improved Unit Range (IUR), Principal Component Analysis (PCA) and unsupervised learning algorithm (Expectation Maximization) to aggregate similar alerts and to reduce the number of alerts. Three different attributes (source port, destination port, and alert type) from the DARPA 2000 dataset were selected and represented as a vector. Shittu et.al. [19] propsed A comperhensive System for Analysing Intrusion Alerts (ACSAnIA). it contains seven components which are: (1) Offline Correlation (2) Online Correlation (3)Meta alert Comparison (4) Meta–alert Prioritisation (5)Meta–alert Clustering (6) Attack Pattern Discovery and (7) Reporting System. The ACSAnIA system uses six of the alerts attributes where as: alert’s timestamp, source IP, source port, destination IP, destination port and intrusion type. Furthermore, based on Ramaki et al.[20] an efficient framework for alert correlation in Early Warning Systems (EWSs) is proposed. An important process in EWSs is the analysis and correlation of alerts aggregated from the installed sensors. The authors mentioned that an alert consists of some features based on specific attributes of network traffic. The most important features that are used for the their alert correlation process are: source IP address, destination IP address, source and destination port numbers, intrusion type or alert type, attack severity and timestamp.

The literature review indicates that there are no standards or specific features used in alert clustering; thus, every researcher selects a different number of feature subsets based on their own experience. Therefore, this paper focuses on applied automated feature selection that improves clustering accuracy and presents accurate attack steps.

3. Feature Selection

The reason for selecting the important and significant features is to represent the attack steps from the alerts pattern correctly and improve the accuracy of the Structural based Alert Correlation (SAC). This section describes the two-tier feature selection, i.e., feature ranking and additional feature. The feature ranking stage employs Information Gain algorithm (IG) that uses a filtering approach. The stage aims at ranking subsets of features based on high information gain entropy in decreasing order. Meanwhile, the additional feature stage is based on the work of Ren et. al. [21] where they mention that identifying relationships between alerts essentially needs to analyses the alerts' attributes, and extracting the basic attributes may not be sufficient to fully discover the relationship between the alerts. Therefore, the aim of this stage is to extend additional features that contribute to the relationship between alerts with a better discriminative ability than the initially ranked features. Fig 1 shows the feature selection procedure that is adopted in this research. The effectiveness of the reduced feature subsets was evaluated on SAC.

3.1 Feature Ranking

Ranking methods are used due to their simplicity and the fact that good success has been reported for practical applications [22]. A suitable ranking criterion is used to score the variables and a threshold is used to remove variables below the threshold. The basic property of feature ranking is to identify the relevance of the features. It essentially states that if a feature is to be relevant it can be independent of the input data but cannot be independent of the class labels, i.e., the feature that has no influence on the class labels can be discarded [22]. The main reason behind the application of feature ranking in this study is based on this property, which ranks the feature that has an influence on the class labels.

Hyper alerts, as well as low-level alerts, can be distinguished based on the type of alert that denotes a certain attack class/step [21]. Furthermore, the absence of truth labeled in DRAPA 2000 datasets directed this research towards proposing alert types as class labels to represent accurate attack steps. Therefore, based on the feature ranking property, high ranking features are most relevant and significant for alert types.

As mentioned earlier, the feature ranking was implemented using Information Gain (IG). IG is frequently employed as a term-goodness criterion in the field of machine learning [23]. It is measured based on the entropy of a system, i.e., of the degree of disorder of the system. Therefore, the entropy of a subset is a fundamental calculation to compute IG. For feature ranking purpose, IG is implemented in all four files of datasets. Twenty-six features are applied to IG for feature ranking. This study manually identified some of those features that have network meaning and is based on the XML file presented in Fig 2.

Referred to Fig 2 the alert is uniquely identified by the alert ident feature. The source and target feature describe the node and service of the sender and the receiver respectively. The node contains the IP address and its category, while service holds the port number and its corresponding protocol. The alert type is given by the Classification name feature. Specifically, this alert simply represents a stealth scan attack via port 23 from 194.007.248.153 to 172.016.113.148 via port 22. Based on this information, eight features have been identified from the XML documents as listed in Table 1 and the rest have been labelled without network meaning. Table 2 is an example of all features in DMZ network for scenario one.

thumbnail
Table 1. Attributes of an alert extracted from the XML document.

https://doi.org/10.1371/journal.pone.0166017.t001

IG looks at each feature in isolation, computes its information gain and measures how important and relevant it is to the class label (alert type). Computing the information gain for a feature involves computing the entropy of the class label (alert type) for the entire dataset and subtracting the conditional entropies for each possible value of that feature. The entropy calculation requires a frequency count of the class label by feature value. In more details, all instances (alerts) are selected with some feature value v, then the number of occurrences of each class within those instances are counted, and the entropy for v is computed. This step is repeated for each possible value v of the feature. The entropy of a subset can be computed more easily by constructing a count matrix, which tallies the class membership of the training examples by feature value. The algorithm of IG implementation is given in Fig 3.

3.1.1 Results on Feature Ranking.

Tables 3, 4, 5 and 6 show that for each dataset file, the different set of features are ranked differently in decreasing order based on their relevance to class label. The reason behind this variation is that the DDoS attack has five attack phases, in each phase different attack steps occurred. The rank score measures how much each feature relates and contributes to class label. These tables indicate that A_ID, D_port, Priority and S_port which are: alert id, destination port, priority and source port (as highlighted in bold) have obtained the same sequence of order with higher scores than other features in all files of datasets. Therefore, we can conclude that those four features are more related and influential to alert type (class label).

thumbnail
Table 3. Feature ranking using IG on DMZ 1 DARPA 2000 dataset.

https://doi.org/10.1371/journal.pone.0166017.t003

thumbnail
Table 4. Feature ranking using IG on Inside 1 DARPA 2000 dataset.

https://doi.org/10.1371/journal.pone.0166017.t004

thumbnail
Table 5. Feature ranking using IG on DMZ 2 DARPA 2000 dataset.

https://doi.org/10.1371/journal.pone.0166017.t005

thumbnail
Table 6. Feature ranking using IG on Inside 2 DARPA 2000 dataset.

https://doi.org/10.1371/journal.pone.0166017.t006

3.1.2 Evaluation Performance on Feature Ranking.

To evaluate the significance of these ranked features on clustering accuracy that present the attack steps and to find the best algorithm that produces the highest clustering accuracy, three clustering algorithms have been applied both before and after the feature selection method. The algorithms are K-Means, EM, and Hierarchical. Clustering refers to unsupervised learning and for that reason, it has no priori data set information. Therefore, many different cluster validity methods have been proposed without prior class information, these are named internal validity. In contrast, there is a number of external cluster validity for which a priori knowledge of dataset information is required. As noted before, the absence of truth labels in DARPA 2000 datasets for evaluating structured (cluster) based alert correlation leads this research to propose a single class label that is an alert type for external validation of the clustered alerts.

Firstly, the clustering algorithms were performed on the original features, a total of 26, for all datasets before applying IG for feature ranking. The average accuracy rates (AR), which include the percentage of alerts that are accurately clustered, are reported in Figs 4, 5 and 6.

thumbnail
Fig 4. Results of K-means with varying number of clusters.

https://doi.org/10.1371/journal.pone.0166017.g004

thumbnail
Fig 6. Results of Hierarchical with varying number of clusters.

https://doi.org/10.1371/journal.pone.0166017.g006

1) Performance using K-means algorithm

Fig 4 presents that K-means algorithm achieves the best result (37% to 43%) with two and three clusters. In average, for all dataset, the best performance of K-means algorithm is 39.7% AR at 2 clusters. After that, the accuracy is slowly downgraded when the number of clusters is increased. The graph in all figures shows that the performance of K-means is poor, it gives in average AR of 39.7%. In addition, the drawback of this algorithm is that there is no efficient and universal method for identifying the initial partitions; and the centroids are varied with different initial points that lead to different results in different iterations.

2) Performance using EM algorithm

Within the EM algorithm, the highest accuracy (64% to 68.9%) is obtained when the number of clusters is between 7 and 16. When the number of clusters is more than this range, the performance of EM decreases. In average for all dataset, the EM's best performance is 67.5%. The result of EM algorithm is shown in Fig 5. The results show that EM is about 27.8% better than K-means. However, when the number of clusters is increased there is a possibility of EM to produce clustering and an incorrect group of alerts pattern. Hence, the experiments were continued with another algorithm.

3) Performance using Hierarchical algorithm

Meanwhile, Fig 6 shows that Hierarchical cluster gives a slight improvement in its curve compared to EM. It produces a consistent result in the range of (72.3% to 93.2%) when the number of clusters equals the number of alerts type (class label) which are 17 clusters and above in all datasets. In average for all dataset, the hierarchical has the best performance 85.6% AR. With this, hierarchical performs better than K-means and EM with approximately 45.9% and 18.1% improvements respectively. The justification for adopting Agglomerative hierarchical cluster is that the algorithm starts with each data point (feature) as a separate class and then each step of the algorithm involves merging two clusters that are most similar. This point is very useful in alert clustering to identify the accurate attack step.

Table 7 details the best clustering Accuracy Rate (AR) produced by, K-means, and EM and Hierarchical algorithms on all datasets before feature selection. The details are based on results from Figs 46.

thumbnail
Table 7. Summary on AR using K-means, EM and Hierarchical algorithm on all datasets before feature selection.

https://doi.org/10.1371/journal.pone.0166017.t007

Secondly, with the feature ranking that is mentioned above, four significant features which are: alert id, destination port, priority and source port are applied to the same clustering algorithms. The reason for this is to empirically prove that the ranked features improve clustering performance. Figs 79 show clustering accuracy rate with varying number of clusters to find optimum results.

1) K-means performance after feature ranking

In regards to K-Means, this study found that a range of 4 to 6 clusters yield the best performance, namely that of 75.1 percent AR for all datasets. After that, any increase in the cluster numbers implies a decrease in the clustering accuracy. There is a 35.4% improvement when compared to cluster performance before feature ranking. The result of K-means after feature selection is shown in Fig 7.

2) EM performance after feature ranking

88.5% has been obtained in average for all dataset within the EM algorithm when the number of clusters is between 11 and 14. The results show that after feature ranking there is an improvement of about 22.2%. The detailed result of EM after feature ranking is presented in Fig 8.

3) Hierarchical performance after feature ranking

In this experiment, the hierarchical clustering algorithm obtained the highest accuracy rate when the number of clusters equaled the number of alerts type in all datasets. As shown in Fig 9, with DMZ1, 100% AR is obtained when the number of clusters equals the number of alert type (19). Therefore, 19 clusters of attack steps are presented. If the number of clusters is more than 19, the clustering accuracy slowly downgrades. Furthermore, the clustering algorithm gives a high accuracy of 100% with 17, 20 and 22 number of clusters in DZM2 for inside1 and inside2 datasets respectively. Consequently, the improvement of hierarchical clustering algorithm after feature ranking is about 14.6%.

Table 8 summarizes the best clustering accuracy offered by K-means, EM and Hierarchical algorithms on all datasets after feature ranking. The details are based on results from Figs 79. The hierarchical algorithm gives the best result among the studied algorithms.

thumbnail
Table 8. Summary of clustering accuracy using K-means, EM and Hierarchical algorithm on all datasets after feature ranking.

https://doi.org/10.1371/journal.pone.0166017.t008

The reported empirical results on the investigated clustering algorithms lead to several observations and discussions:

1) Observation 1: Hierarchical cluster is a suitable candidate to cluster alerts in the structural-based AC because the algorithm starts with each feature as a separate class or cluster and then merges the clusters or the classes that are more similar. This mean that, each ranked feature (i.e., alert id, destination port, priority and source port) has its own cluster. After that, each cluster of a feature is compared to a class label (alert type) that denotes a certain attack class/step to measure the similarity and relevance between the cluster feature and the class label. Therefore, when the cluster of alert id feature compares to a class label (alert type), the algorithm obtains 100% AR because alert id is a dominant feature to alert type. However, this algorithm considers alert id as the main feature and the validation of the algorithm is based on this feature regardless of other features. For this reason, the investigation of other clustering algorithms to evaluate the significance of the other three features is needed.

2) Observation 2: The empirical results prove that the ranked features which are: alert id, destination port, priority and source port yields the best performances for EM and K-means algorithms, namely that of 88.5% and 75.1% AR respectively for all datasets. However, this improvement is still moderate, and according to Ren et al. [21], selecting the basic alert features may not be sufficient to fully discover these patterns. Furthermore, this research seeks to find accurate features that offer high cluster accuracy and represent the step of the attack accurately. Thus, additional features from the available dataset features which are useful in alert correlation are derived.

The above discussions motivate further investigations and experiments to improve the clustering accuracy. The experiments involve the use of three additional features along with ranked features, to produce an enhanced SAC model. The corresponding results are reported in the next section.

3.2 Additional Feature

IDS alerts features capture intrinsic attack characteristics that are mainly for identifying attack strategy such as the IP address of an alert, its port number, and time when the alert is triggered [21]. While the values of these features are the same for low-level alerts grouped on hyper-alert (except time), their values differ among the hyper alerts of the same type. At the same time, feature values of hyper alerts share common patterns that allow describing the hyper-alert type [21]. Therefore, selecting the basic alert features may not be sufficient to fully discover these patterns. For this reason, additional features from the available dataset features which are useful in alert correlation were derived.

Source_IP, target_IP and time have been added as additional features because the source and destination IP addresses are the key of correlation [24]. Also, the time attributes can help to associate and cluster alerts that occur in short intervals [25]. It is concluded that alert ID, source port, destination port, source IP, destination IP, priority and time are the most significant features that are needed in alert correlation as described in Table 9.

thumbnail
Table 9. The description of significant features of DARPA 2000 dataset.

https://doi.org/10.1371/journal.pone.0166017.t009

Fig 10, Fig 11 and Fig 12 show the performance of the clustering accuracy of K-means, EM and Hierarchical algorithms based on the seven selected features. They show that for each clustering algorithm tested K-means, EM, and Hierarchical, the selected features contribute superior results with an average of 77.9%, 90.6% and 100% AR respectively. As mentioned before, the 100% AR obtained by the hierarchical algorithm is due to the dominant feature of alert id. K-means and EM empirical results prove that the additional features improve the clustering accuracy in the average of 2.8% and 2.1% respectively. The Summary of AR using K-means, FCM and EM algorithm on all datasets is presented in Table 10.

thumbnail
Fig 10. Results of K-means based on the seven selected features.

https://doi.org/10.1371/journal.pone.0166017.g010

thumbnail
Fig 11. Results of EM based on the seven selected features.

https://doi.org/10.1371/journal.pone.0166017.g011

thumbnail
Fig 12. Results of Hierarchical based on seven selected features.

https://doi.org/10.1371/journal.pone.0166017.g012

thumbnail
Table 10. Summary on AR using K-means, FCM and EM algorithm on all datasets.

https://doi.org/10.1371/journal.pone.0166017.t010

4.Discussion

In this paper, the practical results have confirmed that using the proposed features (ranking + additional features) gives significantly better clustering performance for representing attack step. As mentioned before each alert cluster represents an attack step of multi-stages attacks. Table 11 represents the attack steps which are listed alphabetically and a number of alerts for each cluster is shown in the brackets. Based on the hierarchical clustering algorithm, the number of attack steps discovered in each dataset is 19, 22, 17 and 20 respectively. A brief meaning of each attack step is described in Table 12. We observe that the Email_Ehlo is the highest number of alerts in all dataset. This indicates that the attacker is trying hard to access information as much as he/she can from the Simple Mail Transfer Protocol (SMTP) configurations. Since SMTP is used to transfer e-mail messages between computers, this attempt is most likely to find a vulnerable path that the attacker can use to send harmful messages or files. When such intrusion is successful, a lot of Email_Almail_Overflow are detected, indicating that the attacker is trying to overflow the email buffer. Other clusters which have a large number of alerts are TelnetTerminaltype alerts, which indicate the beginning of a telnet session using the reported Terminal type has been detected. Furthermore, a high volume of alerts also precedes the File Transfer Protocol (FTP), for example, FTP_Pass, FTP_Syst, FTP_Put, and FTP_User. Based on Table 12, they are related to a standard network protocol which used to copy a file from one host to another over a TCP/IP-based network, such as the Internet. Additionally, FTP is built on a client-server architecture and utilizes separate control and data connections between the client and the server. Although users need to authenticate themselves using a clear-text sign-in protocol, sometimes they can connect anonymously if the server is configured to allow it. This is most probably the best reason for the attacker to exploit the FTP at the destination host.

thumbnail
Table 11. List of attack steps (clusters) discovered on all dataset.

https://doi.org/10.1371/journal.pone.0166017.t011

thumbnail
Table 12. Description of attack steps based on RealSecure Signatures Reference Guide Version 6.0 (Internet Security Systems.

https://doi.org/10.1371/journal.pone.0166017.t012

Furthermore, For comparison with the proposed features, the features subsets selected by Siraj [18] and features subsets selected by Elshoush [26] were compared. The metrics for comparison is the clustering accuracy. Siraj [18] claimed that three significant features were enough to cluster alerts; these are Alert type, Source port, and destination port. While Elshoush [26] suggested seven features to cluster alerts namely: EventID, times, SrcIPAddress, DestPort, DestIPAddress, OrigEventName, and SrcPort. Table 13 shows the performance comparison among these feature subsets. The second row of Table 13 is the clustering accuracy of K-means and EM of all datasets based on the features subset proposed by Siraj[18]. Meanwhile, the final row is the accuracy rate of the same clustering algorithms for the features that were reported by Elshoush [26]. Performance comparison in Table 13 shows that, overall, the selected features proposed in this research give better clustering results compared to the features proposed by [18] and [26]. Fig 13 and Fig 14 illustrate the comparison in graphical formats.

thumbnail
Fig 13. Comparison on accuracy performance of K-means in all datasets.

https://doi.org/10.1371/journal.pone.0166017.g013

thumbnail
Fig 14. Comparison on accuracy performance of EM in all datasets.

https://doi.org/10.1371/journal.pone.0166017.g014

thumbnail
Table 13. Performance comparison with other feature subsets.

https://doi.org/10.1371/journal.pone.0166017.t013

In conclusion of the performance validation and benchmark of the proposed features, some observations have been outlined:

  1. The Alert ID, Destination Port, and Source Port, which are high impact features in ranking, are the same features used in [18], however, giving lower accuracy than the proposed selected features.
  2. Source IP address, Destination IP address and time which are proposed as additional features with the proposed features ranking are the same features used in [26]. Also, [26] gives better performance than [18] which used the same ranking features only.

5. Conclusion

Clustering and finding relationships between alerts is an important issue since alerts are not significant if they are isolated. The pattern of attack steps taken by the attacker is discovered when the similar pattern of alerts are recognized and grouped based on proper features. Different features were selected manually by previous researchers based on their knowledge experience which lead to less accurate in the identification of attack steps and inconsistent performance of clustering accuracy. This paper focuses on presenting accurate attack steps by proposing a 2-tier feature selection method to select appropriate and significant features. The selected features are evaluated in terms of clustering accuracy. The empirical results show that the selected features can significantly identify accurate attack steps and improve the overall clustering performance.

Acknowledgments

We would like to thank Ministry of Higher Education (MoHE) and Universiti Teknologi Malaysia for funding this work under Potential Academic Staff research grant with reference number PY/2014/02899.

Author Contributions

  1. Conceptualization: TA.
  2. Data curation: TA.
  3. Formal analysis: TA MS AZ.
  4. Funding acquisition: MS.
  5. Investigation: TA.
  6. Methodology: TA.
  7. Project administration: MS AZ.
  8. Resources: TA.
  9. Software: TA MS AZ HE FE.
  10. Supervision: MS AZ HE.
  11. Validation: TA MS AZ HE FE.
  12. Visualization: TA MS AZ HE FE.
  13. Writing – original draft: TA MS AZ HE FE.
  14. Writing – review & editing: TA MS AZ HE FE.

References

  1. 1. Srilatha C, Ajith AB, Johnson PT. Feature deduction and ensemble design of intrusion detection systems, 2005.
  2. 2. Anderson D et al. Safeguard final report: Detecting unusual program behavior using the NIDES statistical component. Computer Science Laboratory. SRI International, Menlo Park, CA, Technical Report (1993).
  3. 3. Yang D, Qi H. A network intrusion detection method using independent component analysis. Pattern Recognition (ICPR), 19th International Conference on. 2008; 1–4.
  4. 4. Debar H, Wespi A. Aggregation and correlation of intrusiondetection alerts. Recent Advances in Intrusion Detection. 2001; 85–103.
  5. 5. Qin X, Lee W. Statistical causality analysis of infosec alert data. Recent Advances in Intrusion Detection. 2003; 73–93.
  6. 6. Wang J-X, Wang Z-Y. Intrusion Alert Analysis Based on Pca and the Lvq Neural Network. Neural Information Processing. 217–224.
  7. 7. Forman G, An extensive empirical study of feature selection metrics for text classification, Journal of Machine Learning Research. 2003; 3:1289–1305.
  8. 8. Inza I, Larranaga P, Blanco R, Cerrolaza A. Filter versus wrapper gene selection approaches in dna microarray domains. Artificial Intelligence in Medicine. 2004; 31:91–103. pmid:15219288
  9. 9. Zainal A, Maarof MA, Shamsuddin SM. Feature Selection Using Rough Set in Intrusion Detection. 2006.
  10. 10. Lee W, Stolfo SJ, Mok KW. Adaptive intrusion detection: a data mining approach. AI Review. 20001; 4(6): 533–567.
  11. 11. Saeys Y, Inza I, Larrañaga P. A review of feature selection techniques in bioinformatics. Bioinformatics. 2007; 23(19): 2507–2517. pmid:17720704
  12. 12. Xing E, Jordan M, Karp R. Feature selection for high-dimensional genomic microarray data. In: Proceedings of the 8th International Conference on Machine Learning. 2001; 601–608.
  13. 13. Bolon-Canedo Veronica, Sanchez-Marono N, Alonso-Betanzos A. Feature selection and classification in multiple class datasets. An application to KDD Cup 99 dataset. Expert Systems with Applications. 2011; 38(5): 5947–5957.
  14. 14. Chandrashekar Girish, Sahin F. A survey on feature selection methods. Computers Electrical Engineering. 2014; 40(1): 16–28.
  15. 15. Tjhai GC, Furnell SM, Papadaki M, Clarke NL. A preliminary two-stage alarm correlation and filtering system using SOM neural network and K-means algorithm. Computers & Security. 2010; 29: 712–723.
  16. 16. Man D, Yang W, Wang W, Xuan S. An alert aggregation algorithm based on iterative self-organization. Procedia Engineering. 2012; 29:3033–3038.
  17. 17. Mohamed AB, Idris NB, Shanmugum B. Alert correlation using a novel clustering approach. In: Proceedings of the International Conference on Communication Systems and Network Technologies (CSNT). 2012; 720–725.
  18. 18. Siraj MM. Hybrid of Structural, Causal and Statistical Model for Intrusion Alert Correlation. University Technology Malaysia. PhD Thesis.
  19. 19. Shittu Riyanat, et al. "Intrusion alert prioritisation and attack detection using post-correlation analysis." Computers & Security.2015. 50., 1–15.
  20. 20. Ramaki Ali Ahmadian, Amini Morteza, and Atani Reza Ebrahimi. "RTECA: Real time episode correlation algorithm for multi-step attack scenarios detection." computers & security .2015: 206–219.
  21. 21. Ren Hanli, Stakhanova N, Ali AG. An online adaptive approach to alert correlation. Detection of Intrusions and Malware, and Vulnerability Assessment. Springer Berlin Heidelberg, 2010; 153–172.
  22. 22. Chandrashekar Girish, Sahin F. A survey on feature selection methods. Computers & Electrical Engineering. 2014; 40(1): 16–28.
  23. 23. Yang Yiming, Pedersen JO. A comparative study on feature selection in text categorization. ICM 1997, 97.
  24. 24. Reuben S et al. Using unsupervised learning for network alert correlation. Advances in Artificial Intelligence. Springer Berlin Heidelberg. 2008; 308–319.
  25. 25. Salim Lalla Fatima, Mezrioui A. Improving the quality of alerts with correlation in intrusion detection. IJCSNS International Journal of Computer Science and Network Security. 2007; 7(12): 210–215.
  26. 26. Elshoush, Tagelsir H, Osman IM. An improved framework for intrusion alert correlation. Proceedings of the World Congress on Engineering. 2012; 1.