0% found this document useful (0 votes)
20 views6 pages

An Anomaly Detection Model Based On One-Class

Uploaded by

ouarme.ar
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PDF, TXT or read online on Scribd
0% found this document useful (0 votes)
20 views6 pages

An Anomaly Detection Model Based On One-Class

Uploaded by

ouarme.ar
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PDF, TXT or read online on Scribd
You are on page 1/ 6

2015 11th International Conference on Mobile Ad-hoc and Sensor Networks

An Anomaly Detection Model based on One-class


SVM to Detect Network Intrusions

Ming Zhang, Boyi Xu, Jie Gong


National Key Laboratory of Science and Technology on Information System Security
Beijing Institute of System Engineering
Beijing, China
[email protected]

Abstract—Intrusion detection occupies a decision position in Intrusion detection was first introduced by Anderson in [3].
solving the network security problems. Support Vector Machines Later, lots of researches have been carried out [4]. Generally,
(SVMs) are one of the widely used intrusion detection techniques. there are two main approaches to conduct intrusion detection:
However, the commonly used two-class SVM algorithms are signature-based detection and anomaly-based detection. The
facing difficulties of constructing the training dataset. That is signature-based detection, also called misuse detection, usually
because in many real application scenarios, normal connection builds a model based on known attacks. Any activity that has
records are easy to be obtained, but attack records are not so. We the corresponding known attack signatures is regarded as an
propose an anomaly detection model based on One-class SVM to intrusion. The signature-based detection model has a good prior
detect network intrusions. The one-class SVM adopts only
knowledge of known attacks, but seldom involves new types of
normal network connection records as the training dataset. But
attacks. Hence, in practice, it could miss a significant amount
after being trained, it is able to recognize normal from various
attacks. This just meets the requirements of the anomaly of real attacks [5]. By contrary, the anomaly detection creates a
detection. Experimental results on KDDCUP99 dataset show that profile from normal behaviors and any violation will be
compared to Probabilistic Neural Network (PNN) and C-SVM, reported as an intrusion. Theoretically, it is capable of detecting
our anomaly detection model based on One-class SVM achieves both known and unknown attacks. Under the current
higher detection rates and yields average better performance in complicated network environment, the anomaly detection is
terms of precision, recall and F-value. much more required and has a better application foreground. In
this paper, we focus on the anomaly detection.
Keywords- intrusion detection; anomaly detection; one-class
SVM; scaling strategy
With the network developing at an unprecedented pace, the
traditional intrusion detection approaches are faced with more
and more challenges. So a lot of new techniques have been
I. INTRODUCTION introduced to conduct intrusion detection [6], among which the
Nowadays, we are living in an information age. We Support Vector Machine (SVM) is one of the widely used
immerse ourselves in the endless joy and great convenience techniques [7, 8]. The standard SVM algorithm was proposed
brought by the Internet. Especially, with the rapid growth of by Corinna Cortes and Vapnik in 1993 and published in 1995
Web applications, everything seems so easy. However, in [9]. A SVM algorithm tries to construct a hyperplane that has
recent years, “attack”, “intrusion” and other similar words the largest distance to the nearest training-data point of any
frequently appear in people’s eyes. Like any greatness, the class in a high or infinite dimensional space, which can be used
Internet is also a double-edged sword. We are enjoying the for classification and other tasks. By using the slack variables
benefits of the Internet, but at the same time, we are suffering and kernel tricks, the SVM guarantees to find the hyperplane
from increasing network threats. The well-known internet that achieves a good separation. Whereas in the actual intrusion
security corporation, Symantec, reminds in its annual Internet detection scenarios, the conventional two-class SVM
Security Threat Report (ISTR) that cybercrime remains algorithms may face some minor problems. For example, in
prevalent and damaging threats from cybercriminals continue many cases, normal network records can be obtained easily, but
to loom over businesses and consumers [1]. Another Web intrusion records are not so. So it is difficult to construct the
security company, Cenzic, reported in 2014 that 96% of the training dataset. Actually, the intrusion detection is not a
tested internet applications had vulnerabilities with a median of straightforward binary classification problem. The attacks can
14 per application, resulting in that hackers are increasingly be divided into many categories. Though some researchers
focusing on and are succeeding with layer 7 (application layer) have tried using multi-class SVM for intrusion detection, the
attacks [2]. These reports show that network security should multi-class SVM is just a combination of many two-class
not be ignored and effective security measures are much SVMs and also encounters the similar problems.
needed.
Given this, we propose to adopt the one-class SVM [10],
Among the important ways to solve security problems, which uses the normal connection records as the training
intrusion detection is an effective and high-profile method. dataset and can recognize normal from various attacks, to

978-1-5090-0329-7/15 $31.00 © 2015 IEEE 102


DOI 10.1109/MSN.2015.40
create an anomaly detection model for network intrusions. We III. ONE-CLASS SVM BASED ANOMALY DETECTION
believe our work will help to use SVM to solve practical In this section, we expound our one-class SVM based
intrusion detection problems. intrusion detection model. We first present the framework of
The rest of this paper is organized as follows. In section 2, the model, and then discuss how each constituent module
we introduce some related work about the intrusion detection, works.
especially some SVM-based models. In section 3, we first
present the framework of our one-class SVM based intrusion A. Framework of one-class SVM model
detection model, and then discuss the implementation details.
Experimental results and performance comparison are
described in section 4. Finally, section 5 concludes this paper

II. RELATED WORK


The research on intrusion detection began from Anderson’s
famous literature [3]. In [3], the author proposed a model
established from statistics of users’ normal behaviors, so as to
find the “masquerader” that deviates from the established
normal model, which laid the foundation of intrusion detection
and revealed the basic idea of anomaly detection. Later
researches on anomaly detection also employ various statistical
methods including multivariate statistics [11], Bayesian
analysis [12], principal component analysis [13], and frequency
and simple significance tests [14]. The signature-based
detection (also called misuse detection) was first introduced by
Denning in [15]. The author proposed an intrusion detection
model that can be regarded as a rule-based pattern matching
system. Both the misuse detection and statistics based anomaly
detection have some limitations, such as low intelligence and
poor ability to adapt to various application scenarios. And
when encountering with larger datasets, the detection results
would become worse [16].
To solve the limitations of above models, a number of
machine learning techniques have been used [17, 18], of which
the most widely used techniques may be Artificial Neural
Networks (ANNs) [19] and Support Vector Machines (SVMs)
[20]. A common practice is to use ANN and SVM to construct
the hybrid model to detect intrusions [21, 22]. In this paper, our
work relates to SVM and ANN is used as a comparison.
Figure 1. Framework of one-class SVM based model
Multi-class SVM is also an alternative in intrusion
detection. In [23], the author applied multi-class SVM Our one-class SVM based intrusion detection model
classifiers, using one-against-one method, for anomaly as well consists of the following two modules:
as misuse detection to identify attacks precisely by type. But
like the two-class SVM, the multi-class SVM is also faced with Module I: Feature extracting module.
the difficulties to construct the training dataset. Feature extracting is the necessary step to make the
detection module work correctly. Our intrusion detection
Some other studies concern combining cluster algorithms model integrates a feature extracting module mainly to extract
with SVM techniques. In [24, 25], a hierarchical clustering useful features from the raw data and then generates
method was applied to preprocess the originally enormous manageable formatted data for the detection module.
dataset to provide a reduced dataset for the SVM training. Thus
the intrusion detection system could greatly shorten the training Module II: One-class SVM module.
time. In this paper, we are more concerned about how to Working as the detection module, one-class SVM involves
improve the detection accuracy, and seldom care about the two procedures. The training procedure accepts the training
learning speed. But the clustering method to reduce the dataset data and generates a decision model. The testing procedure
can also be used in our model. takes both the decision model and the testing data as inputs,
and then produces the detection results.
Based on the related work, we propose an anomaly
detection model based on One-class SVM to detect network The framework of one-class SVM based model is
intrusions. One-class SVM can effectively overcome the illustrated in Figure 1. The details of the two modules are
difficulties that the common two-class SVM and multi-class presented in the following sections.
SVM encounter.

103
B. Feature extracting module support vectors.
Almost no intrusion detection model can distinguish If w and ρ solve this problem, we have the following
between intrusive connections and normal connections directly
decision function:
from original packets. They must be inputted with formatted
data. Feature extracting is to obtain useful information from f ( x ) = sgn( w ⋅ φ ( x ) − ρ ). (4)
raw data and then format it, so that it can be interpreted by the
detection module. There is no permanent standard to extract That is, if w ⋅ φ ( x ) − ρ ≥ 0 , x is declared as a normal
features. It may be better to extract features based on the actual event, otherwise, it is declared as intrusive.
network environment to find whether some attacks are hidden
in connections. Extracting proper features helps the detection To solve the quadratic programming (4), we introduce
module to make more accurate predictions. In terms of network multipliers α i , βi ≥ 0 , and get the Lagrangian:
intrusions, some frequently-used features need paying attention
to, such as the length (number of the seconds) of the L ( w, ξ , ρ , α , β )
connection, the type of the protocol, e.g. tcp, udp, etc., the 1 1 l (5)
number of data bytes transferred, the number of “root” accesses = || w ||2 + ¦ ξ i − ρ
2 vl i =1
and so forth. In our one-class SVM based detection model, the
l l
feature extracting module takes the raw data as inputs, and then − ¦ α i ( w ⋅ φ ( xi ) − ρ + ξ i ) − ¦ β i ξ i .
extracts expected features to form the formatted data. i =1 i =1
Moreover, the feature extracting module is charged with
dividing the formatted data into two divisions, the training data The following Karush-Kuhn-Tucker (KKT) conditions
and the testing data. This process is fairly simple. The normal are necessary and sufficient conditions for the quadratic
records comprise the training data and the rest (intrusive) programming problem (3).
records comprise the testing data. This relates to the detection ∂L
mechanism of one-class SVM (detailed later). = 0,
∂w
C. One-class SVM module ∂L
Here, we adopt the one-class SVM proposed by = 0,
∂ρ
Scholkopf [10]. First, consider the training dataset:
∂L (6)
D = {( xi , yi ) | xi ∈ \ n , yi = +1}li =1 , (1) = 0, i = 1, 2, ... , l.
∂ξi
where xi is the feature vector with dimension n , yi = +1 αi ≥ 0 , i = 1, 2, ... , l.
means all the training patterns are normal observations, and βi ≥ 0 , i = 1, 2, ... , l.
l is the number of training patterns. According to the KKT conditions, set the derivatives
The algorithm basically separates all the data points equal to zero with respect to the primal variables w, ξi , ρ ,
from the origin. Suppose the hyperplane has the form: yielding:
w ⋅ φ ( x ) − ρ = 0, (2) l
w = ¦ α iφ ( xi ),
ρ i =1
then the distance from the hyperplane to the origin is .
|| w || αi =
1 1
− βi ⇔ 0 ≤ αi ≤ ,
(7)
Maximizing the distance results to solving the following vl vl
quadratic programming problem: l

1 1 l
¦α i =1
i = 1.
min || w ||2 + ¦ ξi − ρ
w ,ξ , ρ 2 vl i =1 In (7), the xi corresponding to α i = 0 is an inner point
s.t. w ⋅ φ ( xi ) ≥ ρ − ξi , (3)
that is irrelevant to solving w ; the xi corresponding to
ξi ≥ 0.
0 < α i < 1/ (vl ) is called a support vector, which is crucial
Here, φ ( xi ) is the feature mapping function that maps to solve w ; xi corresponding to α i = 1/ (vl ) is an outlier.
xi from its input space to a feature space, ξi is the slack
Now, we can rewrite the decision function (4) using the
variable for outlier xi that allows it to lie on the other side kernel function:
of the decision boundary (hyperplane), and v ∈ (0,1] is the l
regularization parameter that is an upper bound on the f ( x ) = sgn(¦ α i k ( xi , x ) − ρ ). (8)
fraction of outliers and a lower bound on the fraction of i =1

104
In (8), k ( xi , x ) is the kernel function with the steal information or find vulnerabilities, e.g. port scanning. (3)
R2L: remote to local, unauthorized access from a remote
following equivalent form:
machine, e.g. guessing password. (4) U2R: user to root, using
k ( xi , x ) = φ ( xi )φ ( x ). (9) system’s vulnerabilities to get super user (root) privileges for
some purpose, e.g. various “buffer overflow” attacks. Attacks
For the one-class SVM used in our detection model, we in each category are divided into many types, resulting in a
use the Gaussian kernel: total of 24 training attack types. It is important to note that the
testing data contains some specific types not in the training data
k ( xi , x ) = e −γ || xi − x|| .
2
(10) and does not have the same probability distribution as the
training data. This makes the intrusion detection task more
Substituting (7) into (5), we can obtain the dual problem: realistic.

1 l In experiments, we used stratified random sampling [29] to


min
α
¦ α iα j k ( xi , x j )
2 i , j =1
reduce the size of dataset. For one-class SVM used in our
intrusion detection model, the training data must contain only
1 (11) normal patterns and does not contain any attacks. So we
s.t. 0 ≤ α i ≤ , selected a random sample of normal records in the raw training
vl data. The sampling proportion is about 3%. To test the model’s
l

¦α
ability to detect different kinds of attacks, we randomly
i = 1.
i =1
selected different types of records in the raw testing data. The
sampling proportion is about 1%. Some types of attacks such as
The answer to the dual problem (11) is also the answer R2L and U2R were totally selected due to their low proportion
to the primal quadratic programming problem (3). in KDDCUP99 dataset. Finally, 32426 normal connection
Furthermore, solving the dual problem is much easier and records in the raw training data and 31415 connection records
more feasible. We use the SMO (Sequential Minimal in the raw testing data were randomly selected. TABLE 1
Optimization) algorithm [26] to solve the dual problem. shows the details about different categories of records. “Other”
indicates the new types of attacks not present in the four main
We know, any support vector xi satisfies the equality: categories.

w ⋅ φ ( xi ) − ρ = 0. (12) TABLE 1. NUMBER AND DISTRIBUTION OF TRAINING AND TESTING DATA


Category Training dataset Testing dataset
So we can use the support vector xi , i.e. the Normal 32426 100% 6060 19.29%
DOS / / 22429 71.40%
corresponding αi that satisfies 0 < α i < 1/ (vl ) , to recover Probe / / 315 1.00%
ρ by the following formula: R2L / / 622 1.98%
U2R / / 39 0.12%
l
ρ = w ⋅ φ ( xi ) = ¦ α j k ( x j , xi ). (13) Other / / 1950 6.21%
j =1 Total 32426 100% 31415 100%

IV. EXPERIMENTS AND DISCUSSIONS B. Evaluation Criteria


To evaluate the performance of our one-class SVM based In order to evaluate the performance of IDS, some accepted
intrusion detection model, we conducted a series of measurements are proposed. If an intrusive event is declared
experiments on KDDCUP99 [27] dataset. intrusive correctly by IDS, it is a true positive. But if an
intrusion is regarded as a normal behavior mistakenly, it is a
A. Data Preparation false negative. By contrary, if a normal event is declared
normal correctly, it is a true negative, but if it is declared
In 1998, DARPA Intrusion Detection Evaluation Program intrusive mistakenly, it is a false positive. We use TP, FN, TN
was prepared and managed by MIT Lincoln Labs. A standard and FP to represent the number of true positives, false
dataset [28] was provided. The KDDCUP99 dataset used in our negatives, true negatives and false positives, respectively.
experiments is a version of this dataset. Usually, we use the detection rate to evaluate the IDS’ ability
The raw training data contains about five million TCP to detect real attacks. For some category of attacks, the
connection records from seven weeks of network traffic. detection rate is the fraction of detected attacks accounting for
Similarly, the two weeks of testing data yields around three the total ones. In addition to the detection rate, another three
million records. Each connection record has 41 derived features criteria are also widely used for performance evaluation,
that help in distinguishing normal connections from attacks, especially for performance comparison. They are precision,
and is labeled as either normal, or as an attack, with exactly one recall and F-value. Precision is the fraction of true positives in
specific attack type. Attacks fall into four main categories: (1) total determined positives (i.e. the sum of true positives and
DOS: denial-of-service, an attempt to make a machine or false positives). Recall has the same formula as the detection
network resource unavailable to its legitimate users, e.g. syn rate. F-value considers both the precision and the recall to
flood. (2) Probe: surveillance and other probing operations to

105
compute the evaluation value. The precision, recall and F-value shown in TABLE 2 and Figure 2, and are produced in this
are defined as follows. way—first, any attack that can be detected by one-class SVM
TP is declared as abnormal without any distinction. Then we
Precision = (14) compute the detection rate for different category of attacks
TP + FP according to the labels in the testing data. We can see that for
TP DOS attacks, the three models get perfect results (all above
Recall = (15) 99%). For Probe attacks, one-class SVM can reach the top
TP + FN detection rate 100%, while the detection rates of PNN and C-
SVM are relatively lower, respectively 98.73% and 86.98%.
2* Recall* Precision
F −value = (16) We should note that for R2L, U2R and ̌Other̍ categories of
Recall + Precision attacks, the results of all the three models are not very
satisfactory. We believe one of the main reasons is that the
C. Results and Discussions number of attacks in these three categories is relatively small
In this section, we compare our one-class SVM based (see Table 1), so the test results have some limitations. Another
model with other two well-knowns, probabilistic neural reason may be that the attacks are too covert to be detected by
network (PNN) [30] and C-SVM (proposed by Cortes and the models. But even so, the detection results of one-class
Vapnik in [9]), given that they both adopt the radial basis SVM are considerably better than two others’. Furthermore, for
function (Gaussian function or Gaussian kernel) as the one- PNN and C-SVM, the “Other” category of attacks are new
class SVM does and are often used to detect intrusions due to attacks not present in their training data, so it is especially
their good classification performance. PNN used in our difficult for them to detect such attacks. But for one-class
experiments is taken from the MATLAB R2013b toolbox and SVM, the new attacks receive the same treatment as with other
C-SVM from the software LIBSVM [31]. Because the training categories of attacks, without any difference.
data used by PNN and C-SVM must contain both normal and
abnormal records, we conducted a stratified random sampling Next, we use three other criteria, precision, recall and F-
for the raw training data in KDDCUP99 with the proportion value to conduct performance comparison. The results are
around 1%. The final training data consists of 49567 records, shown in Table 3 and Figure 3. As illustrated by Figure 3, one-
including 9728 Normals (19.63%), 39167 DOSs (79.02%), 412 class SVM produces a slightly lower precision than PNN and
Probes (0.83%), 208 R2Ls (0.42%), and 52 U2Rs (0.10%). The C-SVM. But the precisions of all the three models are very
other three models use the same testing data as described in high (above 99%). Apparently, the recall and F-value of one-
TABLE 1. class SVM are higher than others’.
TABLE 3. PRECISION, RECALL AND F-VALUE OF DIFFERENT MODELS
PNN C-SVM One-class SVM
Precision 0.9988 0.9957 0.9903
Recall 0.8916 0.9041 0.9161
F-value 0.9422 0.9477 0.9518

Figure 2. Detection rate comparison of different models

TABLE 2. DETECTION RATES OF DIFFERENT MODELS


PNN C-SVM One-class SVM
DOS 0.9979 0.9958 0.9950
Probe 0.9873 0.8698 1.0000
R2L 0.0804 0.0322 0.2685
U2R 0.1284 0.4872 0.6923
Figure 3. Performance comparison of different models
Other 0.0687 0.1421 0.2067
In experiments, the parameter γ (gamma) in radial basis V. CONCLUSION
function was set to 0.5, the cost parameter c in C-SVM was set We propose a novel anomaly detection model based on
to 1 and the parameter v (nu) in one-class SVM was set to 0.05. One-class SVM to detect network intrusions. One-class SVM is
First, we compare and discuss the detection rates of these three a one-versus-rest classifier, which is very suitable for anomaly
models for different categories of attacks. The results are detection. Although the commonly used two-class SVM
algorithms have been applied in intrusion detection, they are

106
facing the difficulties of constructing the training dataset. That [11] Taylor, Carol, and Jim Alves-Foss. Low cost network intrusion
is because in many true application scenarios, it is easy to detection. (2000).
obtain normal connection records, but difficult to obtain attack [12] Barbara, Daniel, Ningning Wu, and Sushil Jajodia. Detecting Novel
Network Intrusions Using Bayes Estimators. SDM. 2001.
records, or the number of attack records is very limited.
[13] Shyu, Mei-Ling, et al. A novel anomaly detection scheme based on
Whereas to a great extent, the distribution of training records principal component classifier. MIAMI UNIV CORAL GABLES FL
affects the detection results of the two-class SVM. Hence, we DEPT OF ELECTRICAL AND COMPUTER ENGINEERING, 2003.
propose to use one-class SVM, which adopts only the normal [14] Qin, Min, and Kai Hwang. Frequent episode rules for intrusive anomaly
network connection records as the training data, to conduct the detection with internet datamining. USENIX Security Symposium. 2004.
anomaly detection. The experimental results on KDDCUP99 [15] Denning, Dorothy E. An intrusion-detection model. Software
dataset show that compared to PNN and C-SVM, our one-class Engineering, IEEE Transactions on 2 (1987): 222-232.
SVM achieves higher detection rates for different categories of [16] Wang, Gang, et al. A new approach to intrusion detection using
attacks and has an average better performance in terms of Artificial Neural Networks and fuzzy clustering. Expert Systems with
precision, recall and F-value. The deficiency lies in that both Applications 37.9 (2010): 6225-6232.
our one-class SVM based and other two models show relatively [17] Sinclair, Chris, Lyn Pierce, and Sara Matzner. An application of
machine learning to network intrusion detection. Computer Security
low detection rates for low-frequent attacks, such as R2L and Applications Conference, 1999.(ACSAC'99) Proceedings. 15th Annual.
U2R. Affecting the accuracy of results, the insufficient number IEEE, 1999.
of data is partially to blame. But the detection model could also [18] Tsai, Chih-Fong, et al. Intrusion detection by machine learning: A
be enhanced. We leave this as future work. review. Expert Systems with Applications 36.10 (2009): 11994-12000.
[19] Ryan, Jake, Meng-Jang Lin, and Risto Miikkulainen. "Intrusion
ACKNOWLEDGMENT detection with neural networks." Advances in neural information
processing systems (1998): 943-949.
We would like to thank the anonymous reviewers for their [20] Kim, Dong Seong, and Jong Sou Park. "Network-based intrusion
comments and suggestions. Thank our advisors for their detection with support vector machines." Information Networking.
valuable guidance. Thank all the authors in references for their Springer Berlin Heidelberg, 2003.
work to help us carry out research. [21] Sung A H, Mukkamala S. Identifying important features for intrusion
detection using support vector machines and neural
networks[C]//Applications and the Internet, 2003. Proceedings. 2003
REFERENCES Symposium on. IEEE, 2003: 209-216.
[1] Symantec Enterprise. Internet Security Threat Report 2014. [22] Mukkamala, Srinivas, Guadalupe Janoski, and Andrew Sung. "Intrusion
http://www.symantec.com/content/en/us/enterprise/other_resources/b- detection using neural networks and support vector machines." Neural
istr_main_report_v19_21291018.en-us.pdf. (accessed 15th, Apr, 2015) Networks, 2002. IJCNN'02. Proceedings of the 2002 International Joint
Conference on. Vol. 2. IEEE, 2002
[2] Cenzic. Application Vulnerability Trends Report 2014.
http://www.cenzic.com/downloads/Cenzic_Vulnerability_Report_2014. [23] Ambwani, Tarun. "Multi class support vector machine implementation
pdf. (accessed 15th, Apr, 2015) to intrusion detection." Neural Networks, 2003. Proceedings of the
International Joint Conference on. Vol. 3. IEEE, 2003.
[3] Anderson, James P. Computer security threat monitoring and
surveillance. Vol. 17. Technical report, James P. Anderson Company, [24] Khan, Latifur, Mamoun Awad, and Bhavani Thuraisingham. "A new
Fort Washington, Pennsylvania, 1980. intrusion detection system using support vector machines and
hierarchical clustering." The VLDB Journal—The International Journal
[4] Axelsson, Stefan. Intrusion detection systems: A survey and taxonomy.
on Very Large Data Bases 16.4 (2007): 507-521.
Vol. 99. Technical report, 2000.
[25] Horng, Shi-Jinn, et al. "A novel intrusion detection system based on
[5] Kruegel, Christopher, and Thomas Toth. Using decision trees to improve
hierarchical clustering and support vector machines." Expert systems
signature-based intrusion detection. Recent Advances in Intrusion
with Applications 38.1 (2011): 306-313.
Detection. Springer Berlin Heidelberg, 2003.
[26] Platt, John. "Sequential minimal optimization: A fast algorithm for
[6] Patcha, Animesh, and Jung-Min Park. "An overview of anomaly
training support vector machines." (1998).
detection techniques: Existing solutions and latest technological trends."
Computer Networks 51.12 (2007): 3448-3470. [27] UCI KDD Archive. KDDCUP99 dataset.
http://kdd.ics.uci.edu/databases/kddcup99/. (accessed 15th, Apr, 2015)
[7] Li, Yuping, Weidong Li, and Guoqiang Wu. "An intrusion detection
approach using SVM and multiple kernel method." IJACT: International [28] MIT Lincoln Laboratory. DARPA Intrusion Detection Data Sets.
Journal of Advancements in Computing Technology 4.1 (2012): 463- http://www.ll.mit.edu/mission/communications/cyber/CSTcorpora/ideva
469. l/data/index.html. (accessed 15th, Apr, 2015)
[8] Li, Yinhui, et al. "An efficient intrusion detection system based on [29] Georgiou, V. L., et al. Optimizing the performance of probabilistic
support vector machines and gradually feature removal method." Expert neural networks in a bioinformatics task. Proceedings of the EUNITE
Systems with Applications 39.1 (2012): 424-430. 2004 Conference. 2004.
[9] Corinna Cortes and Vladimir Vapnik. "Support-vector networks." [30] Specht, Donald F. "Probabilistic neural networks." Neural networks 3.1
Machine learning 20.3 (1995): 273-297. (1990): 109-118.
[10] Schölkopf, Bernhard, et al. "Estimating the support of a high- [31] Chih-Chung Chang and Chih-Jen Lin, LIBSVM : a library for support
dimensional distribution." Neural computation 13.7 (2001): 1443-1471. vector machines. ACM Transactions on Intelligent Systems and
Technology, 2:27:1--27:27, 2011. Software available at
http://www.csie.ntu.edu.tw/~cjlin/libsvm.

107

You might also like