Anda di halaman 1dari 9

JOURNAL OF NETWORKS, VOL. 5, NO.

4, APRIL 2010 475

A Cooperative Network Intrusion Detection Based


on Fuzzy SVMs
Shaohua Teng, Hongle Du, Naiqi Wu, Wei Zhang, Jiangyi Su
Guangdong University of technology
Guangzhou, Guangdong, China, 510006
shteng@gdut.edu.cn, dhl5597@163.com, NQWu@gdut.edu.cn

AbstractAs the network information includes a large alarm wherever the deviation between a given
number of noise data, in order to reduce or eliminate observation at an instant and the normal behavior exceeds
the noise impact on constructing the hyperplane of a predefined threshold. Another possibility of anomaly
SVM, this paper firstly preprocesses the data. Then detection is to model the abnormal behavior of the
the fuzzy membership function is introduced into system and to raise an alarm when the difference between
SVM. The fuzzy membership function acquires the observed behavior and expected one falls below a
different values for each input data according to given limit. Anomaly detection is used to detect unknown
different effects on the classification result. Because attacks. Both methods need to establish profiles of user
different network protocol has different attributes, behaviors. Of course they can also be used to classify the
that must affect the detection effect. This paper users behavior. There are machine learning methods for
proposes cooperative network intrusion detection pattern recognition such as neural networks, Bayesian
Based on Fuzzy SVMs. Three types of detecting agents theory and genetic algorithm and so on. Support vector
are generated according to TCP, UDP and ICMP machine (SVM) is also a new machine learning method
protocol. Finally, simulate with KDD CUP 1999 data and is widely applied to the field of pattern recognition. It
set, and the experiment results show there are a better will be widely applied in intrusion detection
detection effect with cooperative network intrusion Support Vector Machine is a popular topic based on
detection based on multi fuzzy SVMs. statistical machine learning [1]. In a nutshell, a SVM is
an algorithm that works as follows [2]. It uses a nonlinear
Index TermsFuzzy Support Vector Machine; Intrusion mapping to transform the original training data into a
Detection; Membership Function ; Incremental Learning; higher dimension. Within this new dimension, it searches
Cooperative; Network for the linear optimal separating hyperplane (that is, a
decision boundary separating the tuples of one class
from another). With an appropriate nonlinear mapping to
I. INTRODUCTION a sufficiently high dimension, data from two classes can
Intrusion detection is the second Line of defense in always be separated by a hyperplane. The SVM finds this
network security. Intrusion Detection System can be hyperplane using support vector (training tuples) and
divided into three categories according to protecting margins (defined by the support vectors). Especially in
objects: network intrusion detection system, host high-dimensional data space, the effective overcome of
intrusion detection system and hybrid intrusion detection the dimension disaster and excessive learning problems
system. Network-based intrusion detection system is used are very important. SVM has widely been applied in
to protect the local network or whole network segment, pattern recognition fields [3, 4]. Network connection
monitor the network packets and find out attacks from includes much information of user behavior. The
data packets. Then these attack actions are dealt with traditional SVM-based intrusion detection methods are
correspondingly, such as cut the connection, send out an rarely taken into considering the differences among
alarm signal and so on. Host-based intrusion detection different network protocols. They found SVM by
system is used to protect the critical computer. It adopting unified data formats. That takes much time and
recognizes penetration behavior by fetching and leads to low efficient. In addition, there are also two
analyzing internal system auditable events, system logs, problems which need to been solved: processing
system status, and logs of application program, and then capability of large-scale training set and eliminating the
makes the appropriate response. Depending on the type of impact of noise data. On the one hand with increasing the
analysis carried out, intrusion detection systems are number of training samples, training time and storage
classified as either signature-based or anomaly-based. space will increase dramatically (the time complexity of
Signature-based schemes (also denoted as misuse-based) SVM is O(k3)). On the other hand, the ultimate decision-
seek defined patterns, or signatures, within the analyzed making function depends on the small number of support
data. For this purpose, a signature database corresponding vectors of training samples. Therefore SVM is very
to known attacks is specified a priori. On the other hand, sensitive to outlier and noise sample. Accordingly, this
anomaly detection needs to establish the user's normal paper presents a cooperative network intrusion detection
behavior patterns in the protected system, and generate an based on fuzzy SVMs. According to different network

2010 ACADEMY PUBLISHER


doi:10.4304/jnw.5.4.475-483
476 JOURNAL OF NETWORKS, VOL. 5, NO. 4, APRIL 2010

protocols, the text shows to build a different network samples different power weights in the object functions,
behavior detection classifier. The experimental results samples play different roles during training. That
show that this method can reduce evidently the training increases FSVM efficient.
time and storage space and improve the classification Incremental learning arises out of solving two kinds of
accuracy. problem. One is the training of large-scale data sets, lack
The rest of this paper is organized as follows. Section 2 of memory and too much time. The other is unable to
presents related work on intrusion detection based on obtain a complete data set. We have to use learning on
FSVM. Section 3 gives a cooperative network intrusion line. Data samples are accumulated during continuous
detection model of FSVMs. Section 4 describes v-FSVM, applications to improve the learning accuracy. The key of
fuzzy membership functions and their calculation. incremental learning is to retain the information of
Section 5 presents a detailed process of building a original samples and how to cope with the increased
detection agent about TCP attacks, and proposes a new samples. Syed [13], who is the first man, proposed an
incremental learning algorithm of support vector machine. incremental algorithm of SVM. The idea is to acquire the
Section 6 presents some experiment results done in KDD support vectors by training initial sample set. Both the
CUP 99 data set. It verifies that our method is efficient. new data sets and the previous support vectors form new
Section 7 draws conclusions and outlines future work. samples, and train them to produce new support vectors.
In order to reduce training time, paper [13-14] eliminates
II. RELATED WORK samples outside support vectors. That leads to lower
Intrusion detection can be seen as a classification SVM accuracy due to lack of efficient learning. Paper [15]
problem. According to the network information, network analyzes the relationship between KKT conditions and
behavior can be divided into to normal behavior and sample distribution and presents an increment learning
abnormal behavior. Therefore intrusion detection algorithm. Author thinks that the initial sample set and
problem transforms into pattern recognition problem. the new sample set have the same impact of the final
Paper [4] presents a SVM-based intrusion detection hyperplane. Paper [16] summarizes the current study and
model, and discussed the work process of the model. application of incremental support vector machine and
Paper [5] builds a lightweight intrusion detection system the gives the generalized KKT conditions. The sample
by using a feature selection algorithm, which can reduce points near the hyperplane may be new support vectors in
training time and storage space. Experiments show that the new case due to increment samples. Paper [17]
the method improves the detection efficiency of the proposed redundant incremental learning. It means that
system. Paper [6] combines the C-SVM-based some samples near the hyperplane are added to
supervision algorithm and unsupervised algorithm of participate in new training according to predefined rules.
One-Class SVM, and defines the RBF kernel function That can reduce the loss of information and improve
bases based on HVDM distance. The method is used to training efficiency. In order to solve the problem of
deal with heterogeneous data sources. Paper [7] gives a incremental learning and large-scale learning, paper [18]
SVM based on the fuzzy C-means algorithm. Fuzzy proposed a rapid incremental SVM learning algorithm
membership functions computed iteratively forms a based on active set iteration. Iterations contain the results
member matrix, which is used as power weight of of previous work of the active set.
inputting samples. This method improves the effect of
detecting intrusion. III. A COOPERATIVE INTRUSION DETECTION MODEL
BASED ON FSVMS
To reduce the impact of noise dataLin and et al [8-11]
apply fuzzy technology to support vector machine, called
FSVM. Categories of these samples and measures
A. The Architecture of Cooperative Intrusion Detection
belonging different classes decide the effects on the Network intrusion detection is essentially a
objective function. That assures eliminating or reducing classification problem. Network data is classified as
the impact of the noise and outlier samples for the normal and anomaly data. The network intrusion
objective function. Paper [9] presents a new kernel detection can be transformed into a network
function which is constructed through fuzzy cluster behavior classifier. The recognition of the network
method. That improves the classification effect by fuzzy behavior is related to network protocols. The different
measures of the samples. Paper [10] gives two network protocols have different formats of the network
membership functions to each sample point. These packets, which lead to different network connections,
functions present respectively the degree of positive and such as connection oriented TCP protocol and
negative instances. Based on the multi-SVM classifier unconnection oriented UDP protocol. Therefore this
that is proposed by Weston and Watkins [12], Lin and Li paper use different classifiers for different network
[8, 11]
give an effective method that can eliminate the protocol data. The detection model is shown in flag1,
impact from noise and outlier points for training dataset which includes data collector, data preprocessor, detection
by applying fuzzy member function. Paper [8-11] agent and decision response and saving unit.
describes how to optimize FSVM by adopting different
membership functions. In this paper, fuzzy membership
functions are used in V-SUM, called v-FSVM. By setting

2010 ACADEMY PUBLISHER


JOURNAL OF NETWORKS, VOL. 5, NO. 4, APRIL 2010 477

Figure 1. The architecture of intrusion detection system based on multi-FSVM

B. Components B. Introducing membership into v-SVM


A cooperative network intrusion detection model based Suppose a training data set labeled by class with
on fuzzy SVMs includes data collector, data preprocessor, associated fuzzy
detecting agent and response unit, shown as figure 1. membership
Data collector is used to collect network data. Data s = {( x1, y1 , u( x1 )),( x2 , y2 , u( x2 )),L,( xl , yl , u( xl ))} . Each
preprocessor is used for filtering, cleaning, integrating,
training point xi R is given a label yi {1, 1} and a
n
preprocessing data, attribute selection and format
fuzzy membership 0 < u ( xi ) < 1 . Since the fuzzy
conversion. After data preprocessor completes data
preprocessing, data is sent to corresponding detecting
membership ui is the attitude of the corresponding point xi
agent according to TCP, UDP or ICMP protocol. Because
toward one class and the parameter i is a measure of
SVM just accesses numerical data, non-numerical data
must be transformed into numerical type. SVM requires error in SVM, the term u ( x i ) i is a measure of error with
that all of data have the same dimension. So we extract different weighting. We transform the input data xi into a
the effective network connection information, and the high-dimensional approximately linearly separable
original network data has to transform into digital vector. feature space using a nonlinear mapping Z =(x). Then
Detection agent analyzes the data submitted by data the optimal hyperplane problem is regarded as the
preprocessor, and decides intrusion or not. Response unit following solution:
l
1 1
makes corresponding decision according to the result of m in
2
< w, w > +
l
u i ( i v )
detecting agent. i =1

s.t. yi (< w, xi > +b) i 1


IV. V-FSVM
i 0 0 i = 1, 2,L , l
A. V-SVM From (2), we can see that a smaller ui reduces the
In the C-SVM, there are two contradictory objectives: effect of the parameter i in problem (2) such that the
maximize margin and minimize the training mistakes. corresponding point is treated as less important. Those
The constant C plays the role to reconcile these two also reduce the importance of the corresponding samples
objectives, however it is difficult to select C. To solve xi. In order to solve the problem (2) with constrained
this problem, many scholars proposed v-SVM algorithm optimization problem, define as follows function for the
in 2000 [19], which by introducing parameter v replace Lagrange:
l l
1 T 1
the classical C-SVM algorithm [1]. Parameter V has the L (w, b, , , ) =
2
w w +
l
u i ( i v ) i i
practical significance: v is up-border of the proportion of i =1 i =1
l
classification error samples and low-border of support i [ yi (< w , xi > + b ) + i ]
vector samples in the training data sets. Although v-SVM i =1

and C-SVM may get the same classification hyperplane s.t.


i 0 i 0 i = 1, 2, L, l 0 2
by choosing the appropriate parameters, the parameter v Solving (3) the minimum, respectively w, b, i, for
in v-SVM has a specific and intuitive meaning. It is easier the partial derivative, and let its equivalent be 0
to select it than C in the C-SVM. This will avoid the L l

shortcomings of the choice of parameter C relying mainly w


= w i i xi = 0
i=1 3
on experience in the C-SVM.

2010 ACADEMY PUBLISHER


478 JOURNAL OF NETWORKS, VOL. 5, NO. 4, APRIL 2010

L l
r+ = m ax || x i x + ||
= i yi = 0
b i=1 4 i

And
L 1 r = m ax || x i x ||
= ui i = 0 i
i
i
l 5 So we define the membership function:
|| x i x + ||
u ( xi ) = 1 + xi is normal class
L v l l


= ui + i = 0 r+
l i =1 i =1 6 or
|| x x ||
u ( xi) = 1 i
+ xi
Apply (4)-(7) into (3). Transform the Optimal r is abnormal class.
classification of the problem based on the fuzzy Where > 0is small number predefined to
membership Support Vector Machine into its dual form:
l l l avoid
u ( x i ) = 0 . In this paper, we suppose the number of
1
m ax W (a ) =
a

i=1
i
2

i=1 j=1
i j y i y jK ( xi, x j )
samples in one cluster as the fuzzy membership.
l

i yi = 0
s.t. i =1 V. DETECTION AGENT BASED ON FSVM
1
0 i ui i = 1, 2 , L , l
l (7) A. The architecture of detecting agent
According to the Network protocol type, Network
vui
i packets are divided into three types: TCP packets, UDP
l
Where K(xi, xj) is kernel function, and let K(xi, xj) = packets and ICMP packets. And the Network Data
<(xi), (xj)>.. Flow is looked as three types: TCP Data Flow, UDP Data
After solving this quadratic programming problem, we Flow and ICMP Data Flow. Therefore, three types of
obtain the following classifier: detection agents should be constructed to meet the TCP,
UDP and ICMP detection.
f ( x ) = sgn(
xi sv
a i* y i K ( x , x i ) + b * )
Every detection agent has three processes: construction,
(8)
l adaptive and detection. We give the TCP detection agent
w* = a i * y i ( x i ) as follows. The UDP and ICMP detection agent are
Where i =1 similar with TCP detection agent. The architecture of
1
b * = ( < w * , ( x i ) > + < w * . ( x j ) > ) TCP detection agent is shown in figure 2.
2 . From figure 2, there are tow stages to construct every
C. Membership terms agent in the model: training and predicting. The data after
preprocessing is divided into training data set and testing
In FSVM, many scholars do a deal of research work
data set. The training stage trains support vector machine
for the fuzzy membership. The literature [20] defines
based on the samples of known types applying (7).
membership function to use samples to the class centre
According to (2), we can obtain support vectors and the
distance; because there is the lack of distance
corresponding parameters. Predicting stage is a process. It
membership function because of no taking account to the
is implemented by using support vector machines to
number of samples in one class, the literature [21] gives
classify the network data processed with above method.
the density of membership functions; The literature [22]
According to the discriminant function (8), we can get the
uses the theory of rough set to define membership
computing results of network behavior. The results are
functions. The method can avoid the problem that radius
submitted to the decision-making system. The appropriate
are difficult to compute.
decision is made. The TCP detection agent based on
The principles of determining the size of membership
FSVM is generated after FSVM is tested with testing data
base on the importance of the samples in the category, or
set.
the samples contribution to the category. Class samples to
The self-adaptive module learns the new knowledge
the centre distance are one of the measures of the
through incremental learning to improve the detection
contribution for sample category. To determine the
ability of detection agents; the detecting module gets the
membership based on the distance, the membership of the
network data and predicts results with FSVM, and then
samples is known as the distance from the sample to the
sends them to deciding and responding module to make
centre in feature space
decisions.
Given x + and x are the centers of tow classes, and

we can get radius:

2010 ACADEMY PUBLISHER


JOURNAL OF NETWORKS, VOL. 5, NO. 4, APRIL 2010 479

Detecting intrusion

Data collector

Constructing the model Response & decision

Data preprocessing
Data preprocessing
TCP
Trainning
dataset
FSVM

TCP TCP FSVM


Response unit
Dataset agent
TCP FSVM
Test
model
dataset
Data preprocessing

New New case


dataset
Self-adaptor

Figure 2. The architecture of TCP detection agent

realization process of every detection agent. Data


B Data Preprocess of TCP Trainer
preprocessing is shown in figure 3. After preprocessing,
For every classifier, data preprocessing is similar in we can get data to meet SVM training. And then go to the
training stage and predicting stage. Taking the classifier next process.
of TCP data flow as an example, we discuss the

Figure 3. Data preprocessing

Figure 3 is described as follow: 18 attributes for ICMP detection agent from 41 attributes.
Data cleaning is a process to deal with the problem of Experiments verify that the results are the same as before
data inconsistency through filling in missing values, discretization.
smoothing noise data, recognizing or deleting the data Cluster chooses the simple unsupervised clustering
including very small information. algorithm (UC).
Attribute selection reduce the size of analyzed data set
C. The SVM based on clustering
through deleting the unrelated attributes or redundant
attributes. For example, the formats of TCP, UDP, and Based on above analyzing, this paper presents a new
ICMP are different, so detection agent of UDP does not incremental learning algorithm combined SVM with
need the specific attributes of TCP. And it ensures that clustering algorithm. In this algorithm, first, we deal with
the experiment results with smaller data set are similar or training set using clustering algorithm (Class label is
same with the original data set. looked as an attribute of data set. And the samples that
Data integration combines the data that comes from belong to the same cluster have same class label if
different data source. Data is stored respectively radius r > 1 ). Then we can get clusters O(ni , oi , yi )
according to TCP, UDP and ICMP protocol.
(Where ni is the number of sample point; oi is the center
Data transformation transforms the non-numerical
values into numerical attributes in order to meet FSVM of the cluster; yi is class label). Next, we construct new
training.
Data discretization is used to generalize data. It training data set using centers of clusters and ni are fuzzy
discretizes the attributes to reduce the amount of analyzed membership function. Then we train new training data set
data through with FSVM and obtain support vectors. There are two
After preprocessing, we select 32 attributes for TCP strategies to deal with new adding data set: one is to add
detection agent, 21 attributes for UDP detection agent and new adding samples into the support vector set that is get

2010 ACADEMY PUBLISHER


480 JOURNAL OF NETWORKS, VOL. 5, NO. 4, APRIL 2010

in the first step using clustering algorithm; another is only classification. Then we can get the new training set
to add samples that contrary to KKT condition using UC T = {( n1 , o1 ', y1 ), ( n2 , o2 ', y2 ), L , (n p , o p ', y p )}
.Where ni is
algorithm and thrown away the samples that are satisfied
the fuzzy membership function. Finally, retrain the new
training set T using FSVM and get classifier and the
with KKT condition. At last we can get new training set
and train it using FSVM again. In this paper, we will
compare two treatment methods with experimental results support vector set SVs;
and give the corresponding analysis. Step 3. There are two methods to deal with new adding
1) UC algorithm set B. The first method is that we look the set SVs as the
Literature [23] presents a simple UC algorithm. clustering center set and deal with set B using clustering
Compared with tradition K-means algorithm, the algorithm of section 2.2. Then we can get new clustering
algorithm doesnt need to pre-specify the number of center set O. Second method, according to KKT
classification. There is a high speed of clustering, so we condition, discard the samples that meet the KKT
deal with the training data set using the UC algorithm in condition and deal with the samples that contrary to KKT
this paper. Given the training set using the first method;
T = {x1 , x2 , L , xl }, xi R n +1 Step 4Do with O using the method of step 2; then we
. Here, we look the label of can get the new classifier ' and new the support vector
the SVM training set as a dimension, so the number of
dimensions is n+1. C_number is the number of cluster. set SVs ' .
Then algorithm can be described as follows: The process of training is shown in figure 4.
Step 1. Read one record xi from training set. If
C_number = 0, Create a new cluster center O1 and set O1
= xi; otherwise go to 2;
Step 2. Compute distance di between sample xi and
each ok
di = ( x j1 ok1 )2 + ( x j 2 ok 2 )2 + L+ ( x jm okm )2
Set d m = min ( di ) , where n is number of cluster
i =1,2,L, n

particles and m is index of cluster particle


d m < r , then xi is added into cluster om
Step 3. If
om * nm + x j
om =
and reset om: nm + 1 ; and reset n = n + 1;
k k
Otherwise recreate a cluster and set C_number =
o =x n =1
C_number + 1, C _ number j
, C _ number ;
Step 4. If all samples are dealt with, then stop;
otherwise go to 1.
2) Combining SVM with UC algorithm
In order to keep more classification information of Figure 4. The process of training
original samples and have a high accuracy of
classification, the clustering radius should be smaller. If r D. An Adaptive mechanism
= 0, that is to say, we do not cluster. But in order to keep
more boundary support vectors and improve the speed of Adaptive mechanism for detecting agents mainly
achieves through incremental learning for a new support
training, we should make the radius larger. In this paper,
we set r = 1. Given training sample set with label of vector set. During incremental learning of SVM, support
vectors will change into none-support vectors and none-
categories A = {( xi , yi ), i = 1, 2, L , l} and new adding support vectors will change into support vectors to ensure
data set B = {( xi , yi ), i = 1, 2, L , p} , xi R n , the maximize margin and minimize error rate of classifier.
Based on above analyzing, this paper presents a new
y i {1, 1} . l, p are number of samples. The incremental learning algorithm combining SVM with
Incremental Learning SVM algorithm based on cluster clustering algorithm. In this algorithm, first, we deal with
can be described as follows: training set using clustering algorithm (Class label is
Step 1. Deal with the data set A using the algorithm of looked as an attribute of data set. And the samples that
section 2.2. Then we can get the clustering center set belong to the same cluster have same class label if
O = {(n1 , o1 ), (n2 , o2 ), L , ( n p , o p )} radius r > 1 ). Then we can get clusters O(ni , oi , yi )

Step 2. Reconstruct training set with clustering center (Where ni is the number of sample point; oi is the center
set O. Firstly, we separate the last characteristic of the
of the cluster; yi is class label). Next, we construct new
oi '
vectors Oi and get and Yi. Here Yi is the label of

2010 ACADEMY PUBLISHER


JOURNAL OF NETWORKS, VOL. 5, NO. 4, APRIL 2010 481

training data set using centers of clusters and ni is fuzzy uses a simple substitution of symbols with numerical data
types. The protocol-type, service and flag replaced by
membership function. Then we train new training data set digital attributes. For examples, three kinds of Protocol-
with FSVM and obtain support vectors. There are two type (TCP, UDP and ICMP) are instead with 1, 2, and 3.
strategies to deal with new adding data set: one is to add 71 kinds of service are substituted with 1, 2,, 71. Label
new adding samples into the support vector set that is get of attack instead of 1 or -1, where normal record is 1 and
in the first step using clustering algorithm; another is only abnormal record is -1. Last, normalize the input data set
to add samples that contrary to KKT condition using UC with Libsvm [25].
algorithm and thrown away the samples that are satisfied In order to reduce the training time and ensure
with KKT condition. At last, we can get new training set representation of the chosen data, use the same interval to
and training it using FSVM again. selected data sets. We select four training sets: training
VI. EXPERIMENTS set is that get one record every 15 records from first and
[24] all are 32935 records from 10 percent of the training data
KDD CUP 1999 data are standard data sets for set of KDD CUP 1999; Test set is that get one record
Intrusion Detection, including the training data sets and every 20 records from fifth and all are 15552 records
test data sets. The training data sets include 494,022 from the Correct with label of KDD CUP 1999.
records and test data sets include 311030 records. There
are 24 types of attacks in training data sets and increase B. Experiments Results and Analysis
of new 14 kinds of attacks in the test data sets. They can We select different attributes to experiment for
be divided into four major categories: Probing, Denial of different network protocol. Select 32 attributes from TCP
Service (DoS), User-to-Root (U2R) and Remote-to-Local data set and 21 attributes from UDP data set and 18
(R2L). Each a complete TCP connection is considered as attributes from ICMP data set. The accuracy rates are
a record, including four types of attributes collection: exactly same with the experimental results obtained with
time-based traffic features, host-based traffic features, the 41 attributes. The results are shown in table 1 and
content features, basic features. A total of 41 different table 2.
attributes, of which include 32 consecutive attributes and In table1 and table 2, N-train is the number of records
nine discrete attributes. In all experiments, we use of training data set, N-test is the number of records of
personal computer (P4 3.0 GHZ, 512M RAM), and the testing data set, TN is the number of correctly detected
operating system is Windows XP. and TR is true rate of correctly classified and define as:
A. Selection experiment data sets TR (True rate) = numbers correctly classified / total
numbers of samples
These attributes are the types of values, and others are R-error is the percentage of records that the corrected
types of characters, but SVM can only deal with records are detected as attack. T-time is time of training
numerical vector. Therefore, before training we must FSVM.
make the input data numerical and normal. This study

TABLE I. THE DETECTION RESULTS OF 41 ATTRIBUTES


detector N-train N-test TN Accuracy (%) R-error (%) T-time(s)

TCP FSVM 12674 6000 5149 85.8167 1.8513 217


UDP FSVM 1346 1297 787 60.6785 3.2856 2
ICMP FSVM 18915 8255 8249 99.9273 0 4

TABLE II. THE DETECTION RESULTS OF REDUCED ATTRIBUTES


detector N-rain N-test TN Accuracy (%) R-error (%) T-time(s)

TCP FSVM 12674 6000 5149 85.8167 1.8513 189


UDP FSVM 1346 1297 787 60.6785 3.2856 2
ICMP FSVM 18915 8255 8249 99.9273 0 3

As shown in table 1 and table 2, training set of ICMP ICMP. The accuracy rate is only 60.6785% after UDP
is very larger than others, but training time is shorter and training from the table. In order to increase the number of
there is a higher accuracy rate of prediction. Analyzing records of UDP training set, we select all 21865 records
the training data set of ICMP, we can see that there are of UDP from 10% training set of KDD CUP 1999 data
18819 attack records and only 96 normal records in set and look as the training set to train the classifier. The
training set of ICMP. There is a small computation to accuracy rate of prediction is still only 61.6573% and still
construct the hyperplane for this unbalance training set, bellows the accuracy rate of TCP and ICMP. So the
and so the training time is short. And reducing attributes, reason of low accuracy rate is not less records. Analyzing
the training time is not too large change. The records in the UDP training set, we can find that there are only 77
training set of UDP are significantly less than TCP and attack records in 1346 records (the proportion of 5.7207%)

2010 ACADEMY PUBLISHER


482 JOURNAL OF NETWORKS, VOL. 5, NO. 4, APRIL 2010

and 1498 records in 21865 records (the proportion of is 5149+787+8249=14185 records with cooperative
6.8511% ). But there are 18819 attack records in 18915 detection based on multi FSVMs, and accuracy rate is
records (the proportion of 99.5925%) of ICMP training 91.2101%. The number of detection records is 12841
set and 7554 attack records in 12674 records (the records with single FSVM, and accuracy rate is
proportion of 59.6023%) of TCP. So the accuracy rate of 82.5682%. Training time of detection based on multi
prediction for UDP is not because of training set but also FSVMs is only 194 seconds and training time of single
because of unbalance training set of UDP. And it affects FSVM is 816 seconds. We can see that training time and
to construct the hyperplane of FSVM, and meanwhile it accuracy rate of cooperative detection based on multi
affects detection accuracy rate. FSVMs is better than single FSVM.
Table 3 is the results comparison between single
FSVM and multi FSVM. All number of detection records

TABLE III. THE RESULTS COMPARISON BETWEEN SINGLE FSVM AND MULTI FSVMS.

algorithm TN Accuracy (%) R-error (%) T-time(s)

Single FSVM 12841 82.5682 6.3714 816

Multi FSVM 14185 91.2101 5.1369 194

intrusion detection [J]. Chinese Journal of Software.2008,


VII. CONCLUSION 19(8):2140-2148
In this paper, we propose a cooperative network [8] Lin, C.F. Wang, S.D. Training algorithms for fuzzy
support vector machines with noisy data[C]. IEEE XIII
intrusion detection system based on multi FSVMs. Firstly,
Workshop on Neural Networks for Signal Processing, pp.
the v-FSVM reasoning process is given. And then 517-526 (2003)
construct the different detection agents according to [9] Yanyou Hao, Zhongxian Chi, Deqin Yan, Xun Yu. An
different network protocol and give the process of Improved Fuzzy Support Vector Machine for Credit
adaptive learning. This method divides the network data Rating[C]. IFIP International Federation for Information
flow with network protocol, so it can improve the speed Processing 2007. 2007 495-505
of each detection agent and the accuracy rate of [10] Chiang JH, Hao PY.A new kernel-based fuzzy clustering
prediction. Meanwhile, experiment results prove the approach: Support vector clustering with cell
method. But the accuracy rate of the UDP detection agent growing[C].IEEE Trans on fuzzy systems, 2003,
11(4):518-527
is low because of too lack of attack records in training set.
[11] K.L LI, H.K Wang, S.F Tian, Z.P Liu, IU Zhi Qang. Fuzzy
How to improve the accuracy of UDP detection agent in Multi-Class Support Vector Machine and Application in
existing data set will be the major work of the next stage. Intrusion Detection [J]. Chinese Journal of
Computers.2005, 28(2):274-280
VIII. ACKNOWLEDGEMENT [12] Weston J.Watkins C.Multi-class support vector
This work was supported by Guangdong Provincial machines[C]. Department of Computer Science, Royal
Natural Science Foundation (Grant No. 06021484), Holloway University of London Technical Report, SD-TR-
Guangdong Provincial science & technology project 98-04, 1998
(Grant No. 2005B16001095, Grant No. 2005B10101077), [13] Syed N Liu H Sung K Incremental learning with
Yuexiu Zone Guangzhou city science & technology support vector machines [A].Proceedings of the Workshop
on Support Vector Machines at the International Joint
project (Grant No. 2007-GX-075).
Conference on Artificial Intelligence (IJCAI-99) [C].
IX. REFERENCES Stockholm, Sweden: M organ Kaufmann, 1999.876892.
[14] Yangguang Liu,Qinming He,Qi Chen.Incremental Batch
[1] Vapnik V. The Nature of Statistical Learning Theory [M]. Learning with Support Vector Machines[C].Proceedings of
New York: Springer-Verlag, 1995 the 5th World Congress on Intelligent Control and
[2] J. Han and M. Kamber, Data mining concepts and Automation.Hangzhou,China.2004 (2):1857-1861
techniques (Second Edition), China machine press, 2006 [15] WANG Ding-cheng, JIANG Bin. Review of SVM-based
[3] Burges C.A A tutorial on support vector machines for Control and Online Training Algorithms [J]. Chinese
pattern recognition[C]. Data Mining and Knowledge Journal of System Simulation, 2007, 19(6):1177-1181
discovery, 1998, 2(2):121-167. [16] Xiaodan Wang, Chunying Zheng, Chongming Wu. A New
[4] Rao Xian, Dong Chun-Xi, Yang Shao-Quan. An Intrusion Algorithm for SVM Incremental Learning[C].ICSP2006
Detection System Based on Support Vector Machines [J]. Proceedings.Beijing, China.2006
Chinese Journal of Software.2003, 14(4):798-803 [17] WEN-JIAN WANG. A redundant incremental learning
[5] CHEN You, SHEN Hua-Wei, LI Yang, CHENG Xue-Qi. algorithm for SVM[C]. Proceedings of the Seventh
An Efficient Feature Selection Algorithm Toward Building International Conference on Machine Learning and
Lightweight Intrusion Detection System [J]. Chinese Cybernetics[C], Kunming, China. 2008.734-738.
Journal of Computers, 2007, 30(8):1398-1408 [18] TAO Liang. Fast Incremental SVM Learning Algorithm
[6] Li Hui, Guan Xiaohong, Zan Xin. Network Intrusion Based on Active Set Iterations [J]. Chinese Journal of
Detection Based on Support Vector Machine [J]. Journal of System Simulation, 2006, 18(11):3305-3308
Computer Research and Development, 2003, 40(6):799- [19] .Bernhard Scholkopf, Smola A.Williamson R.C.et a1. New
807 support vector algorithms [J], Neural Computation.2000,
[7] XIAO Lizhong,Shao Zhiqing Ma Hanhua. An algorithm 12(5):1207-1245
for automatic clustering number determination in networks

2010 ACADEMY PUBLISHER


JOURNAL OF NETWORKS, VOL. 5, NO. 4, APRIL 2010 483

[20] Lin, C.F. Wang, S.D. Fuzzy Support Vector Machines[C]. in the School of Industrial Engineering, Purdue University,
IEEE Transactions on Neural Networks 2002, 13(2):464- West Lafayette, USA. In 1999, 2004, and 2007-2009, he was a
471 visiting professor with the Department of Industrial Engineering,
[21] Huang, H.P. Liu, Y.H. Fuzzy Support Vector Machines for Arizona State University, Tempe, USA, the Department of
Pattern Recognition and Data Mining [J]. International Electrical and Computer Engineering, New Jersey Institute of
Journal of Fuzzy Systems, 2002 4(3):826-835 Technology, Newark, USA, and Industrial Systems
[22] Yanyou Hao, Zhongxian Chi, Deqin Yan, Xun Yu. An Engineering Department, Industrial Systems Optimization
Improved Fuzzy Support Vector Machine for Credit Laboratory, University of Technology of Troyes, Troyes,
Rating[C]. IFIP International Federation for Information France, respectively. He is currently a Professor of Industrial
Processing 2007. 2007 495-505 and Systems Engineering in the Department of Industrial
[23] Li XLLiu JMShi ZZA Chinese Web page classifier Engineering, School of Mechatronics Engineering, Guangdong
based on support vector machine and unsupervised University of Technology, Guangzhou, China. His research
clustering Chinese Journal of Computers 2001 24 interests include production planning and scheduling,
(1):6268. manufacturing system modeling and control, discrete event
[24] http://kdd.ics.uci.edu/databases/kddcup99/kddcup99.html systems, Petri net theory and applications, and information
[25] Chih-Chaung Chang and Chih-Jen Lin, LIBSVM: a library assurance. He is the author or coauthor of many papers
for support vector machines, 2001.Software available at published in International Journal of Production Research, IEEE
http://www.csie.ntu.tw/~cjlin/libsvm Transactions on Systems, Man, and Cybernetics, IEEE
Transactions on Robotics and Automation, IEEE Transactions
Shaohua Teng is a Professor of on Automation Science and Engineering, IEEE Transactions on
Guangdong University of Technology in Semiconductor Manufacturing, IEEE/ASME Transactions on
China. He was born on Jan in 1962. He is Mechatronics, Journal of Intelligent Manufacturing, Production
responsible for teaching data mining in Planning and Control, and Robotics and Computer Integrated
Faculty of Computer. He is engaged in Manufacturing.
education and technology transfer on Dr. Wu is an associate editor of the IEEE Transactions on
knowledge discovery issues, and research in Systems, Man, & Cybernetics, Part C and IEEE Transactions on
network security, machine learning and statistical pattern Automation Science and Engineeringand editor in chief of
recognition. Dr. Teng earned a Ph.D. in Industry Engineering at Industrial Engineering Journal. He was a Program Committee
Guangdong University of Technology. He has published 50 Member of the 2003 to 2009 IEEE International Conference on
papers on computer magazines and international conferences Systems, Man, & Cybernetics, a Program Committee Member
and 2 books. The three Papers are listed as follows in recent of the 2005 to 2009 IEEE International Conference on
years: Networking, Sensing and Control, a Program Committee
1. Shaohua Teng, Wenwei Tan, Video Temporal Member of the 2006 IEEE International Conference on service
Segmentation Using Support Vector Machine, Lecture systems and service management, a Program Committee
Notes in Computer Science, Vol. 4993, LNCS, 2008:442- Member of the 2007 International Conference on Engineering
447 and Systems Management, and reviewer for many international
2. Shaohua Teng, Wei Zhang, and et al, Cooperative journals.
intrusion detection model based on state transition
analysis, Lecture Notes in Computer Science, Vol. 5236, Wei Zhang is a associate Professor of
2008: 419-431 Guangdong University of Technology in
3. Shaohua Teng, Wenwei Tan, Wei Zhang, Cooperative China. She was responsible for teaching data
shot boundary detection for video, Lecture Notes in mining in Faculty of Computer. She is
Computer ScienceVol. 5236, 2008: 99-110 engaged in network security, machine
Hongle DU is a graduate of learning and statistical pattern recognition.
Guangdong University of Technology in Mrs. Zhang earned a M.S. in Software
China. He was born in 1979. His major Engineering from the South China University
research interests include Network of Technology. She has published 20 papers on computer
security, intrusion detection based on magazines and international conferences. The three Papers are
Support Vector Machines and machine listed as follows in recent years:
learning. 1. Wei Zhang, Shaohua Teng, Xiufen Fu, Haibin Zhu, Roles
in Learning Systems, SMC 2008, 2008 IEEE International
NaiQi Wu (M04-SM05) received the Conference on Systems, Man, and Cybernetics, 2008:
M. S. and Ph. D. Degree in Systems 2213-2218
Engineering both from Xian Jiaotong 2. Wei Zhang, Shaohua Teng, Xiufen Fu, Scan attack
University, Xian, China in 1985 and 1988, detection based on distributed cooperative model,
respectively. CSCWD 2008743-748
From1988 to 1995, he was with the Chinese Wei Zhang, Shaohua Teng, Zhaohui Zhu, Xiufen Fu, Haibin
Academy of Sciences, Shenyang Institute of Zhu, An improved Least-Laxity-First scheduling algorithm of
Automation, Shenyang, China, and from variable time slice for periodic tasks, IEEE Conf., ICCI 2007,
1995 to 1998, with Shantou University, 548-553
Shantou, China. From 1991 to 1992, he was a Visiting Scholar

2010 ACADEMY PUBLISHER

Anda mungkin juga menyukai