• Keine Ergebnisse gefunden

Active Learning for Network Intrusion Detection

N/A
N/A
Protected

Academic year: 2022

Aktie "Active Learning for Network Intrusion Detection"

Copied!
8
0
0

Wird geladen.... (Jetzt Volltext ansehen)

Volltext

(1)

Active Learning for Network Intrusion Detection

Nico Görnitz, Marius Kloft, Konrad Rieck, and Ulf Brefeld

Technische Universität Berlin Machine Learning Group

Franklinstr. 28/29 10587 Berlin, Germany

{goernitz,mkloft,rieck,brefeld}@cs.tu-berlin.de

ABSTRACT

Anomaly detection for network intrusion detection is usu- ally considered an unsupervised task. Prominent techniques, such as one-class support vector machines, learn a hyper- sphere enclosing network data, mapped to a vector space, such that points outside of the ball are considered anoma- lous. However, this setup ignores relevant information such as expert and background knowledge. In this paper, we rephrase anomaly detection as an active learning task. We propose an effective active learning strategy to query low- confidence observations and to expand the data basis with minimal labeling effort. Our empirical evaluation on net- work intrusion detection shows that our approach consis- tently outperforms existing methods in relevant scenarios.

Categories and Subject Descriptors

C.2.0 [Computer-Communication Networks]: Gen- eral—Security and protection; I.2.6 [Artificial Intelli- gence]: Learning—Parameter learning; I.5.2 [Pattern Recognition]: Design Methodology—Classifier design and evaluation

General Terms

Algorithms, Experimentation, Security

Keywords

Machine learning, anomaly detection, support vector data description, active learning, intrusion detection, network se- curity

1. INTRODUCTION

Computer systems linked to the Internet are exposed to a plethora of network attacks and malicious code. Several threats, ranging from zero-day exploits to Internet worms,

Permission to make digital or hard copies of all or part of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. To copy otherwise, to republish, to post on servers or to redistribute to lists, requires prior specific permission and/or a fee.

AISec’09,November 9, 2009, Chicago, Illinois, USA.

Copyright 2009 ACM 978-1-60558-781-3/09/11 ...$10.00.

target network hosts every day; networked systems are gen- erally at risk to be remotely compromised and misused for illegal purposes. While early attacks have been developed rather for fun than for profit, proliferation of current net- work attacks is driven by a criminal underground economy.

Compromised systems are often misused for monetary gains including the distribution of spam messages and theft of confidential data. The success of these illegal businesses poses a severe threat to the security of network infrastruc- tures. Alarming reports on an expanding dissemination of advanced attacks render sophisticated security systems in- dispensible, e.g. [13, 27].

Conventional defenses against such network threats rest on the concept of misuse detection. That is, attacks are identified in network traffic using known patterns of mis- use, so-called attack signatures. While misuse detection ef- fectively protects from known threats, it increasingly fails to cope with the amount and diversity of attacks. The time span required for crafting a signature from a newly discovered attack is insufficient for protecting from rapidly propagating malicious code, e.g. [14, 24]. Moreover, poly- morphism employed in recent attacks obstructs modelling accurate signatures [25], such that there is a demand for alternative techniques for detection of attacks during their initial propagation.

Anomaly detection methods provide means for identifying unknown and novel attacks in network traffic and thereby complement regular security defenses, e.g. [11, 10, 8, 19, 6, 21]. Anomaly detection methods proceed by learning a model of normal network data and identifying unusual contents and potential attacks as deviations thereof – ir- respective of the employed intrusion techniques. Although anomaly detection methods enable tracking novel threats, their practical deployment poses a dilemma to the security practitioner. On the one hand, recent attacks and network traffic are required to properly calibrate and validate a learn- ing method during operation. On the other hand, providing labels for network traces on a regular basis renders appli- cation of learning methods intractable in practice. Unfor- tunately, calibrating a method using unlabeled data only is not an option either, as the learned model of normality may be easily foiled by adversarial traffic [17, 7].

In this paper, we consider payload-based anomaly detec- tion methods, such as PAYL [32], Anagram [31] and Mc- PAD [16], which model normality by mapping network pay- loads to a vector space and enclosing the resulting vectors in a hypersphere. In contrast to previous approaches, how- ever, we phrase anomaly detection as anactive learning task.

(2)

Figure 1: Left: An exemplary solution of the SVDD. Right: Illustration of ActiveSVDD that incorporates unlabeled (green) as well as labeled data of the normal class (red) and attacks (blue).

That is, we present a learning method that processes unla- beled data but actively queries labels for particular network payloads. The selection process is designed to find unlabeled examples in the data which – once labeled – lead to the max- imal improvement of the learned hypersphere; the labeling effort for the practitioner is hereby significantly reduced.

We first present an effective active learning strategy to query network events of low confidence. The strategy cali- brates the threshold of the hypersphere-based learner. Sec- ondly, we extend hypersphere-based approaches to so-called semi-supervisedmodels that enable processing unlabeled as well as labeled examples. Our method is initially trained on unlabeled examples and then subsequently refined by in- corporating labeled data that have been queried by active learning rules. The training process can be terminated at any time, for instance when the desired predictive perfor- mance is obtained. The devised method contains unsuper- vised approaches such as centroids [32, 16, 21], as a special case that is obtained when no label information is used.

Empirical results on network intrusion detection demon- strate the benefit of combining anomaly detection and active learning. The active learning strategy significantly reduces the manual labeling effort for the practitioner. By labeling only a fraction of 1.5%, the detection rate was improved from 64% to 96% at a false-positive rate below 0.0015%.

This demonstrates the merits of active learning in practice.

Our paper is structured as follows. Section 2 introduces hypersphere-based anomaly detection and presentes our ex- tension using semi-supervised and active learning strategies.

Section 3 reports on empirical results of our approach us- ing real network traffic and attacks. Finally, Section 4 con- cludes.

2. METHODOLOGY

In this section, we present our methodology. Firstly, we describe how we derive numerical features from network pay- load data. Then, we review the classical hypersphere-based approach to anomaly detection. We discuss how anomaly detection can be equipped with an active learning strategy to adjust an anomaly threshold. Finally, we propose an in- tegrated method to compute hyperspheresand thresholds simultaneously.

2.1 From Network Payload to Feature Spaces

The detection of unknown and novel attacks requires an expressive representation of network contents, accessible to means of intrusion detection and machine learning. To this end, we apply a technique for embedding of network pay- loads in vector spaces derived from concepts of information retrieval [22] and recently applied in the realms of intrusion detection [19]. A network payloadx(the data contained in a network packet or connection) is mapped to a vector space using a set of stringsS and an embedding functionφ. For each strings ∈ S the functionφs(x) returns 1 if s is con- tained in the payloadxand 0 otherwise. By applyingφs(x) for all elements ofS we obtain the following map

φ:X →R|S|, φ:x7→(φs(x))s∈S, (1) where X is the domain of all network payloads. Defining a set S of relevant strings a priori is difficult in advance, as typical patterns of novel attacks are not available prior to their disclosure. As an alternative, we define the set S implicitly and associate S with all strings of lengthn. The resulting set of strings is often referred to asn-grams.

As a consequence of usingn-grams, the network payloads are mapped to a vector space with 256ndimensions, which apparently contradicts with efficient detection of intrusions.

Fortunately, a payload of lengthT comprises at most (T− n−1) different n-grams and, consequently, the map φ is sparse, that is, the vast majority of dimensions is zero. This sparsity can be exploited to derive linear-time algorithms for extraction and comparison of embedded vectors. Instead of operating with full vectors, only non-zero dimensions are considered, where the extracted strings associated with each dimension can be maintained in efficient data structures, such as hash tables [3], Bloom filters [31] or Tries [19].

2.2 Hypersphere-based Anomaly Detection

In this section, we briefly review anomaly detection using one-class support vector machines. In particular we study a variant proposed by Tax and Duin [28] referred to as support vector domain description (SVDD). Several approaches to anomaly detection for network intrusion detection, e.g. [20, 32, 31], can be shown to resemble special cases of the SVDD, given that an appropriate embedding of network events into some vector space as described in the previous section is per-

(3)

0 0.5 1 1.5 2 0

0.25 0.5 0.75 1

yf(x)

loss

linear

quadratic

linear hinge loss Huber loss

Figure 2: The differentiable Huber loss`∆=1,=0.5. formed. Further discussion and comparison of hypersphere- based anomaly detection is provided in [23, 29].

The goal of the SVDD is to find a concise description of the normal data such that anomalous data can be easily identified as outliers. In the underlying one-class scenario, this translates to finding a minimal enclosing hypersphere (i.e., centercand radiusR) that contains the normal input data [28], see Figure 1 (left). Given the function

f(x) =kφ(x)−ck2−R2,

the boundary of the hypersphere is described by the set{x: f(x) = 0∧x∈ X }. That is, the parameters off are to be chosen such thatf(x) ≤0 for normal data and f(x) >0 for anomalous points. The centercand the radiusRcan be computed accordingly by solving the following optimization problem [28]

R,c,ξmin R2

n

X

i=1

ξi

s.t. ∀ni=1: kφ(xi)−ck2≤R2i (2)

ni=1: ξi≥0.

The trade-off parameter η adjusts point-wise violations of the hypersphere. That is, a concise description of the data might benefit from omitting some data points in the com- putation of the solution. Discarded data points induce slack that is absorbed by variablesξi. Thus, in the limitη→ ∞, the hypersphere will contain all input data, whileη→0 im- pliesR→0 and the centercreduces to the centroid of the data.

Some actions to the SVDD have been proposed to incor- porate labeled data in the learning process, e.g. [5, 9, 30, 29], the resulting optimization problems are no longer con- vex and the proposed optimization in dual space might suffer from duality gaps. Techniques for actively guiding the learn- ing of the SVDD have been not considered so far although active learning for anomaly detection has been studied by [26, 15, 1]. [1] take a max-margin approach and propose to query points that lie close to the decision hyperplane and violate the margin criterion in order to minimize the error rate. By contrast, the approach by [15] aims at detecting rejection categories in the data using as few queries as possi- ble. Finally, the approach taken in [26] combines the former two active learning strategies to find interesting regions in feature spaceandto decrease the error-rate simultaneously.

In this section we devise an efficient strategy to query net- work events which lie in low-confidence regions of the feature space, hence guiding the security expert in the labeling pro- cess. Thisactive learningstrategy selects an instance of the unlabeled data pool and presents it to the security expert.

The selection process is designed to yield the maximal im- provement of the actual model.

Our strategy takes unlabeled as well as already labeled examples into account. Without loss of generality, we denote the unlabeled examples byx1, . . . ,xn and the labeled ones byxn+1, . . . ,xn+m, wherenm. Every labeled example xi is annotated with a label yi ∈ {+1,−1}, depending on whether it is classified as benign (yi= +1) or malicous (yi=

−1) data.

We begin with a commonly used active learning strat- egy which simply queries borderline points. The strategy is sometimes called margin strategy and can be expressed by asking the user to label the point x0 that is closest to the decision hypersphere [1, 33]

x0 = argmin

xi∈{x1,...,xn}

λ1(xi) (3)

:= argmin

xi∈{x1,...,xn}

|R2− kφ(xi)−ck2|

Ω ,

where Ω is a normalization constant and given by Ω = maxi|R2− kφ(xi)−ck2|.

However, when dealing with many non-stationary outlier and/or attack categories, it is beneficial to identify novel at- tacks as soon as possible. We translate this into an active learning strategy as follows. Let A = (ast)s,t=1,...,n+m be an adjacency matrix, for instance obtained by ak-nearest- neighbor approach, where aij = 1 if xi is among the k- nearest neighbors of xj and 0 otherwise. Equation (4) im- plements the above idea and returns the unlabeled instance according to

x0 = argmin

xt∈{x1,...,xn}

λ2(xt) (4)

:= argmin

xt∈{x1,...,xn}

Pn

i=1ait+Pn+m j=n+1yjajt

2k .

The above strategy explores unknown regions in feature space and subsequently deepens the learned knowledge by querying clusters of potentially similar objects to allow for good generalizations.

Nevertheless, using Equation (4) alone may result in querying points lying close to the center of the hypersphere or far from its boundary. These points will hardly contribute to an improvement of the hypersphere. In other words, only a combination of both strategies (3) and (4) guarantees the active learning to query points of interest. Our final active learning strategy is therefore given by

x0= argmin

xt∈{x1,...,xn}

τ λ1(xt) + (1−τ)λ2(xt) (5) forτ ∈[0,1]. The combined strategy queries instances that are close to the boundary of the hypersphereand lie in po- tentially anomalous clusters with respect to the k-nearest neighbor graph. Depending on the actual value of τ, the strategy jumps from cluster to cluster and thus helps to identify interesting regions in feature space. For the special case of no labeled points our combined strategy reduces to the margin strategy.

(4)

2.3 An integrated approach: ActiveSVDD

The active learning strategy from the previous section queries low-confidence points to improve the current hypoth- esis. The idea is, that the model can be re-trained after quering some points, using the unlabeledas well as thethe newly labeled data. Unfortunately, the vanilla SVDD can- not make use of labeled data. In this section, we extend the SVDD to support active learning and propose the integrated method ActiveSVDD which determines a hypersphere and a radius simultaneously.

As in Section 2.2, we aim at finding a model f(x) =

||φ(x)−c||2−R2that generalizes well on unseen data, how- ever, the model is now devised on the basis of labeled and unlabeled data. A straight-forward extension of the SVDD in Equation (2) using both, labeled and unlabeled examples, is given by

R,γ,c,ξmin R2−κγ+ηu n

X

i=1

ξil n+m

X

j=n+1

ξj

s.t. ∀ni=1: kφ(xi)−ck2≤R2i

n+mj=n+1: yj

`kφ(xj)−ck2−R2´

≤ −γ+ξj (6)

ni=1: ξi≥0,

n+mj=n+1: ξj≥0.

The optimization problem has additional constraints for the labeled examples that have to fulfill the margin criterion with marginγ. Trade-off parametersκ,ηu, andηlbalance margin-maximization and the impact of unlabeled and la- beled examples, respectively. To avoid cluttering the no- tation unnecessarily, we omit the obvious generalization of allowing different trade-offs η+l and ηl for positively and negatively labeled instances, respectively. The additional slack variablesξj are bound to labeled examples and allow for point-wise relaxations of margin violations by labeled ex- amples. The solution of the above optimization problem is illustrated in Figure 1 (right).

The inclusion of negatively labeled data turns the above optimization problem non-convex and optimization in the dual is prohibitive. As a remedy, we translate Equation (6) into anunconstrainedproblem [2, 34] as follows,

min

R,γ,c R2−κγ+ηu n

X

i=1

``

R2− ||φ(xi)−c||2´

(7)

l n+m

X

j=n+1

`` yj

`R2− ||φ(xj)−c||2´

−γ´ . where `(t) = max{−t,0}is the common hinge loss. Note that the optimization problems in Equations (6) and (7) are equivalent so far. Nevertheless, the non-smoothness of the objective prohibits an efficient optimization. Hence, we substitute the Huber loss for the hinge loss to obtain a smooth and differentiable function that can be optimized with gradient-based techniques. The Huber loss`∆, is dis- played in Figure 2 and given by

`∆,(t) = 8

<

:

∆−t : t≤∆−

(∆+−t)2

4 : ∆−≤t≤∆ + 0 : otherwise.

0 1 2 3 4 5 7.5 10 15

0.7 0.75 0.8 0.85 0.9 0.95 1

labeled data in %

AUC [0,0.01]

activeSVDD SVDD

Figure 3: Results for normal vs. malicious.

For our purposes ∆ = 0 suffices and the final objective function can be stated as,

min

R,γ,c R2−κγ+ηu n

X

i=1

`0,

`R2− ||φ(xi)−c||2´ (8)

l n+m

X

j=n+1

`0,

`yj

`R2− ||φ(xj)−c||2´

−γ´ . Notice that by rephrasing the problem as an unconstrained, smooth optimization problem, its intrinsic complexity has not changed. However, the local minima of Optimization Problem (8) can now easily be found with gradient-based techniques such as conjugate gradient descent, see Appendix A for details. Note that, in general, unconstrained optimiza- tion is also easier to implement than constrained optimiza- tion. We will observe the benefit of this approach in the following.

3. EMPIRICAL EVALUATION

We proceed to present an empirical evaluation of our novel method ActiveSVDD for intrusion detection using real net- work traffic. In particular, we are interested in studying the performance gain attained by our active learning strat- egy in comparison to the unsupervised formulation of the SVDD [29]. Several approaches to learning-based intrusion detection constitute special cases of the SVDD, e.g. [32, 21, 16], and hence are implicitly reflected in our experiments.

The ActiveSVDD is trained by solving Equation (8) using conjugate gradient descent, where the optimization problem underlying the SVDD is solved using SMO [18]. Parameters of the active learning strategy are set tok= 10,α= 0.1 for simplicity.

3.1 Data Corpus

For our experiments, we consider HTTP traffic recorded within 10 days at Fraunhofer Institute FIRST. The data set comprises 145,069 unmodified connections with an av- erage length of 489 bytes. The incoming byte stream of each connection is mapped to a vector space using 3-grams as detailed in Section 2.1. We refer to the FIRST data as thenormal pool. Themalicious pool contains 27 real attack classes generated using the Metasploit framework [12]. It covers 15 buffer overflows, 8 code injections and 4 other at-

(5)

0 3 6 10 15 0.6

0.65 0.7 0.75 0.8 0.85 0.9 0.95 1

labeled data in %

AUC [0,0.01]

SVDD

activeSVDD with random sampling

0 1 2 3 4 5 7.5 10 15

0.6 0.65 0.7 0.75 0.8 0.85 0.9 0.95 1

labeled data in %

AUC [0,0.01]

SVDD

activeSVDD with active learning

Figure 4: Results for normal vs. cloaked. Left: Random samling. Right: Active learning.

tacks including HTTP tunnels and cross-site scripting. Ev- ery attack is recorded in 2–6 different variants using virtual network environments and decoy HTTP servers, where the attack payload is adapted to match characteristics of the normal data pool.

To study the robustness of our approach in a more realistic scenario, we also consider techniques to obfuscate malicious content by adapting attack payloads to mimic benign traffic in feature space [4]. As a consequence, the extracted features deviate less from normality and the classifier is likely to be fooled by the attack. For our purposes, it already suffices to study a simple cloaking technique by adding common HTTP headers to the payload while the malicious body of the attack remains unaltered. We apply this technique to the malicious pool and refer to the obfuscated set of attacks ascloaked pool.

3.2 Active Learning Experiment

In our first experiment we focus on two scenarios: nor- mal vs. malicious and normal vs. cloaked data. For both settings, we randomly draw 966 training examples from the normal pool and 34 attacks either from the malicious or the cloaked pool, depending on the scenario. Holdout and test sets are also drawn at random and consist of 795 normal con- nections and 27 attacks, each. We make sure that attacks of the same attack class occur either in the training, or in the test set but not in both. Note that all attacks in the training data are unknown to the learning methods, unless an active labeling strategy is performed. We report on 10 repetitions with distinct training, holdout, and test sets and measure the performance by the area under the ROC curve in the false-positive interval [0,0.01] (AUC0.01).

Figure 3 shows the results for normal vs. malicious data pools, where the x-axis depicts the percentage of labeled in- stances which are selected using random sampling. Irrespec- tively of the amount of labeled data, the malicious traffic is detected by all methods equally well, as the intrinsic nature of the attacks is sufficiently captured by the representation of 3-grams. There is no significant difference between the detectors. However, our next experiment shows the fragility of these results in the presence of simple cloaking techniques.

Simply obfuscating the attacks by copying normal headers into the malicious payload leads to dramatically different results.

Figure 4 (left) displays the results for normal vs. cloaked data, where network connections to be labeled for the Ac- tiveSVDD are chosen randomly. First of all, the perfor- mance of the unsupervised SVDD drops to only 70%, as the cloaked attacks successfully foil the detection process. By contrast, the ActiveSVDD benefits from labeled data and clearly shows a reasonable accuracy. For only 2% labeled data, the ActiveSVDD easily outperforms the vanilla SVDD and for labeling 5% of the available data it separates almost perfectly between normal and cloaked malicious traffic.

0 1 2 3 4 5 7.5 10 15

0 5 10 15 20 25 30 35

labeled data in %

attacks optimal

active learning margin

random (empirical) random (theoretical)

Figure 5: Number of attacks found by different ac- tive learning strategies.

Nevertheless, labeling 30% of the data is not realistic for practical applications. We thus explore the benefit of ac- tive learning for inquiring label information of borderline and low-confidence points. Figure 4 (right) shows the re- sults for normal vs. cloaked data where the labeled data for ActiveSVDD is chosen according to the active learning strategy in Equation (5). The unsupervised SVDD does not make use of label information and remains at an AUC0.01of 70%. Compared to the results for a random labeling strat- egy (Figure 4, left), the performance of the ActiveSVDD clearly improves for active learning. Using active learning, we need to label only 3% of the data for attaining an almost perfect separation, compared to 30% for a random labeling strategy. Our active learning strategy effectively boosts the

(6)

1 2 3 4 5 0

0.2 0.4 0.6 0.8 1

chunk

AUC [0,0.01]

activeSVDD with random sampling activeSVDD with active learning SVDD

0 0.02 0.04 0.06 0.08 0.1

0 0.2 0.4 0.6 0.8 1

FPR

TPR

activeSVDD with active learning activeSVDD with random sampling

Figure 6: Online application of ActiveSVDD over different chunks. Left: Progress over chunks. Right: ROC curve for all chunks.

performance and reduces the manual labeling effort signifi- cantly.

Figure 5 details the impact of our active learning strat- egy in Equation (5). We compare the number of outliers detected by the combined strategy with the margin-based strategy in Equation (3) (see also [1, 26]) and by randomly drawing instances from the unlabeled pool. As a sanity check, we also included the theoretical outcome for ran- dom sampling. The results show that the combined strat- egy effectively detects malicious traffic much faster than the margin-based strategy.

3.3 Online Application

When employing network intrusion detectors in practice, one faces a steadily increasing amount of unlabeled network data. Holding all instances in memory gets infeasible over time such that means for learning online become a crucial requirement. The next experiment aims at investigating our method ActiveSVDD in anonline learningscenario, i.e.

when the normal data pool steadily increases. To this end, we draw a sample of 3,750 network events from the normal pool, where 1,250 connections are used as test set and the remaining data is decomposed into five chunks of equal size for online application. Cloaked attacks are mixed into all samples and we take care that the same attack classes are not present in the training and test data. The ActiveSVDD is then trained on an increasing number of chunks, starting from the first and finally using all five chunks. For each chunk we adjust the active learning strategy such that in average only 10 data points need to be labeled.1

Figure 3.2 shows the change in accuracy of the Ac- tiveSVDD over the different chunks. Results are averaged over 10 random draws of data splits. One can see that with increasing amount of network data the active learning strat- egy steadily drives the learner to high accuracy predictions while the random strategy is too slow to adapt. The vanilla SVDD performs worse since it doesn’t profit from the labels.

Figure 3.2 shows a ROC curve for the ActiveSVDD and the regular SVDD obtained after learning on all five chunks. By

1We feel that such a small amount of data labelings is re- alistic in the light of massive incoming traffic and the high costs of a human experts.

only labeling a fraction of 1.5% the ActiveSVDD enables detecting 96% of the cloaked attacks at a false-positive rate at 0.0015%. By contrast, the vanilla SVDD identifies only 64% attacks at the same false-positive rate.

3.4 Threshold Adaption

The previous experiments demonstrate the advantages of active learning for network intrusion detection. So far, all results have been obtained using our method ActiveSVDD, however, the active learning techniques devised in Section 2 are also applicable for calibrating other learning-based meth- ods. We herein focus on the vanilla SVDD with param- eter ν = 1, which corresponds to classical centroid-based anomaly detection, such that results directly transfer to anomaly detectors as Anagram and PAYL.

We again draw a set of 3,750 network connections form the pool of normal data and split the resulting set into a training set of 2,500 connections and a test partition of 1,250 events. Both sets are mixed with cloaked attack instances.

The SVDD is then trained on the normal training set. For application of the learned hypersphere to the test set, we evaluate different strategies for determining a radius using random sampling and active learning. In both cases, the selected connections are labeled and a threshold is obtained by computing the mean of all labeled instances.

Figure 7 shows for various levels of labeled data the ROC curve of the SVDD and the computed thresholds that have been derived from the radius outputed by the SVDD. Re- sults have been averaged over 10 random draws of working sets. One can see that even for small amounts of labeled data the active learning strategy finds a reasonable radius while the random strategy and the vanilla SVDD completely fail with false-positive rate of 0.5 and 1 respectively. This result demonstrates that active learning strategies enable calibrating anomaly detection with significantly reduced ef- fort in comparison to random sampling and hence provide a valuable instrument when deploying learning methods in practice.

4. CONCLUSION

In this paper, we proposed to view anomaly detection as an active learning problem to allow for the inclusion of prior

(7)

0 0.2 0.4 0.6 0.8 1 0

0.2 0.4 0.6 0.8 1

FPR

TPR

ROC curve SVDD threshold active learning 0.05%

random sampling 0.05%

Figure 7: Treshold adaption for the SVDD

and expert knowledge. To reduce the labeling effort for the practitioner, we devised an active learning strategy to query instances that are not only close to the boundary of the hypersphere but also likely members of novel rejection cat- egories. To use labeled as well as unlabeled instances in the training process, we proposed ActiveSVDDs as a general- ization of SVDDs. The resulting unconstraint, smooth opti- mization problem can be optimized with efficient gradient- based techniques.

Empirically, we showed for network intrusion detection, that rephrasing the unsupervised problem setting as an ac- tive learning task is worth the effort. ActiveSVDDs prove robust in scenarios where the performance of baseline ap- proaches deteriorate due to obfuscation techniques. More- over, we observe the effectiveness of our active learning strategy which significantly improves the quality of the Ac- tiveSVDD and spares practitioners from labeling unneces- sarily many data points. For experiments on sequentially arriving data chunks, the ActiveSVDDs achieve a perfect separation of normal and attack data and outperform its unsupervised counterpart significantly.

Acknowledgements

This work was supported in part by the German Bundesmin- isterium f¨ur Bildung und Forschung (BMBF) under the project ReMIND (FKZ 01-IS07007A) and by the FP7-ICT Programme of the European Community, under the PAS- CAL2 Network of Excellence, ICT-216886.

5. REFERENCES

[1] M. Almgren and E. Jonsson. Using active learning in intrusion detection.Proc. IEEE Computer Security Foundation Workshop, 2004.

[2] O. Chapelle and A. Zien. Semi-supervised

classification by low density separation. InProceedings of the International Workshop on AI and Statistics, 2005.

[3] M. Damashek. Gauging similarity withn-grams:

Language-independent categorization of text.Science, 267(5199):843–848, 1995.

[4] P. Fogla, M. Sharif, R. Perdisci, O. Kolesnikov, and W. Lee. Polymorphic blending attacks. InProceedings of USENIX Security Symposium, 2006.

[5] C.-H. Hoi, C.-H. Chan, K. Huang, M. Lyu, and I. King. Support vector machines for class

representation and discrimination. InProceedings of the International Joint Conference on Neural Networks, 2003.

[6] K. L. Ingham, A. Somayaji, J. Burge, and S. Forrest.

Learning DFA representations of HTTP for protecting web applications.Computer Networks,

51(5):1239–1255, 2007.

[7] M. Kloft and P. Laskov. A poisoning attack against online anomaly detection. InNIPS Workshop on Machine Learning in Adversarial Environments for Computer Security, 2007.

[8] C. Kruegel, G. Vigna, and W. Robertson. A multi-model approach to the detection of web-based attacks.Computer Networks, 48(5), 2005.

[9] Y. Liu and Y. F. Zheng. Minimum enclosing and maximum excluding machine for pattern description and discrimination. InICPR ’06: Proceedings of the 18th International Conference on Pattern Recognition, pages 129–132, Washington, DC, USA, 2006. IEEE Computer Society.

[10] M. Mahoney. Network traffic anomaly detection based on packet bytes. InProc. of ACM Symposium on Applied Computing, pages 346 – 350, 2003.

[11] M. Mahoney and P. Chan. PHAD: Packet header anomaly detection for identifying hostile network traffic. Technical Report CS-2001-2, Florida Institute of Technology, 2001.

[12] K. Maynor, K. Mookhey, J. F. R. Cervini, and K. Beaver. Metasploit toolkit. InSyngress, 2007.

[13] Microsoft. Microsoft security intelligence report:

January to June 2008. Microsoft Corporation, 2008.

[14] D. Moore, C. Shannon, and J. Brown. Code-Red: a case study on the spread and victims of an internet worm. InProc. of Internet Measurement Workshop (IMW), pages 273–284, 2002.

[15] D. Pelleg and A. Moore. Active learning for anomaly and rare-category detection.Proc. Advances in Neural Information Processing Systems, 2004.

[16] R. Perdisci, D. Ariu, P. Fogla, G. Giacinto, and W. Lee. McPAD: A multiple classifier system for accurate payload-based anomaly detection.Computer Networks, 2009. in press.

[17] R. Perdisci, D. Dagon, W. Lee, P. Fogla, and

M. Sharif. Misleading worm signature generators using deliberate noise injection. InProc. of IEEE

Symposium on Security and Privacy, pages 17–31, 2006.

[18] J. C. Platt. Fast training of support vector machines using sequential minimal optimization. InAdvances in kernel methods: support vector learning, 1999.

[19] K. Rieck and P. Laskov. Detecting unknown network attacks using language models. InDetection of Intrusions and Malware, and Vulnerability

Assessment, Proc. of 3rd DIMVA Conference, LNCS, pages 74–90, July 2006.

[20] K. Rieck and P. Laskov. Language models for detection of unknown attacks in network traffic.

Journal in Computer Virology, 2(4):243–256, 2007.

[21] K. Rieck, S. Wahl, P. Laskov, P. Domschitz, and K.-R.

(8)

M¨uller. A self-learning system for detection of anomalous sip messages. InPrinciples, Systems and Applications of IP Telecommunications (IPTCOMM), Second International Conference, LNCS, pages 90–106, 2008.

[22] G. Salton, A. Wong, and C. Yang. A vector space model for automatic indexing.Communications of the ACM, 18(11):613–620, 1975.

[23] B. Sch¨olkopf, J. Platt, J. Shawe-Taylor, A. Smola, and R. Williamson. Estimating the support of a

high-dimensional distribution.Neural Computation, 13(7):1443–1471, 2001.

[24] C. Shannon and D. Moore. The spread of the Witty worm.IEEE Security and Privacy, 2(4):46–50, 2004.

[25] Y. Song, M. Locasto, A. Stavrou, A. Keromytis, and S. Stolfo. On the infeasibility of modeling polymorphic shellcode. InConference on Computer and

Communications Security (CCS), pages 541–551, 2007.

[26] J. W. Stokes and J. C. Platt. Aladin: Active learning of anomalies to detect intrusion. Technical report, Microsoft Research, 2008.

[27] Symantec. Symantex report on the underground economy: July 07 to June 08. Symantec Corporation, 2008.

[28] D. M. Tax and R. P. Duin. Support vector data description.Machine Learning, 54:45–66, 2004.

[29] D. M. Tax.One-class classification. PhD thesis, Technical University Delft, 2001.

[30] J. Wang, P. Neskovic, and L. N. Cooper. Pattern classification via single spheres.Computer Science:

Discovery Science (DS), 2005.

[31] K. Wang, J. Parekh, and S. Stolfo. Anagram: A content anomaly detector resistant to mimicry attack.

InRecent Adances in Intrusion Detection (RAID), pages 226–248, 2006.

[32] K. Wang and S. Stolfo. Anomalous payload-based network intrusion detection. InRecent Adances in Intrusion Detection (RAID), pages 203–222, 2004.

[33] M. K. Warmuth, J. Liao, G. R¨atsch, M. Mathieson, S. Putta, and C. Lemmen. Active learning with support vector machines in the drug discovery process.

Journal of Chemical Information and Computer Sciences, 43(2):667–673, 2003.

[34] A. Zien, U. Brefeld, and T. Scheffer. Transductive support vector machines for structured variables. In Proceedings of the International Conference on Machine Learning, 2007.

APPENDIX

In this section, we compute the gradient of ActiveSVDD.

For problem (6), the slacks can be expressed as ξi=``

R2− ||φ(xi)−c||2´ ξj=``

yj`

R2− ||φ(xj)−c||2´

−γ´ ,

respectively. Furthermore the derivative of the Huber loss

`∆, is given by

`0∆,(t) = 8

<

:

−1 : t≤∆−

12(∆−t + 1) : ∆−≤t≤∆ + 0 : otherwise .

For notational convenience, we focus on the Huber loss for

`∆=0,(t). Using the Huber loss`0,, computing the gradients of the slack variablesξiassociated with unlabeled examples with respect to the primal variablesRandcyields

∂ξi

∂R= 2R`0(R2− ||φ(xi)−c||2)

∂ξi

∂c = 2(φ(xi)−c)`0(R2− ||φ(xi)−c||2).

The derivatives of their counterparts ξj for the labeled ex- amples with respect toR,γ, andcare given by

∂ξj

∂R = 2yjR`0` yj

`R2− ||φ(xj)−c||2´

−γ´

∂ξj

∂γ =−`0

`yj

`R2− ||φ(xj)−c||2´

−γ´

∂ξj

∂c = 2yj(φ(xj)−c)`0` yj

`R2− ||φ(xj)−c||2´

−γ´ . Substituting the partial gradients, we resolve the gradient of Equation (7) with respect to the primal variables:

∂EQ7

∂R = 2R+ηu n

X

i=1

∂ξi

∂R+ηl n+m

X

j=n+1

∂ξj

∂R, (9)

∂EQ7

∂γ =−κ+ηl n+m

X

j=n+1

∂ξj

∂γ, (10)

∂EQ7

∂c =ηu n

X

i=1

∂ξi

∂c +ηl n+m

X

j=n+1

∂ξj

∂c. (11) The above equations can be plugged directly into off-the- shelf gradient-based optimization tools to optimize Equation (7) in the input space for the identityφ(x) =x. However, predictive power is often related to (possibly) non-linear mappings φ of the input data into some high-dimensional feature space. In the following, we extend our approach to allow for the use of non-linear feature embeddings. An ap- plication of the representer theorem shows that the centerc can be expanded as

c=X

i

αiφ(xi) +X

j

αjyjφ(xj). (12) According to the chain rule, the gradient of Equation (7) with respect to theαi/j is given by

∂EQ7

∂αi/j

= ∂EQ7

∂c

∂c

∂αi/j

. Using Equation (12), the partial derivatives ∂α∂c

i/j resolve to

∂c

∂αi

=φ(xi) and ∂c

∂αj

=yjφ(xj), (13) respectively. Applying the chain-rule to Equations (9),(10),(11), and (13) gives the gradients of Equation (7) with respect to theαi/j.

Referenzen

ÄHNLICHE DOKUMENTE

Throughout the 1980s, numerous leaders in the field of higher education (Cross 1987) and a series of national reports (Study Group 1984) repeatedly urged college and university

Dieses Dokument enthält eine Hardware-/Software-Kompatibilitätsmatrix für die Cisco Intrusion Prevention System (IPS) Appliances (4210, 4215, 4220, 4230, 4235, 4240, 4250,

To formalize this definition, we here specify two aspects more precisely: a “concept of normality” and what “deviates considerably” signifies. We define a concept of normality as

Trained on normal sequences, the model proved to work for the detection of anomalous patterns in multivariate sensor data of varying length.. In contrast, the GAN-based approach

Machine Learning, Fraud Detection, Financial Data, Big Data, Support Vector Machine, CRISP- DM, One Class Support Vector Machine, Principal Component Analysis, Ensemble

On the other hand, the second part is an adaptive classifier that includes a classifier Model based on the Enhanced Growing Hierarchical Self Organizing Map (EGHSOM), a

The aim of this thesis has been to improve the detection capability of attacks in local area networks. With a glance at the area of safety engineering, it appears efficient to

The state-of-the-art statistical method for binary classification, the Logistic regression, will be explained and its idea will be used for construction of the deep learning