1 / 32

A Geometric F ramework for Unsupervised A nomaly D etection: Detecting Intrusions in Unlabeled Data

A Geometric F ramework for Unsupervised A nomaly D etection: Detecting Intrusions in Unlabeled Data. Authors: Eleazar E skin , Andrew Arnold, Michael Prerau , Leonid Portnoy , Sal Stolfo. Presenter: Marbin Pazos -Revilla Cognitive Radio Group TTU-2011. Motivation.

bernad
Télécharger la présentation

A Geometric F ramework for Unsupervised A nomaly D etection: Detecting Intrusions in Unlabeled Data

An Image/Link below is provided (as is) to download presentation Download Policy: Content on the Website is provided to you AS IS for your information and personal use and may not be sold / licensed / shared on other websites without getting consent from its author. Content is provided to you AS IS for your information and personal use only. Download presentation by click this link. While downloading, if for some reason you are not able to download a presentation, the publisher may have deleted the file from their server. During download, if you can't get a presentation, the file might be deleted by the publisher.

E N D

Presentation Transcript


  1. A Geometric Framework for Unsupervised Anomaly Detection: Detecting Intrusions in Unlabeled Data Authors: EleazarEskin, Andrew Arnold, Michael Prerau, Leonid Portnoy, Sal Stolfo Presenter: MarbinPazos-Revilla Cognitive Radio Group TTU-2011

  2. Motivation • Machine Learning Algorithms • Cluster • K-Means • SVM • Datasets • KDD Cup • Intrusion Detection • Among best ROC curves and overall IDS performance

  3. Contributions • The authors proposed three improved methods for clustering, K-NN and SVM to be used in Unsupervised Intrusion Detection • The methods show to have very good performance (ROC curves)

  4. Introduction • Commercially available methods for intrusion detection employ signature based detection • The signature database has to be manually revised for newly discovered signatures and until a new update is applied systems are left vulnerable to new attacks

  5. IDS Types • Misuse • Each instance in a set of data is labeled as normal or intrusion, and a machine learning algorithm is trained over the labeled data • Classification rules • Manuel updates are needed • Anomaly • A given normal set data is given • A new set of data is tested and system is supposed to detect whether it is normal or not • It can detect new types of attacks

  6. Supervised Anomaly Detection • Supervised Anomaly Detection require a set of purely normal data from which they train their model. If intrusions are present in “normal” data, then these intrusions won’t be detected. • It is hard in practice to have labeled or purely normal data • In the event of having labeled data by simulating intrusions, we would be limited by the set of known attacks in the simulation

  7. Unsupervised Anomaly Detection • Goal is to differentiate normal elements from anomalous elements buried in the data • Do not require a purely normal training set • No need for labeled data • Raw data is much easier to obtain

  8. Geometric Framework • Maps Data to a d-dimentional Feature Space • Better capture intrusion in this feature space • Represent and map different types of data • Data-dependent normalization feature map • Spectrum Kernel feature map • Points can be classified as outliers (anomalies) based on their position in this space • In general anomalies tend to be distant from other points (parallel with sparse)

  9. Datasets and Algorithms • Datasets • KDD CUP 99 data (IDS dataset) • Lincoln Labs DARPA intrusion detection evaluation • Algorithms • Clustering • KNN • SVM

  10. Unsupervised Anomaly Detection • Intrusions are buried in the data • Can help in forensic analysis • Assumptions • Most (significant) of the elements are normal • Anomalies are qualitatively different than normal instances • With the previous assumptions anomalies will appear to be rare and different from normal elements and show as outliers

  11. Geometric Framework for Unsupervised Anomaly Detection • Mapping records from audit stream to a feature space • The distance between two elements in the feature space then becomes or

  12. In many cases is difficult to map data instances to a feature space and calculate distances • High Dimentionality of the feature space (memory considerations) • Explicit map might be difficult to determine • We can define a kernel function to compute these dot products in the feature space (Hilbert) Then we could get distances by using Kernel functions

  13. Radial Basis Kernel Function Defined over input spaces which are vector spaces • Using Convolution kernels we can then use arbitrary input spaces. • The author suggests the use of convolution kernels to avoid converting audit data into a vector in

  14. Detecting Outliers • Detecting points that are distant from other points or in relatively sparse regions of the feature space

  15. Cluster-based Estimation • Count the number of points within a sphere of radius w around the point • Sort clusters based on size • The points in the small clusters are labeled anomalous

  16. Cluster-based Estimation • Any points x1,x2 are considered near if their distance is less than or equal to • Define N(x) to be the number of points that are within w of point x • Since we have to compute the pairwise distance among points the computation of N(x) for all points has complexity • We are interested in the outliers

  17. To reduce computation, an approximation can be done via fixed width clustering • The first point is the center of the first cluster • For every subsequent point, if it is within w of a cluster center, it is added to that cluster • Otherwise it becomes the center of a new cluster • Points may be added to several clusters • Complexity with c number of clusters and n number of data points • A threshold on n is used to find outliers

  18. K-Nearest Neighbor • Find points that lie in a sparse region of the feature space by computing the distances to the k-nearest neighbors of the point • Dense regions will have many points near them and will have a small k-NN score • If k exceeds the frequency of any given attack and the images of the attack elements are far from the images of the normal elements, then the k-NN score can be used to detect attacks

  19. K-NN is computationally expensive • Since we’re interested in only the k-nearest points to a given point we can reduce the computational cost by using canopy clustering • Canopy Clustering is used to reduce the space into smaller subsets avoiding the need to check every data point

  20. Modified Canopy Clustering • Cluster data with fixed-width approach with the variation of placing each element in only one cluster • For each two points x1,x2 in a cluster • And in all cases

  21. Let C be the set of clusters (initially containing all clusters in the data) • At any step, we have a set of points which are potentially among the k-nearest neighbor points. This set is denoted as P. • We also have a set of points that are in fact among the k-nearest points. This set is denotes as K. • Initially K and P are empty

  22. Pre-compute the distance from x to each cluster. • For the cluster with center closest to x we remove it from C and add all its points to P. Called Opening the Cluster • We can use the lower bound on distance given by • For each point xi in P we compare distances to other points in P • If this distance is <dminwe can guarantee that xi is closer to point x than all the points in the clusters in C

  23. In this case we remove xi from P and add it to K • If distance is >dminthen we open the closest cluster and add all the points to P and remove that cluster from C • Every time we remove a cluster from Cdmin will increase • Once K has k elements we terminate

  24. Computation is spent checking distance between points in D to the cluster centers, which is more efficient than computing pairwise distances among all points • Choice of w effects only the efficiency, not the K-NN score • Intuitively we want to choose a w that splits the data into reasonably sized clusters

  25. One Class SVM • Map feature space into a second feature space with a radial basis kernel Standard SVM requires supervised learning algorithms (it requires labeled data)

  26. A newly modified SVM was adapted to unsupervised learning algorithm • Attempts to separate the entire set of data from the origin with maximal margin • Classes will be labeled as +1 and -1

  27. The hyperplane is estimated by the hyperplane’s normal vector in the feature space w and offset from the origin Decision function

  28. Optimization is solved with a variant of Sequential Minimal Optimization

  29. Feature Space • Data Sets • Network Records with 41 features and 4,900,00 instances (KDD Cup 1999 Data) • System Call Traces (process) from 5 weeks from the Basic Security Module of the MIT Lincoln Labs IDS Evaluation created on 1999

  30. Experimental Results

  31. ROC Curves

  32. Questions

More Related