Presentation is loading. Please wait.

Presentation is loading. Please wait.

Learning Rules for Anomaly Detection of Hostile Network Traffic Matthew V. Mahoney and Philip K. Chan Florida Institute of Technology.

Similar presentations


Presentation on theme: "Learning Rules for Anomaly Detection of Hostile Network Traffic Matthew V. Mahoney and Philip K. Chan Florida Institute of Technology."— Presentation transcript:

1 Learning Rules for Anomaly Detection of Hostile Network Traffic Matthew V. Mahoney and Philip K. Chan Florida Institute of Technology

2 Problem: How to detect novel intrusions in network traffic given only a model of normal traffic  Normal web server request GET /index.html HTTP/1.0 GET /index.html HTTP/1.0  Code Red II worm GET /default.ida?NNNNNNNNN… GET /default.ida?NNNNNNNNN…

3 What has been done  Firewalls Can’t block attacks on open ports (web, mail, DNS) Can’t block attacks on open ports (web, mail, DNS)  Signature Detection (SNORT, BRO) Hand coded rules (search for “default.ida?NNN”) Hand coded rules (search for “default.ida?NNN”) Can’t detect new attacks Can’t detect new attacks  Anomaly Detection (eBayes, ADAM, SPADE) Learn rules from normal traffic for low-level protocols (IP, TCP, ICMP) Learn rules from normal traffic for low-level protocols (IP, TCP, ICMP) But application protocols (HTTP, mail) are too hard to model But application protocols (HTTP, mail) are too hard to model

4 Learning Rules for Anomaly Detection (LERAD)  Associative mining (APRIORI, etc.) learns rules with high support and confidence for one value  LERAD learns rules with high support (n) and a small set of allowed values (r)  Any value seen at least once in training is allowed If port = 80 and word1 = “GET” then word3  {“HTTP/1.0”, “HTTP/1.1”} (r = 2)

5 LERAD Steps 1. Generate candidate rules 2. Remove redundant rules 3. Remove poorly trained rules LERAD is fast because steps 1-2 can be done on a small random sample (~100 tuples)

6 Step 1. Generate Candidate Rules Suggested by matching attribute values SamplePortWord1Word2Word3 S180GET/index.htmlHTTP/1.0 S280GET/banner.gifHTTP/1.0 S325HELOpascalMAIL  S1 and S2 suggest: port = 80 if port = 80 then word1 = “GET” if word3 = “HTTP/1.0” and word1 = “GET then port = 80  S2 and S3 suggest no rules

7 Step 2. Remove Redundant Rules Favor rules with higher score = n/r SamplePortWord1Word2Word3 S180GET/index.htmlHTTP/1.0 S280GET/banner.gifHTTP/1.0 S325HELOpascalMAIL Rule 1: if port = 80 then word1 = “GET” (n/r = 2/1) Rule 2: if word2 = “/index.html” then word1 = “GET” (n/r = 1/1) Rule 2 has lower score and covers no new values, so it is redundant

8 Step 3. Remove Poorly Trained Rules Rules with violations in a validation set will probably generate false alarms Train Validate Test r (number of allowed values) Fully trained rule (kept) Incompletely trained rule (removed)

9 Attribute Sets  Inbound client packets (PKT) IP packet cut into 24 16-bit fields IP packet cut into 24 16-bit fields  Inbound client TCP streams Date, time Source, destination IP addresses and ports Length, duration TCP flags First 8 application words Anomaly score = tn/r summed over violated rules, t = time since previous violation

10 Experimental Evaluation  1999 DARPA/Lincoln Laboratory Intrusion Detection Evaluation (IDEVAL) Train on week 3 (no attacks) Train on week 3 (no attacks) Test on inside sniffer weeks 4-5 (148 simulated probes, DOS, and R2L attacks) Test on inside sniffer weeks 4-5 (148 simulated probes, DOS, and R2L attacks) Top participants in 1999 detected 40-55% of attacks at 10 false alarms per day Top participants in 1999 detected 40-55% of attacks at 10 false alarms per day  2002 university departmental server traffic (UNIV) 623 hours over 10 weeks 623 hours over 10 weeks Train and test on adjacent weeks (some unlabeled attacks in training data) Train and test on adjacent weeks (some unlabeled attacks in training data) 6 known real attacks (some multiple instances) 6 known real attacks (some multiple instances)

11 Experimental Results Percent of attacks detected at 10 false alarms per day

12 UNIV Detection/False Alarm Tradeoff Percent of attacks detected at 0 to 40 false alarms per day

13 Run Time Performance (750 MHz PC – Windows Me)  Preprocess 9 GB IDEVAL traffic = 7 min.  Train + test < 2 min. (all systems)

14 Anomalies are due to bugs and idiosyncrasies in hostile code No obvious way to distinguish from benign events UNIV attack How detected Inside port scan HEAD / HTTP\1.0 (backslash) Code Red II worm TCP segmentation after GET Nimda worm host: www Scalper worm host: unknown Proxy scan host: www.yahoo.com DNS version probe (not detected)

15 Contributions  LERAD differs from association mining in that the goal is to find rules for anomaly detection: a small set of allowed values  LERAD is fast because rules are generated from a small sample  Testing is fast (50-75 rules)  LERAD improves intrusion detection Models application protocols Models application protocols Detects more attacks Detects more attacks

16 Thank you


Download ppt "Learning Rules for Anomaly Detection of Hostile Network Traffic Matthew V. Mahoney and Philip K. Chan Florida Institute of Technology."

Similar presentations


Ads by Google