1 / 57

Correlating Network Attacks Using Bayesian Multiple Hypothesis Tracking

Correlating Network Attacks Using Bayesian Multiple Hypothesis Tracking. Daniel J. Burroughs Institute for Security Technology Studies Thayer School of Engineering Dartmouth College. May 1, 2002. Outline. Institute for Security Technology Studies Needs and goals System overview

miya
Download Presentation

Correlating Network Attacks Using Bayesian Multiple Hypothesis Tracking

An Image/Link below is provided (as is) to download presentation Download Policy: Content on the Website is provided to you AS IS for your information and personal use and may not be sold / licensed / shared on other websites without getting consent from its author. Content is provided to you AS IS for your information and personal use only. Download presentation by click this link. While downloading, if for some reason you are not able to download a presentation, the publisher may have deleted the file from their server. During download, if you can't get a presentation, the file might be deleted by the publisher.

E N D

Presentation Transcript


  1. Correlating Network Attacks Using Bayesian Multiple Hypothesis Tracking Daniel J. Burroughs Institute for Security Technology StudiesThayer School of EngineeringDartmouth College May 1, 2002

  2. Outline • Institute for Security Technology Studies • Needs and goals • System overview • Sensor Modeling • Attacker Modeling • Hypothesis Management • Testing and Evaluation • Summary and Future Work

  3. Institute for Security Technology Studies • Security and counter-terrorism research center • Funded by the NIJ • Main focus is on computer security • Investigative Research for Infrastructure Assurance (IRIA) • Joint effort with Thayer School of Engineering

  4. Outline • Institute for Security Technology Studies • Needs and goals • System overview • Sensor Modeling • Attacker Modeling • Hypothesis Management • Testing and Evaluation • Summary and Future Work

  5. The Internet and Security in a Nutshell IDS IDS ALERT! ALERT!

  6. What is the Need? • Distributed and/or coordinated attacks • Increasing rate and sophistication • Infrastructure protection • Coordinated attack against infrastructure • Attacks against multiple infrastructure components • Overwhelming amounts of data • Huge effort required to analyze • Lots of uninteresting events

  7. Outline • Institute for Security Technology Studies • Needs and goals • System overview • Sensor Modeling • Attacker Modeling • Hypothesis Management • Testing and Evaluation • Summary and Future Work

  8. What is the System? SHADOW RealSecure Tracking System Snort Security Database • Reorganization of existing data • Data fusion • Building situational knowledge • Not an intrusion detection system

  9. Network Centered View • Network viewed in isolation • Limited view of attacker’s activity • Defensive posture

  10. Distributed Attack Denial of Service

  11. Attacker Centered View • More complete picture • Information gathering • Requires cooperation and data fusion

  12. Radar Tracking RealSecure Snort SHADOW • Multiple sensors • Multiple targets • Heterogeneous sensors • Real-time tracking • Incomplete data • Inaccurate data

  13. Gather and Correlate • Collecting data • Time correlation, communications, common formatting, etc. • These issues are addressed by numerous projects • IDEF, IDMEF, CIDF, D-Shield, Incidents.org, etc. • Correlating data • How can we tell what events are related? • Attacker’s goals determine behavior • Multiple hypothesis tracking

  14. Multiple Hypothesis Tracking PortScan BufferOverflow PortScan PortScan Attack 1: Attack 1: • Scenario created Stream 1 • Alternate hypothesis BufferOverflow Attack 2: • Events analyzed on arrival BufferOverflow OR

  15. Hypothesis Evaluation • Hypotheses are evaluated based on the behaviors of the sensor and target • What real-world event caused the given sensor output? • How likely is it that the target moved to this position?

  16. Outline • Institute for Security Technology Studies • Needs and goals • System overview • Sensor Modeling • Attacker Modeling • Hypothesis Management • Testing and Evaluation • Summary and Future Work

  17. IDS Overview • Two methods of intrusion detection • Signature detection (pattern matching) • Low false positive / Detects only known attacks • Statistical anomaly detection • High false positive / Detects wider range of attacks • Two domains to be observed • Network • Host

  18. Signature Detection vs. Anomaly Detection • Modeling signature detection is easy • If a known attack occurred in an observable area, then p(detection) = 1, else p(detection) = 0 • Modeling anomaly detection is more difficult • Noisy and/or unusual attacks are more likely seen • Denial of Service, port scans, unused services, etc. • Other types of attacks may be missed • Malformed web requests, some buffer overflows, etc.

  19. Event Measurements • Minimal feature set is extracted from reports • Source IP, destination IP • Source port, destination port • Type of attack • Time • These are then used to describe a hyperspace through which the attack moves

  20. Bayesian Inference • Forward response of sensor is well known • Given real-world event x, what is H(x)? • We need to reason backwards • Given sensor output H(x), what is x? • Forward response and prior distribution of x • Probability of H(x) given x • Probability of a particular x existing

  21. Outline • Institute for Security Technology Studies • Needs and goals • System overview • Sensor Modeling • Attacker Modeling • Hypothesis Management • Testing and Evaluation • Summary and Future Work

  22. Attacker Model • Attackers are not as easy to observe • Often we are only able to observe them through the sensors (IDS) • State of the attack is difficult to describe • We have three sources of attack data • Simulation • Dartmouth / Thayer network • Def Con

  23. Simulation • Purely generated data • Models for generating attack sequences and noise • Highly controllable – good for development • Generated attacks with ‘background noise’ • Use Thayer IDS for background noise • More interesting for testing

  24. Dartmouth / Thayer Network ISTS Snort SignalQuest SHADOW Switch Snort Snort SHADOW Switch Switch Switch

  25. Def-Con Capture-The-Flag • Hacker game • Unrealistic data in some aspects • Lack of stealth, lack of firewall, etc. • Many attacks, many scenarios • 16,250 events in 2.5 hours • 89 individual scenarios • Classified by Oliver Dain at Lincoln Labs

  26. State Problem • Desire to describe state as Markovian process • Reduces computational complexity and space • Easy for an aircraft, difficult for an attack • Non-linear, non-contiguous space ? X, Y, Z Yaw, Pitch, Roll Position & Velocity

  27. State Problem • No simple method for describing state • Use a history of events in the track • Increases computational complexity • Increases memory requirements • Use a weighted window of past events • Calculate various relationships between past and current events.

  28. Windowed History • Minimum history needed to differentiate state • Weighting of events to lend more value to recent events • Relationships calculated between pairs and sequences of events Xt-6 Xt-5 Xt-4 Xt-3 Xt-2 Xt-1 Xt

  29. Common History 1a • Don’t care which path was taken • Just need to distinguish current state State1 2a 1b 2b State2 1c 2c

  30. Predictive Model • To determine likelihood of event belonging in series, predictive models are needed • Based on current state, what is the probability distribution for the target motion? • Different types of attacks have different distributions

  31. Attacker Motion Probability Distributions Motion update for scanning Motion update for DoS(Denial of Service) Events are readily distinguishable based on arrival time and source IP distance

  32. Feature Extraction • Historical data sets used to determine good differentiating feature sets • These are used in combination to measure the fitness of new events to scenarios • Use neural net to discover complex patterns

  33. Neural Net • Empirically derived probability distributions work well for simple attacks • But is difficult to compute for more complex ones • Machine learning is applied to solve this • Neural net feeds from event feature set values • Fitness function is calculated from this

  34. Neural Net • Fitness functions created for various feature subsets • i.e., rate of events vs. IP source velocity • These values feed a neural net • NN then determines overall fitness value

  35. Outline • Institute for Security Technology Studies • Needs and goals • System overview • Sensor Modeling • Attacker Modeling • Hypothesis Management • Testing and Evaluation • Summary and Future Work

  36. Hypothesis Management • In the brute-force approach, each new event doubles the number of hypotheses • Without pruning, complexity grows exponentially

  37. Branch and Prune • Calculate all possible hypotheses • Prune back unlikely or completed ones • Must be very aggressive in pruning • Many hypotheses are not kept long • Inefficient method of controlling growth

  38. Selective Branching • Often times, there is a clear winner • Why bother creating hypotheses for other? • Measure difference between fitness of top choice and fitness of second choice • If it is greater than a predetermined threshold, no branching is needed • Number of branches can be determined with threshold

  39. Preprocessing and Multi-pass • Some sequences of events are simply related • Port scans • Noisy • Many events • Require many evaluations • Easily grouped • Preprocessing groups these into single larger events

  40. Multi-Pass Approach a b c d f g h k l m • Develop small attack sequences initially • Chain sequences together in later passes • Small sequences become atomic events • May aid ‘missing data’ problem a-b-c-d f-g-h k-l-m a-b-c-d-f-g-h-k-l-m

  41. Outline • Institute for Security Technology Studies • Needs and goals • System overview • Sensor Modeling • Attacker Modeling • Hypothesis Management • Testing and Evaluation • Summary and Future Work

  42. Testing and Evaluation • Testing has been performed with data collected from the Thayer network and DefCon data sets • Thayer testing used earlier probability distribution method • DefCon testing used machine learning approach • Arranging for a live run at DefCon

  43. Thayer Testing and Evaluation • Testing performed on Thayer data • Roughly 1500 events • 20 Scenarios • Roughly half of data were single events

  44. Thayer Testing and Evaluation • Accuracy measured by number of correctly placed scenario events • Best hypothesis had ~20% of the single events included in tracks • Most confident hypothesis not always most accurate

  45. DefCon Testing and Evaluation • Testing performed on DefCon data • 2.5 Hour time slice • Roughly 16,000 events • 89 Scenarios • Hand classified by Oliver Dain at Lincoln Labs • Neural net approach used • Trained with random time slice of data

  46. DefCon Testing and Evaluation • Testing performed on DefCon data • 2.5 Hour time slice • Roughly 16,000 events • 89 Scenarios • Hand classified by Oliver Dain at Lincoln Labs • Neural net approach used • Trained with random time slice of data

  47. DefCon Testing and Evaluation From Dain & Cunningham (October, 2001)

  48. DefCon Testing and Evaluation • Accuracy measured by number of correctly placed scenario events • Achieved higher accuracy, but less stable with fewer hypotheses

  49. Outline • Institute for Security Technology Studies • Needs and goals • System overview • Sensor Modeling • Attacker Modeling • Hypothesis Management • Testing and Evaluation • Summary and Future Work

  50. Summary • Reorganize data already being collected • Provide ‘Higher level’ view of situation • Reduce the work of the security analyst • Radar tracking analogy • Multisensor data fusion • Multiple hypothesis tracking

More Related