Propagation and Containment - PowerPoint PPT Presentation

1 / 72
About This Presentation
Title:

Propagation and Containment

Description:

It would take 40 years to infect 50% of vulnerable population with random scanning ... Scan-based & self-propagation. Email. Windows File Sharing. Hybrid ... – PowerPoint PPT presentation

Number of Views:32
Avg rating:3.0/5.0
Slides: 73
Provided by: awo3
Category:

less

Transcript and Presenter's Notes

Title: Propagation and Containment


1
Propagation and Containment
  • Presented by Jing Yang, Leonid Bolotnyy, and
    Anthony Wood

2
Analogy between Biological and Computational
Mechanisms
  • The spread of self-replicating program within
    computer systems is just like the transition of
    smallpox several centuries ago 1
  • Mathematical models which are popular in the
    research of biological epidemiology can also be
    used in the research of computer viruses 1
  • Kephart Whites work first time to explicitly
    develop and analyze quantitative models which
    capture the spreading characteristics of computer
    viruses

3
Kephart Whites Work
  • Based on the assumption that viruses are spread
    by program sharing
  • Benefits of using mathematically models mentioned
    in their paper
  • Evaluation and development of general polices and
    heuristics for inhibiting the spread of viruses
  • Apply to a particular epidemic, such as
    predicting the course of a particular epidemic

4
Modeling Viral Epidemics on Directed Graphs
  • Directed Graph 1
  • Representing an individual system as a node in a
    graph
  • Directed edges from a given node j to other nodes
    represent the set of individuals that can be
    infected by j
  • A rate of infection is associated with each edge
  • A rate at which infection can be detected and
    cured is associated with each node

5
SIS Model on A Random Graph
  • Random Graph a directed graph constructed by
    making random, independent decisions about
    whether to include each of the N(N-1) possible
    directional edges 1
  • Techniques used by Kephart White
  • Deterministic approximation
  • Approximate probabilistic analysis
  • Simulation

6
Deterministic Approximation
  • ß infection rate along each edge
  • d cure rate for each node
  • ß ß p (N - 1) average total rate at which a
    node attempts to infect its neighbors
  • ? d / ß if ? gt 1, the fraction of infected
    individuals decays exponentially from the initial
    value to 0, i.e. there is no epidemic if ? lt 1,
    the fraction of infected individuals grows from
    the initial value at a rate which is initially
    exponential and eventually saturates at the value
    1 - ?

7
Probabilistic Analysis
  • Including new information such as
  • Size of fluctuations in the number of infected
    individuals
  • Possibility that fluctuations will result in
    extinction of the infection
  • Conclusion
  • A lot of variance in the number of infected
    individuals from one simulation to another
  • In equilibrium the size of the infected
    population is completely insensitive to the
    moment at which the exponential rise occurred
  • Extinction probability and metastable
    distribution can be calculated

8
Simulations
  • Results
  • Higher extinction probability
  • Lower average number of infected individuals
  • Suspected reason
  • No details of which nodes are infected
  • No variation in the number of nodes that a given
    node could infect

9
Simulations (cont.)
  • Scenario
  • A random graph in which most nodes are isolated
    and a few are joined in small clusters
  • Anything contributes to containment?
  • Build isolated cells worms can spread unimpeded
    in a cell, but containment system will limit
    further infection between cells

10
Improvements of SIS Model on A Random Graph
  • Kephart Whites work
  • Weak links give a node a small but finite
    chance of infecting any node which is not
    explicitly connected to it
  • Hierarchical model extend a two-type model of
    strong and weak link to a hierarchy
  • Wangs work
  • Effects of infection delay
  • Effects of user vigilance

11
Does SIS SIR Models Take Containment into
Consideration?
  • No to SIS and maybe Yes to SIR
  • In SIS, it may be more appropriate to be called
    treatment
  • In SIR, deployment of containment is limited to
    the individual node, which means that every cell
    only contains one node more appropriate to be
    called treatment
  • Not to be automatic
  • No cooperation has been applied

12
Model without Containment
  • Remember the assumption by Kephart Whites
    work?
  • Viruses spread by program sharing
  • Modern worms spread so quickly that manual
    containment is impossible
  • A model without containment should be built first
    (SI model) and then different containment methods
    are added to test the results

13
IPv6 vs. Worms
  • It will be very challenging to build Internet
    containment systems that prevent widespread
    infection from worm epidemics in the IPv4
    environment 2
  • It seems that the only effective defense is to
    increase worm scanning space
  • Upgrading IPv4 to IPv6

14
How Large is IPv6 Address
  • IPv6 has 2128 IP addresses 3
  • Smallest subnet has 264 addresses 3
  • 4.4 billion IPv4 internets
  • Consider a sub-network 3
  • 1,000,000 vulnerable hosts
  • 100,000 scans per second (Slammer - 4,000)
  • 1,000 initially infected hosts
  • It would take 40 years to infect 50 of
    vulnerable population with random scanning

15
Worm Classification
  • Spreading Media 3
  • Scan-based self-propagation
  • Email
  • Windows File Sharing
  • Hybrid
  • Target Acquisition 3
  • Random Scanning
  • Subnet Scanning
  • Routing Worm
  • Pre-generated Hit List
  • Topological
  • Stealth / Passive

16
Can IPv6 Really Defeats Worms?
  • Traditional scan-based worms seems to be
    ineffective, but there may be some way to improve
    the scan methods 4
  • More sophisticated hybrid worms may appear, which
    use a variety of ways to collect addresses for a
    quick propagation
  • Polymorphic worms may significantly increase the
    time to extract the signature of a worm

17
Improvement in Scan Methods
  • Subnet Scanning
  • The first goal may be a /64 enterprise network
    instead of the whole internet
  • Routing Worm
  • Some IP addresses are not allocated
  • Pre-generated Hit List Scanning
  • Speedup the propagation and the whole address
    pace can be equally divided for each zombie

18
Improvement in Scan Methods (cont.)
  • Permutation Scanning
  • Avoid waste of scanning one host many times
  • Topological Scanning
  • Use the information stored on compromised hosts
    to find new targets
  • Stealth / Passive Worm
  • Waiting for the vulnerable host to contact you
    may be more efficient to scan such a large
    address space

19
What Can IPv6 Itself Contribute?
  • Public services need to be reachable by DNS
  • At least we have some known addresses in advance
    4
  • DNS name for every host because of the long IPv6
    address
  • DNS Server under attack can yield large caches of
    hosts 4
  • Standard method of deriving the EUI field
  • The lower 64 bits of IPv6 is derived from the
    48-bit MAC address 5
  • IPv6 neighbor-discovery cache data
  • One compromised host can reveal the address of
    other hosts 4
  • Easy-to-remember host address in the transition
    from IPv4 to IPv6
  • To scan the IPv6 address space is no difference
    to scan the IPv4 address space 4

20
More Sophisticated Hybrid Worms
  • Humans are involved
  • Different methods are used to compromise a host,
    so the vulnerability density increases relatively
  • Nimdas success

21
Polymorphic Worms
  • Very effective containment system extracts a
    worms feature to do content filtering
  • Even though some methods exist to detect
    polymorphic worms, the successful rate may not be
    100

22
What Should We Do Then?
  • To find out whether the following methods which
    just speedup the worms propagation in IPv4 can
    just make worms quick propagation in IPv6
    possible
  • Improvement in scan methods IPv6 inherent
    features
  • More sophisticated hybrid worms
  • Polymorphic worms

23
Futures Work
  • Use traditional models to see whether each method
    or a combination of them can make a quick
    propagation of worm in IPv6 possible
  • Add new features of worm spread in IPv6 to build
    new models, which can represent the reality more
    precisely
  • If quick propagation can be true in IPv6,
    relative containment methods should be figured
    out it should be much more possible than in
    IPv4

24
References
  • 1 Jeffrey O. Kephart, Steve R. White.
    Directed-Graph Epidemiological Models of Computer
    Viruses
  • 2 David Moore et al. Internet Quarantine
    Requirements for Containing Self-Propagating Code
  • 3 Mark Shaneck. Worms Taxonomy and Detection.
  • 4 Sean Convery et al. IPv6 and IPv4 Threat
    Comparison and Best Practice Evaluation.
  • 5 Michael H. Warfield et al. Security
    Implications of IPv6.

25
Propagation and Containment of Worms
  • General modeling of worm propagation and
    containment strategies

26
Ways to mitigate the threat of worms
  • Prevention
  • Prevent the worm from spreading by reducing the
    size of vulnerable hosts
  • Treatment
  • Neutralize the worm by removing the vulnerability
    it is trying to exploit
  • Containment
  • Prevent the worm from spreading from infected
    systems to the unaffected, but vulnerable hosts

27
Containment Approaches
  • La Brea
  • Intercepts probes to unallocated addresses
  • Connection-history based anomaly detection
  • Analyzes connection traffic trying to detect an
    anomaly
  • Per host throttling
  • Restricts connection rate to new hosts
  • Blocking access to affected ports
  • Prevents affected hosts from accessing vulnerable
    ports on other machines
  • NBAR
  • Filters packets based on the content using worm
    signatures
  • It was very effective in preventing the spread of
    Code-Red

28
General model for worm infection rate
29
Modeling Containment Systems
  • Reaction Time
  • Time required to detect the infection, spread the
    information to all the hosts participating in the
    system, and to activate containment mechanisms
  • Containment Strategy
  • Strategy that isolates the worm from uninfected
    susceptible systems (e.g. address blacklisting
    and content filtering)
  • Deployment Scenario
  • Who, Where and How of the containment
    strategy implementation

30
Simulation parameters
  • Population 232 (assuming IPv4)
  • Number of vulnerable hosts 360,000 (same as for
    Code-Red v2)
  • Any probe to the susceptible host results in an
    infection
  • A probe to the infected or non-vulnerable host
    has no effect
  • The first host is infected at time 0
  • If a host is infected in time t, then all
    susceptible hosts are notified at time t R
    where R is the reaction time of the system
  • Simulation is run 100 times

31
Simulation Goals
  • Determine the reaction time needed to limit the
    worm propagation for address-blacklisting and
    content filtering
  • Compare the two containment strategies
  • Realize the relationship between reaction time
    and worm probe rate

32
Idealized Deployment Simulationfor Code-Red
33
Idealized Deployment Simulationfor Code-Red
Conclusions
  • The strategy is effective if under 1 of
    susceptible hosts are infected within the 24 hour
    period with 95 certainty
  • Address-blacklisting is effective if reaction
    time is less than 20 minutes
  • Note if reaction time gt 20 minutes, all
    susceptible hosts will eventually become infected
  • Content filtering is effective if reaction time
    is less than two hours
  • How many susceptible hosts will become infected
    after time R (reaction time)?

34
Idealized Deployment Simulationfor General Worm
(1)
  • The authors generalize the definition of the
    effectiveness as the reaction time required to
    contain the worm to a given degree of global
    infection
  • Worm aggressiveness rate at which infected host
    probes others to propagate
  • Note The rate of host probes does not take into
    account the possibility of preferential status
    that some addresses may have

35
Idealized Deployment Simulationfor General Worm
(2)
36
Idealized Deployment Simulationfor General Worm
conclusions
  • Worms that are more aggressive than Code-Red
    having higher probe rate of 100 probes/second
    require a reaction time of under three minutes
    using Address-blacklisting and under 18 minutes
    for Content Filtering to contain the worm to 10
    of total susceptible population.

37
Practical Deployment
  • Analyzing practical deployment, authors
    concentrate on content filtering because of
    seemingly much lower requirements on the reaction
    time compared to address-blacklisting. This may
    be premature because the technique is still
    useful. It would be very beneficial to see a
    hybrid containment strategy that uses both
    content filtering and address-blacklisting.

38
Practical Deployment simulation parameters
  • The topology of the Internet is taken at the time
    of the spread of Code-Red v2.
  • The number of vulnerable hosts is 338,652 (some
    hosts map to multiple autonomous systems they
    have been removed only infected hosts in the
    first 24 hours are inc.)
  • The number of autonomous systems is 6,378
  • The packet is assumed to travel along the
    shortest path through autonomous systems

39
Practical Deployment for Code-Red
  • Reaction time is two hours (less than 1 infected
    in idealized simulation)

40
Practical Deployment for Code-Red conclusions
  • ISP deployment is more effective by itself than
    the Customer deployment
  • 40 top ISPs can limit the infection to under 5
    whereas top 75 of Customer Autonomous systems
    can only limit infection to 25
  • The results could have been anticipated based on
    the role of ISPs (their topology)

41
Practical Deployment for Generalized Worm
  • We investigate reaction time requirements

42
Practical Deployment for Generalized Worm
conclusions
  • For probe rate of 100 probes/second or larger,
    neither deployment can effectively contain the
    worm
  • In the best case, effective containment is
    possible for 30 or fewer probes by the TOP 100
    ISPs and only 2 or fewer probes by the 50
    Customers
  • Note TOP 100 ISPs cannot prevent a worm from
    infecting less than 18 of the hosts if the probe
    rate is 100 probes/second (not on the graph)

43
Conclusions of the modeling scheme (1)
  • Automated means are needed to detect the worm and
    contain it.
  • Content filtering is more effective than
    address-blacklisting, but a combination of
    several strategies may need to be employed.
  • The reaction time has to be very small, on the
    order of minutes to be able to combat aggressive
    worms.
  • It is important to deploy the containment
    filtering strategy at most top ISPs.

44
Conclusions of the modeling scheme (2)
  • The parameters of the model have changed
  • Other containment strategies need to be
    considered
  • What will happen to the population parameter?
  • What may happen to beta soon?
  • Combination of prevention, treatment and
    containment strategies are needed to combat
    aggressive worms

45
LaBrea (1)
  • LaBrea is a Linux-based application which works
    at the network application layer creating virtual
    machines for nonexistent IP addresses when a
    packet to such an address reaches the network.
  • Once the connection is established, LaBrea will
    try to hold the connection as long as possible
    (by moving connections from established state to
    persistent state, it can hold connections almost
    indefinitely).

46
LaBrea (2)
  • Any connection to LaBrea is suspect because the
    IP address to which the packet is sent does not
    exist, not even in DNS.
  • It can also analyze the range of IP addresses
    that are requested giving it a broader view of a
    potential attack (all ports on virtual machines
    appear open).
  • It requires 8bps to hold 3 threads of Code-Red.
    If there were 300,00 infected machines each with
    100 threads, 1,000 sites would require 5.2 of
    the full T1 line bandwidth each to hold them.

47
Connection-history based anomaly detection
  • The idea is to use GriDS based intrusion
    detection approach and make some modifications to
    it to allow for worm containment
  • Goals
  • Automatic worm propagation determination
  • Worm detection with a low false positive rate
  • Effective countermeasures
  • Automatic responses in real-time
  • Prevent infected hosts from infecting other hosts
  • Prevent non-infected hosts from being infected

48
Connection-history based anomaly detection model
  • Monitoring station collects all recent connection
    attempts data and tries to find anomaly in it.
  • Patterns of a worm
  • Similarity of connection patterns
  • The worm tries to exploit the same vulnerability
  • Causality of connection patterns
  • When one event follows after another
  • Obsolete connections
  • Compromised hosts try to access services at
    random IPs

49
Very Fast Containment of Scanning Worms
  • Weaver, Staniford, Paxson

50
Outline
  • Scanning
  • Suppression Algorithm
  • Cooperation
  • Attacks
  • Conclusion

51
What is Scanning?
  • Probes from adjacent remote addresses?
  • Dist. probes that cover local addresses?
  • Horizontal vs. Vertical
  • Factor in connection rates?
  • Temporal and spatial interdependence
  • How to infer intent?

52
Scanning Worms
  • Blaster, Code Red, CR II, Nimda, Slammer
  • Does not apply to
  • Hit lists (flash worms)
  • Meta-servers (online list)
  • Topology detectors
  • Contagion worms

53
Scanning Detection
  • Key properties of scans
  • Most scanning fails
  • Infected machines attempt many connections
  • Containment is based on worm behavior, not
    signatures (content)
  • Containment by address blocking (blacklisting)
  • Blocking can lead to DoS if false positive rate
    is high

54
Scan Suppression
  • Goal 1 protect the enterprise forget the
    Internet
  • Goal 2 keep worm below epidemic threshold, or
    slow it down so humans notice
  • Divide enterprise network into cells
  • Each is guarded by a filter employing the scan
    detection algorithm

55
Inside, Outside, Upside Down
  • Preventing scans from Internet is too hard
  • If inside node is infected, filter sees all
    traffic
  • Cell (LAN) is outside, Enterprise network is
    inside
  • Can also treat entire enterprise as cell,
    Internet as outside

Inside
Scan detectors
56
Scan Suppression
  • Assumption benign traffic has a higher
    probability of success than attack traffic
  • Strategy
  • Count connection establishment messages in each
    direction
  • Block when misses hits gt threshold
  • Allow messages for existing connections, to
    reduce impact of false positives

57
Constraints
  • For line-speed hardware operation, must be
    efficient
  • Memory access speed
  • On duplex gigabit ethernet, can only access DRAM
    4 times
  • Memory size
  • Attempt to keep footprint under 16MB
  • Algorithm complexity
  • Want to implement entirely in hardware

58
Mechanisms
  • Approximate caches
  • Fixed memory available
  • Allow collisions to cause aliasing
  • Err on the side of false negative
  • Cryptographic hashes
  • Prevent attackers from controlling collisions
  • Encrypt hash input to give tag
  • For associative cache, split and save only part
    as tag in table

59
Connection Cache
  • Remember if weve seen a packet in each direction
  • Aliasing turns failed attempt into success
    (biases to false negative)
  • Age is reset on each forwarded packet
  • Every minute, bg process purges entries older
    than Dconn

60
Address Cache
  • Track outside addresses
  • Counter keeps difference between successes and
    failures
  • Counts are decremented every Dmiss seconds

61
Algorithm Pseudo-code
62
In
Out
Connection cache Address cache
A,X OutIn -
A, OutIn -
A,Z OutIn -
A,Y OutIn -
B
A
A,B OutIn -
A,B OutIn InOut
A 1
A 2
A 3
A 1
A T
A Cmax
  • UDP Probe
  • A ? X fwd
  • A ? Y fwd
  • Normal Traffic
  • A ? B fwd
  • B ? A fwd, bidir
  • Scanning again
  • A ? fwd until T
  • A ? Z blocked
  • A ? B ?
  • block SYN/UDP, fwd TCP

63
Performance
  • For 6000-host enterprise trace
  • 1MB connection cache, 4MB 4-way address cache
    5MB total
  • At most 4 memory accesses per packet
  • Operated at gigabit line-speed
  • Detects scanning at rates over 1 per minute
  • Low false positive rate
  • About 20 false negative rate
  • Detects scanning after 10-30 attempts

64
Scan Suppression Tuning
  • Parameters
  • T miss-hit difference that causes block
  • Cmin minimum allowed count
  • Cmax maximum allowed count
  • Dmiss decay rate for misses
  • Dconn decay rate for idle connections
  • Cache size and associativity

65
Cooperation
  • Divide enterprise into small cells
  • Connect all cells via low-latency channel
  • A cells detector notifies others when it blocks
    an address (kill message)
  • Blocking threshold dynamically adapts to number
    of blocks in enterprise
  • T T ?X, for very small ?
  • Changing ? does not change epidemic threshold,
    but reduces infection density

66
Cooperation Effect of ?
67
Cooperation Issues
  • Poor choice of ? could cause collapse
  • Lower thresholds increase false positives
  • Should a complete shutdown be possible?
  • How to connect cells (practically)?

68
Attacking Containment
  • False positives
  • Unidirectional control flows
  • Spoofing outside addresses (though this does not
    prevent inside systems from initiating
    connections)
  • False negatives
  • Use a non-scanning technique
  • Scan under detection threshold
  • Use a whitelisted port to test for liveness
    before scanning

69
Attacking Containment
  • Detecting containment
  • Try to contact already infected hosts
  • Go stealthy if containment is detected
  • Circumventing containment
  • Embed scan in storm of spoofed packets
  • Two-sided evasion
  • Inside and outside host initiate normal
    connections to counter penalty of scanning
  • Can modify algorithm to prevent, but lose
    vertical scan detection

70
Attacking Cooperation
  • Attempt to outrace containment if threshold is
    permissive
  • Flood cooperation channels
  • Cooperative collapse
  • False positives cause lowered thresholds
  • Lowered thresholds cause more false positives
  • Feedback causes collapse of network

71
Conclusion
72
Additional References
  • Weaver, Paxson, Staniford, Cunningham, A Taxonomy
    of Computer Worms, ACM Workshop on Rapid Malcode,
    2003.
  • Williamson, Throttling Viruses Restricting
    Propagation to Defeat Mobile Malicious Code,
    ACSAC, 2002.
  • Jung, Paxson, Berger, Balakrishnan, Fast Portscan
    Detection Using Sequential Hypothesis Testing,
    IEEE Symposium on Security and Privacy, 2004.
Write a Comment
User Comments (0)
About PowerShow.com