Title: The Internet Motion Sensor: A Distributed Blackhole Monitoring System
1The Internet Motion SensorA Distributed
Blackhole Monitoring System
- Michael Bailey, Evan Cooke, Farnam Jahanian,
- Jose Nazario, David Watson
-
Presenter Anup Goyal
2What is a Blackhole?
3Telescope/ Blackhole/ DarkNet/ Sink
- Monitoring of unused Address space is very useful
- CAIDA-Network Telescope
- Internet Motion Sensor blackhole
- Team Cymru DarkNets
- IUCC/IDC Internet Telescope
- Isink
- Investigating DDOS
- Tracking Worms
- Characterizing emerging Internet Thread
4Motivation
- Passive Monitoring has been around for a while,
why build another one? - Need broader coverage
- Need to be able to differentiate threats
- Our Solutions
- Distributed Monitoring Infrastructure
- Increas coverage of threat activity
- Lightweight Active Responder
- Capture payloads to differentiate threats
- Payload signature and caching
- Improve performance and scalability
5Why do you need to be distributed?
6Local Preference across sensors
7Distributed Monitoring Infrastructure
8Differentiate Threats?
9Why not Build Service Modules?
10(No Transcript)
11Lightweight Active Responder
- Goal obtain enough fidelity to differentiate
threats with the minimum resource cost - TCP needs to establish a connection before data
is sent - Use Lightweight SYN-ACK active Responder
12Blaster Worm Live Host
13Blaster Worm Passive Monitor
14Blaster Worm - IMS
15Active Responder Limitations
- Application-level responder may be required to
differentiate threats (e.g. Agobot) - Threats like Agobot can be differentiated using
scanning behaviour - Sensors can be fingerprinted and avoided
- IMS focused on globally scoped threats (threat
model does not include targated manual attacks) - Many sensors of different sizes in many networks
near live hosts makes avoidance very hard - Rotate active responders within address block
16Payload Signature and Caching
17IMS Deployment
- Initial /8 deployment in 2001. Currently 60
address blocks at 18 networks on 3 continent - Tier 1 ISPs, Large Enterprise, Broadband,
Academic, National ISPs, Regional ISPs
18Deployment Statistics
- 17,094,016 IPs monitored
- 27 /8 block with an IMS sensor
- 1.25 of routed IPv4 space
- 21 of all routable /8 blocks have at least one
sensor - IMS Has provided insight into
- Internet Worms
- Reconnaissance/Scanning
- DDos
19Internet Worms
20Scanning
21SCO DDoS
22Summary Outreach
- Summary
- IMS provides a lightweight but effective platform
for tracking, and characterizing Internet threats - Successful deployment on 60 address blocks at 18
organization - Outreach
- IMS currently used daily by operator community to
investigate new threats - We provide reports and forensics to NSP-SEC
- We are now focusing academic networks
- Data traces will be available through DHS PREDICT
project
23Scalability, Fidelity, and Containment in
thePotemkin Virtual Honeyfarm
- Michael Vrable, Justin Ma, Jay Chen, David Moore,
Erik Vandekieft, Alex C. Snoeren, Geoffrey M.
Voelker,Stefan Savage - Presenter Anup Goyal
24Background
- Large-scale host exploitation a serious problem
- Worms, viruses, bots, spyware
- Supports an emerging economic criminal enterprise
- SPAM, DDoS, phishing, piracy, ID theft
- Quality and sophistication of malware increasing
rapidly
25Motivation
- Intelligence about new threats is critical for
defenders - Principal tool is the network honeypot
- Monitored system deployed for the purpose of
being attacked - Honeyfarm Collection of honeypots
- Provide early warning, accurate inference of
global activity,cover wide range of software - Design issues
- Scalability How many honeypots can be deployed
- Fidelity How accurately systems are emulated
- Containment How well innocent third parties are
protected - Challenge tension between scalability and
fidelity
26Honeyfarm Scalability/Fidelity Tradeoff
27Approach
- Dedicated honeypot systems are overkill
- Can provide the illusion of dedicated systems via
aggressive resource multiplexing at network and
host levels
28Network-level Multiplexing
- Most addresses dont receive traffic most of the
time - Apply late binding of IP addresses to honeypots
- Most traffic that is received causes no
interesting effects - - Allocate honeypots only long enough to identify
interesting behavior - Recycle honeypots as soon as possible
- How many honeypots are required?
- - For a given request rate, depends upon
recycling rate
29Effectiveness of Network-Level Multiplexing
30Host-Level Multiplexing
- CPU utilization in each honeypot quite low
(milliseconds to process traffic) - Use VMM to multiplex honeypots on a single
physical machine - Few memory pages actually modified when handling
network data - Share unmodified pages among honeypots within a
machine - How many virtual machines can we support?
- - Limited by unique memory required per VM
31Effectiveness of Host-Level Multiplexing
- Further 2-3 order of magnitude improvement
32The Potemkin Honeyfarm Architecture
- Two components
- Gateway
- VMM
- Basic operation
- Packet received by gateway
- Dispatched to honeyfarm server
- VM instantiated
- Adopts IP address
33Potemkin VMM Requirements
- VMs created on demand
- VM creation must be fast enough to maintain
illusion - Many VMs created
- - Must be resource-efficient
34Potemkin VMM Overview
- Modified version of Xen 3.0 (pre-release)
- Flash cloning
- Fork copies from a reference honeypot VM
- Reduces VM creation timeno need to boot
- Applications all ready to run
- Delta virtualization
- Copy-on-write sharing (between VMs)
- Reduces per-VM stateonly stores unique data
- Further reduces VM creation time
35Flash Cloning Performance
- Time required to clone a 128 MB honeypot
- Control tools overhead 124 ms
- Low-level clone 11 ms
- Device setup 149 ms
- Other management overhead 79 ms
- Networking setup overhead 158 ms
- Total 521 ms
- 0.5 s already imperceptible to external observers
unless looking for delay, but we can do even
better
36Delta Virtualization Performance
- Deployed using 128 MB Linux honeypots
- Using servers with 2 GB RAM, have memory
available to support 1000 VMs per physical host - Currently tested with 100 VMs per host
- Hits artificial resource limit in Xen, but this
can be fixed
37Containment Policies
- Must also care about traffic going out
- We deliberately run unpatched, insecure software
in honeypots - Containment Should not permit attacks on third
parties - As with scalability, there is a tension between
containment and fidelity - Various containment policies we support
- Allow no traffic out
- Allow traffic over established connections
- Allow traffic back to original host
38Containment Implementation in Gateway
- Containment policies implemented in network
gateway - Tracks mappings between IP addresses, honeypots,
- connections
- Modular implementation in Click
- Gateway adds insignificant overhead ( 1 ms)
39Traffic Reflection
- Example gateway policy
- Redirect traffic back to
- Honeyfarm
- Packets sent out to third parties. . .
- . . .may be redirected back into honeyfarm
- Reuses honeypot creation functionality
40Challenges
- Honeypot detection
- If malware detects it is in a honeypot, may act
differently - How easy it is to detect virtualization?
- VMware detection code used in the wild
- Open arms race between honeypot detection and
camouflage - Resource exhaustion
- Under high load, difficult to maintain accurate
illusion - Large-scale outbreak
- Honeypot denial-of-service
- Challenge is intelligently shedding load
41Summary
- Can achieve both high fidelity and scalability
- Sufficient to provide the illusion of scale
- Potemkin prototype 65k addresses -gt 10 physical
hosts - Largest high-fidelity honeypot that we are aware
of - Provides important tool for study of and defenses
against malware