Incrementally Improving Lookup Latency in Distributed Hash Table Systems - PowerPoint PPT Presentation

1 / 31
About This Presentation
Title:

Incrementally Improving Lookup Latency in Distributed Hash Table Systems

Description:

Incrementally Improving Lookup Latency in Distributed Hash Table Systems ... Could Chord do better, e.g., O(1) stretch, without much change? 6/14/2003. Sigmetrics'03 ... – PowerPoint PPT presentation

Number of Views:43
Avg rating:3.0/5.0
Slides: 32
Provided by: Kat6265
Category:

less

Transcript and Presenter's Notes

Title: Incrementally Improving Lookup Latency in Distributed Hash Table Systems


1
Incrementally Improving Lookup Latency in
Distributed Hash Table Systems
Hui Zhang1, Ashish Goel2, Ramesh
Govindan1 1University of Southern
California 2Stanford University
2
Outline
  • Latency stretch problem in Distributed Hash Table
    (DHT) systems, with Chord as an example
  • Two latency stretch theorems
  • Lookup-Parasitic Random Sampling (LPRS)
  • Simulation Internet measurement results
  • Conclusion future work

3
DHT systems
  • A new class of peer-to-peer routing
    infrastructures
  • CAN, Chord, Pastry, Tapestry, etc.
  • Support a hash table-like functionality on
    Internet-like scale
  • a global key space each data item is a key in
    the space, and each node is responsible for a
    portion of the key space.
  • given a key, map it onto a node.
  • Our research results apply to frugal DHT systems.
  • The search space for the key decreases by a
    constant factor after each lookup hop.
  • Examples Chord, Pastry, Tapestry.

4
Chord key space
Network node
A Chord network with 8 nodes and 8-bit key space
5
Chord routing table setup
Network node
Pointer
0
255
In node is routing table One entry is created
to point to to the first node in its jth ranges
i2j-1, i2j), 1 ? j ? m.
A Chord network with N(8) nodes and m(8)-bit
key space
6
Latency stretch in Chord
Network node
Overlay routing
physical link
0
255
A Chord network with N(8) nodes and m(8)-bit
key space
7
Latency stretch Ratnasamy et al. 2001
latency for each lookup on the overlay topology

average latency on the underlying topology
  • In Chord, ?(logN) hops per lookup in average
  • ?(logN) stretch in original Chord.
  • Could Chord do better, e.g., O(1) stretch,
    without much change?

8
Our contributions
  • Theory
  • Latency expansion characteristic of the
    underlying network topology decides latency
    optimization in frugal DHT systems.
  • Exponential latency expansion bad news.
  • Power-law latency expansion good news.
  • System
  • Lookup-Parasitic Random Sample (LPRS), an
    incremental latency optimization technique.
  • Achieve O(1) stretch under power-law latency
    topologies.
  • Internet measurement.
  • The Internet router-level topology resembles
    power-law latency expansion.

9
Latency expansion
  • Let Nu(x) denote the number of nodes in the
    network G that are within latency x of node u.
  • - power-law latency expansion Nu(x) grows (i.e.
    expands') proportionally to xd, for all nodes
    u.
  • Examples ring (d1), mesh (d2).
  • - exponential latency expansion Nu(x)
    grows proportionally to ?x for some constant ? gt
    1.
  • Examples random graphs.

10
Latency-stretch theorem - I
  • Bad news Theorem
  • If the underlying topology G is drawn from a
    family of graphs with exponential latency
    expansion, then the expected latency of Chord is
    ?(LlogN), where L is the expected latency
    between pairs of nodes in G.

11
Latency-stretch theorem - II
  • Good news Theorem
  • If
  • (1) the underlying topology G is drawn from a
    family of graphs with d-power-law latency
    expansion, and
  • (2) for each node u in the Chord network, it
    samples (log N)d nodes in each range with uniform
    randomness and keeps the pointer to the nearest
    node for future routing,
  • then the expected latency of a request is
    O(L), where L is the expected latency between
    pairs of nodes in G.

12
Two remaining questions
  • How does each node efficiently achieve (log N)d
    samples from each range?
  • Do real networks have power-law latency expansion
    characteristic?

13
Uniform sampling in terms of ranges
Node x the node at hop x
Node 0 the request initiator
Node t the request terminator
routing path
14
Lookup-Parasitic Random Sampling
1. Recursive lookup. 2. Each intermediate hop
appends its IP address to the lookup message.
3. When the lookup reaches its target, the
target informs each listed hop of its
identity. 4. Each intermediate hop then sends one
(or a small number) of pings to get a reasonable
estimate of the latency to the target, and update
its routing table accordingly.
15
LPRS-Chord convergence time
Convergence Time
16
LPRS-Chord topology with power-law expansion
Ring Stretch
(at time 2logN)
17
Whats the latency expansion characteristic of
Internet?
18
Internet router-level topology latency
measurement
  • Approximate link latency by geographical latency
  • - assign geo-locations to nodes using
    GeotrackPadmanabhan2001.
  • A large router-level topology dataset
  • - 320,735 nodes, mapped to 603 distinct cities
    all over the world.
  • - 92,824 node pairs are sampled to tractably
    compute the latency expansion of this large
    topology.

19
Internet router-level topology latency expansion
latency expansion
20
LPRS-Chord on router-level topology
Stretch on the router-level subgraphs (at time
2logN)
21
Conclusion
  • LPRS has significant practical applicability as
    a general latency reduction technique for frugal
    DHT systems.
  • Future work
  • - Studying the interaction of LPRS scheme with
    the dynamics of P2P systems.

22
Thank you!
23
Backup slides
24
A simple random sampling solution
Network node
Pointer
Distance measurement
2m-1
0
A Chord network with m-bit key space
25
A simple random sampling solution
Network node
Pointer
Distance measurement
2m-1
0
A Chord network with m-bit key space
26
Term definition (II)
  • Range
  • - for a given node in a Chord overlay with ID j,
    its i-th range Ri(j) is the interval j2i-1,
    j2i) on the key space, where 1 ? i ? m.
  • Frugal routing
  • 1. after each hop, the search space for the
    target reduces by a constant factor, and
  • 2. If w is an intermediate node in the route, v
    is the destination, and v ? Ri(w), then the node
    after w in the route depends only on w and i.

27
LPRS-Chord simulation methodology
Phase 1. N nodes join the network
one-by-one. Phase 2. each node on average
inserts four documents into the network. Phase 3.
each node generates, on average 3logN data
requests one-by-one. - LPRS actions are enabled
only in Phase 3 - Performance measurement begins
at Phase 3
28
Comparison of 5 sampling strategies definitions
  • Consider a lookup that is initiated by node x0,
    then forwarded to node x1, x2, ..., and finally
    reaches the request terminator, node xn
  • 1. Node xi samples node xn, 0 ? i lt n
  • 2. Node xn samples nodes x0, , xn-1
  • 3. Node xi samples node xi-1, 1 ? i ? n
  • 4. Node x0 samples nodes xn
  • 5. Node xi samples node x0, 0 lt i ? n

29
Comparison of 5 sampling strategies simulation
result
30
Zipf-ian document popularity
31
Impact of skewed request distributions
Write a Comment
User Comments (0)
About PowerShow.com