* Your assessment is very important for improving the workof artificial intelligence, which forms the content of this project
Download Exploiting Route Redundancy via Structured Peer to Peer Overlays
Network tap wikipedia , lookup
Multiprotocol Label Switching wikipedia , lookup
Deep packet inspection wikipedia , lookup
Backpressure routing wikipedia , lookup
Computer network wikipedia , lookup
Recursive InterNetwork Architecture (RINA) wikipedia , lookup
List of wireless community networks by region wikipedia , lookup
Airborne Networking wikipedia , lookup
IEEE 802.1aq wikipedia , lookup
Exploiting Route Redundancy via Structured Peer to Peer Overlays Ben Y. Zhao, Ling Huang, Jeremy Stribling, Anthony D. Joseph, and John D. Kubiatowicz University of California, Berkeley ICNP 2003 Challenges Facing Network Applications  Network connectivity is not reliable   Disconnections frequent in the wide-area Internet IP-level repair is slow    Wide-area: BGP  3 mins Local-area: IS-IS  5 seconds Next generation network applications     Mostly wide-area Streaming media, VoIP, B2B transactions Low tolerance of delay, jitter and faults Our work: transparent resilient routing infrastructure that adapts to faults in not seconds, but milliseconds November 7, 2003 ICNP 2003 ravenben@eecs.berkeley.edu Talk Overview       Motivation Why structured routing Structured Peer to Peer overlays Mechanisms and policy Evaluation Summary November 7, 2003 ICNP 2003 ravenben@eecs.berkeley.edu Routing in “Mesh-like” Networks   Previous work has shown reasons for long convergence [Labovitz00, Labovitz01] MinRouteAdver timer  Necessary to aggregate updates from all neighbors    Contributes to lower bound of BGP convergence time Internet becoming more mesh-like [Kaat99,labovitz99]   Commonly set to 30 seconds Worsens BGP convergence behavior Question  Can convergence be faster in context of structured routing? November 7, 2003 ICNP 2003 ravenben@eecs.berkeley.edu Resilient Overlay Networks (MIT)   Fully connected mesh Allows each node full knowledge of network    D Fast, independent calculation of routes Nodes can construct any path, maximum flexibility Cost of flexibility   Protocol needs to choose the “right” route/nodes Per node O(n) state   Monitors n - 1 paths O(n2) total path monitoring is expensive S November 7, 2003 ICNP 2003 ravenben@eecs.berkeley.edu Leveraging Structured Peer-to-Peer Overlays  0 Key based routing (IPTPS 03)  Large sparse ID space N root(k) (160 bits: 0 – 2160)  Nodes in overlay network have nodeIDs  N  Given some key k  N, overlay deterministically maps k to its root node (live node in the network)  route message to root (k) source k  Distributed Hashtables (DHT) is interface on KBR  Key is leveraging underlying routing mesh November 7, 2003 ICNP 2003 ravenben@eecs.berkeley.edu Proximity Neighbor Selection  PNS = network aware overlay construction    Important for routing     Within routing constraints, choose neighbors closest in network distance (latency) Generally reduces # of IP hops Reduce latency Reduce susceptibility to faults  Less IP links = smaller chance of link/router failure Reduce overall network bandwidth utilization We use Tapestry to demonstrate our design   P2P protocol with PNS overlay construction Topology-unaware P2P protocols will likely perform worse November 7, 2003 ICNP 2003 ravenben@eecs.berkeley.edu System Architecture v v v v v v OVERLAY v v v v v v v Internet    Locate nearby overlay proxy Establish overlay path to destination host Overlay traffic routes traffic resiliently November 7, 2003 ICNP 2003 ravenben@eecs.berkeley.edu Traffic Tunneling A, B are IP addresses Legacy Node A P’(B) B Legacy Node B register register Proxy P’(A) = A P’(B) = B Proxy get (hash(B)) P’(B) put (hash(A), P’(A)) put (hash(B), P’(B)) Structured Peer to Peer Overlay   Store mapping from end host IP to its proxy’s overlay ID Similar to approach in Internet Indirection Infrastructure (I3) November 7, 2003 ICNP 2003 ravenben@eecs.berkeley.edu Tradeoffs of Tunneling via P2P  Less neighbor paths to monitor per node: O(log(n))  Large reduction in probing bandwidth: O(n)  O(log(n))    Actively maintain path redundancy      Manageable for “small” # of paths Redirect traffic immediately when a failure is detected Eliminate on-the-fly calculation of new routes Restore redundancy when a path fails End result   Increase probing frequency Faster fault detection with low bandwidth consumption Fast fault detection + precomputed paths = increased responsiveness to faults Cons  Overlay imposes routing stretch (more IP hops), generally < 2 November 7, 2003 ICNP 2003 ravenben@eecs.berkeley.edu Some Details  Efficient fault detection    Use soft-state to periodically probe log(n) neighbor paths “Small” number of routes  reduced bandwidth Exponentially weighted moving average in link quality estimation     Avoid route flapping due to short term loss artifacts Loss rate Ln = (1 - )  Ln-1 +   p p = instantaneous loss rate,  = hysteresis factor Maintaining backup paths  Each hop has flexible routing constraint    Create and store backup routes at node insertion Restore redundancy via “intelligent” gossip after failures Simple policies to choose among redundant paths November 7, 2003 ICNP 2003 ravenben@eecs.berkeley.edu First Reachable Link Selection (FRLS)     Use estimated loss results to choose shortest “usable” path Sort next hop paths by latency Use shortest path with 2299 minimal quality > T Correlated failures   Reduce with intelligent topology construction Key is to leverage redundancy available November 7, 2003 ICNP 2003 2225 2274 2046 2286 2281 2530 1111 ravenben@eecs.berkeley.edu Evaluation  Metrics for evaluation    How much routing resiliency can we exploit? How fast can we adapt to faults? What is the overhead of routing around a failure?    Proportional increase in end to end latency Proportional increase in end to end bandwidth used Experimental platforms  Event-based simulations on transit stub topologies   Data collected over different 5000-node topologies PlanetLab measurements    Microbenchmarks on responsiveness Bandwidth measurements from 200+ node overlays Multiple virtual nodes run per physical machine November 7, 2003 ICNP 2003 ravenben@eecs.berkeley.edu % of All Pairs Reachable Exploiting Route Redundancy (Sim) 1 0.9 0.8 0.7 0.6 0.5 0.4 0.3 0.2 0.1 0 Instantaneous IP 0 0.05 Tapestry / FRLS 0.1 0.15 0.2 Proportion of IP Links Broken   Simulation of Tapestry, 2 backup paths per routing entry Transit-stub topology shown, results from TIER and AS graphs similar November 7, 2003 ICNP 2003 ravenben@eecs.berkeley.edu Responsiveness to Faults (PlanetLab) Time to Switch Routes (ms) 2500 2000 1500 1000 660 alpha=0.2 alpha=0.4 500 0 0 200 300 400 600 800 1000 1200 Link Probe Period (ms)   Response time increases linearly with probe period Minimum link quality threshold T = 70%, 20 runs per data point November 7, 2003 ICNP 2003 ravenben@eecs.berkeley.edu Link Probing Bandwidth (Planetlab) Bandwidth Per Node (KB/s) 7 PR=300ms PR=600ms 6 5 4 3 2 1 0 1 10 100 1000 Size of Overlay   Medium sized routing overlays incur low probing bandwidth Bandwidth increases logarithmically with overlay size November 7, 2003 ICNP 2003 ravenben@eecs.berkeley.edu Related Work  Redirection overlays      Topology estimation techniques      Detour (IEEE Micro 99) Resilient Overlay Networks (SOSP 01) Internet Indirection Infrastructure (SIGCOMM 02) Secure Overlay Services (SIGCOMM 02) Adaptive probing (IPTPS 03) Peer-based shared estimation (Zhuang 03) Internet tomography (Chen 03) Routing underlay (SIGCOMM 03) Structured peer-to-peer overlays  Tapestry, Pastry, Chord, CAN, Kademlia, Skipnet, Viceroy, Symphony, Koorde, Bamboo, X-Ring… November 7, 2003 ICNP 2003 ravenben@eecs.berkeley.edu Conclusion  Benefits of structure outweigh costs  Structured routing lowers path maintenance costs   Can no longer construct arbitrary paths    Structured routing with low redundancy gets very close to ideal in connectivity Incur low routing stretch Fast enough for highly interactive applications    Allows “caching” of backup paths for quick failover 300ms beacon period  response time < 700ms On overlay networks of 300 nodes, b/w cost is 7KB/s Future work   Deploying a public routing and proxy service on PlanetLab Examine impact of   Network aware topology construction Loss sensitive probing techniques November 7, 2003 ICNP 2003 ravenben@eecs.berkeley.edu Questions…  Related websites: Tapestry http://www.cs.berkeley.edu/~ravenben/tapestry  Pastry http://research.microsoft.com/~antr/pastry  Chord http://lcs.mit.edu/chord   Acknowledgements  Thanks to Dennis Geels and Sean Rhea for their work on the BMark benchmark suite November 7, 2003 ICNP 2003 ravenben@eecs.berkeley.edu Backup Slides November 7, 2003 ICNP 2003 ravenben@eecs.berkeley.edu Another Perspective on Reachability Proportion of All Paths 1 0.9 0.8 Portion of all pairwise paths where no failure-free paths remain 0.7 A path exists, but neither IP nor FRLS can locate the path 0.6 0.5 0.4 0.3 0.2 0.1 Portion of0 all paths where IP0 and FRLS both route successfully November 7, 2003 0.05 0.1 0.15 Proportion of IP Links Broken ICNP 2003 FRLS finds path, 0.2 where short-term IP routing fails ravenben@eecs.berkeley.edu Constrained Multicast     Used only when all paths are below quality threshold Send duplicate messages on multiple paths Leverage route convergence 2299 2274  Assign unique message IDs  Mark duplicates  Keep moving window of IDs 2046 2281  Recognize and drop duplicates ? ? ? Limitations  Assumes loss not from congestion 1111  Ideal for local area routing November 7, 2003 ICNP 2003 2225 2286 2530 ravenben@eecs.berkeley.edu Latency Overhead of Misrouting Proportional Latency Overhead Hop 0 Hop 1 Hop 4 Hop 3 Hop 2 1 0.9 0.8 0.7 0.6 0.5 0.4 0.3 0.2 0.1 0 20 40 60 80 100 120 160 140 180 Latency from source to destination (ms) November 7, 2003 ICNP 2003 ravenben@eecs.berkeley.edu Bandwidth Cost of Constrained Multicast Proportional Bandwidth Overhead Hop 0 Hop 1 Hop 2 Hop 3 Hop 4 1 0.9 0.8 0.7 0.6 0.5 0.4 0.3 0.2 0.1 0 20 40 60 80 100 120 140 160 Latency from source to destination (ms) November 7, 2003 ICNP 2003 ravenben@eecs.berkeley.edu Challenge 2: Tunneling Application Traffic  Basic idea    Desired properties     Tunnel “legacy” traffic via overlay proxies Should be protocol independent Traffic redirection transparent to end hosts Stable mapping between a proxy and its legacy nodes Incremental deployment Details  Making a connection to D     Determine if D is reachable from overlay If so, retrieve D’s overlay ID from storage in overlay Redirect traffic through overlay to D’s overlay ID Assign legacy nodes IDs in the overlay  Routing to IDs automatically reach respective proxies November 7, 2003 ICNP 2003 ravenben@eecs.berkeley.edu Assigning Overlay Identifiers     0 A proxy close by in network assigns a legacy node a non-random overlay ID Assign closest possible ID to proxy’s ID  proxy proxy source node ID such that: root (ID) = proxy Chord/Tapestry: ID = proxyID – 1 Stable ID mapping   Probability of a new node “hijacking” legacy node very low 1 million nodes, 1000 nodes / proxy, 160 bit names, Ph < 10-40 November 7, 2003 ICNP 2003 destination node ID ravenben@eecs.berkeley.edu Unused Slides November 7, 2003 ICNP 2003 ravenben@eecs.berkeley.edu Summary  Highly adaptive to link failures  Fast adaptation time w/ low monitoring BW cost    300ms beacon period, 7KB/s (300 overlay nodes), response time < 700ms Low latency and BW overhead for misrouting Limitation: less useful for congested IP links    Constrained multicast can exacerbate congestion Can utilize more physically-aware overlay construction Issue: using loss as congestion indicator  Can fix with more intelligent protocols such as XCP November 7, 2003 ICNP 2003 ravenben@eecs.berkeley.edu Possible Approaches  Modify basic IP protocols    “Helper” infrastructures to assist protocols   BGP, OSPF, IS-IS Millions of deployed routers  Significant deployment challenge Manipulate input to improve performance without modification Deploy intelligent protocols above IP layer   Overlay approach taken by many Relies on basic point to point IP routing November 7, 2003 ICNP 2003 ravenben@eecs.berkeley.edu