One Hop Lookups for Peer-to-Peer Overlays Anjali Gupta, Barbara Liskov, Rodrigo Rodrigues Laboratory for Computer Science, MIT.

Slides:



Advertisements
Similar presentations
C. Mastroianni, D. Talia, O. Verta - A Super-Peer Model for Resource Discovery Services in Grids A Super-Peer Model for Building Resource Discovery Services.
Advertisements

Correctness of Gossip-Based Membership under Message Loss Maxim GurevichIdit Keidar Technion.
An Adaptive Compulsory Protocol for Basic Communication in Ad-hoc Mobile Networks Ioannis Chatzigiannakis Sotiris Nikoletseas April 2002.
Alex Cheung and Hans-Arno Jacobsen August, 14 th 2009 MIDDLEWARE SYSTEMS RESEARCH GROUP.
P2P data retrieval DHT (Distributed Hash Tables) Partially based on Hellerstein’s presentation at VLDB2004.
Ion Stoica, Robert Morris, David Karger, M. Frans Kaashoek, Hari Balakrishnan MIT and Berkeley presented by Daniel Figueiredo Chord: A Scalable Peer-to-peer.
Peer to Peer and Distributed Hash Tables
Pastry Peter Druschel, Rice University Antony Rowstron, Microsoft Research UK Some slides are borrowed from the original presentation by the authors.
Peter Druschel, Rice University Antony Rowstron, Microsoft Research UK
Scalable Content-Addressable Network Lintao Liu
CHORD – peer to peer lookup protocol Shankar Karthik Vaithianathan & Aravind Sivaraman University of Central Florida.
Chord: A scalable peer-to- peer lookup service for Internet applications Ion Stoica, Robert Morris, David Karger, M. Frans Kaashock, Hari Balakrishnan.
1 1 Chord: A scalable Peer-to-peer Lookup Service for Internet Applications Dariotaki Roula
P2psip WG, IETF841 Jin Peng Qing Yu Yuan Li One Hop Lookups Algorithm Plugin for RELOAD draft-peng-p2psip-one-hop-plugin-02.
Chord: A Scalable Peer-to-peer Lookup Service for Internet Applications Ion StoicaRobert Morris David Liben-NowellDavid R. Karger M. Frans KaashoekFrank.
Massively Distributed Database Systems Distributed Hash Spring 2014 Ki-Joune Li Pusan National University.
Pastry Peter Druschel, Rice University Antony Rowstron, Microsoft Research UK Some slides are borrowed from the original presentation by the authors.
LightFlood: An Optimal Flooding Scheme for File Search in Unstructured P2P Systems Song Jiang, Lei Guo, and Xiaodong Zhang College of William and Mary.
Common approach 1. Define space: assign random ID (160-bit) to each node and key 2. Define a metric topology in this space,  that is, the space of keys.
Small-world Overlay P2P Network
Scribe: A Large-Scale and Decentralized Application-Level Multicast Infrastructure Miguel Castro, Peter Druschel, Anne-Marie Kermarrec, and Antony L. T.
Web Caching Schemes1 A Survey of Web Caching Schemes for the Internet Jia Wang.
Peer to Peer File Sharing Huseyin Ozgur TAN. What is Peer-to-Peer?  Every node is designed to(but may not by user choice) provide some service that helps.
Scalable Application Layer Multicast Suman Banerjee Bobby Bhattacharjee Christopher Kommareddy ACM SIGCOMM Computer Communication Review, Proceedings of.
Spring 2003CS 4611 Peer-to-Peer Networks Outline Survey Self-organizing overlay network File system on top of P2P network Contributions from Peter Druschel.
Chord: A Scalable Peer-to-peer Lookup Service for Internet Applications Ion Stoica, Robert Morris, David Karger, M. Frans Kaashoek and Hari alakrishnan.
Chord-over-Chord Overlay Sudhindra Rao Ph.D Qualifier Exam Department of ECECS.
Adaptive Self-Configuring Sensor Network Topologies ns-2 simulation & performance analysis Zhenghua Fu Ben Greenstein Petros Zerfos.
Topics in Reliable Distributed Systems Fall Dr. Idit Keidar.
1 CS 194: Distributed Systems Distributed Hash Tables Scott Shenker and Ion Stoica Computer Science Division Department of Electrical Engineering and Computer.
Wide-area cooperative storage with CFS
1 Peer-to-Peer Networks Outline Survey Self-organizing overlay network File system on top of P2P network Contributions from Peter Druschel.
“Umbrella”: A novel fixed-size DHT protocol A.D. Sotiriou.
Chord & CFS Presenter: Gang ZhouNov. 11th, University of Virginia.
“Intra-Network Routing Scheme using Mobile Agents” by Ajay L. Thakur.
Peer to Peer Research survey TingYang Chang. Intro. Of P2P Computers of the system was known as peers which sharing data files with each other. Build.
1 The Design of a Robust Peer-to-Peer System Rodrigo Rodrigues, Barbara Liskov, Liuba Shrira Presented by Yi Chen Some slides are borrowed from the authors’
Using the Small-World Model to Improve Freenet Performance Hui Zhang Ashish Goel Ramesh Govindan USC.
Ion Stoica, Robert Morris, David Karger, M. Frans Kaashoek, Hari Balakrishnan MIT and Berkeley presented by Daniel Figueiredo Chord: A Scalable Peer-to-peer.
Content Addressable Network CAN. The CAN is essentially a distributed Internet-scale hash table that maps file names to their location in the network.
Network Layer4-1 Distance Vector Algorithm Bellman-Ford Equation (dynamic programming) Define d x (y) := cost of least-cost path from x to y Then d x (y)
Presentation 1 By: Hitesh Chheda 2/2/2010. Ion Stoica, Robert Morris, David Karger, M. Frans Kaashoek, Hari Balakrishnan MIT Laboratory for Computer Science.
Chord: A Scalable Peer-to-peer Lookup Service for Internet Applications.
Kaleidoscope – Adding Colors to Kademlia Gil Einziger, Roy Friedman, Eyal Kibbar Computer Science, Technion 1.
1 Peer-to-Peer Technologies Seminar by: Kunal Goswami (05IT6006) School of Information Technology Guided by: Prof. C.R.Mandal, School of Information Technology.
Peer to Peer A Survey and comparison of peer-to-peer overlay network schemes And so on… Chulhyun Park
The new protocol of freenet Taken from Ian Clarke and Oskar Sandberg (The Freenet Project)
15 November 2005LCN Collision Detection and Resolution in Hierarchical Peer-to-Peer Systems Verdi March 1, Yong Meng Teo 1,2, Hock Beng Lim 2, Peter.
1 Secure Peer-to-Peer File Sharing Frans Kaashoek, David Karger, Robert Morris, Ion Stoica, Hari Balakrishnan MIT Laboratory.
Chord Advanced issues. Analysis Theorem. Search takes O (log N) time (Note that in general, 2 m may be much larger than N) Proof. After log N forwarding.
1 Distributed Hash Table CS780-3 Lecture Notes In courtesy of Heng Yin.
LightFlood: An Efficient Flooding Scheme for File Search in Unstructured P2P Systems Song Jiang, Lei Guo, and Xiaodong Zhang College of William and Mary.
1. Outline  Introduction  Different Mechanisms Broadcasting Multicasting Forward Pointers Home-based approach Distributed Hash Tables Hierarchical approaches.
Chord Advanced issues. Analysis Search takes O(log(N)) time –Proof 1 (intuition): At each step, distance between query and peer hosting the object reduces.
CS 6401 Overlay Networks Outline Overlay networks overview Routing overlays Resilient Overlay Networks Content Distribution Networks.
Algorithms and Techniques in Structured Scalable Peer-to-Peer Networks
Energy Efficient Data Management for Wireless Sensor Networks with Data Sink Failure Hyunyoung Lee, Kyoungsook Lee, Lan Lin and Andreas Klappenecker †
1 Roie Melamed, Technion AT&T Labs Araneola: A Scalable Reliable Multicast System for Dynamic Wide Area Environments Roie Melamed, Idit Keidar Technion.
Adlib : A Self-Tuning Index for Dynamic Peer-to-Peer Systems Proceedings of the 21st International Conference on Data Engineering(ICDE'05) Prasanna Ganesan.
Chord: A Scalable Peer-to-Peer Lookup Service for Internet Applications * CS587x Lecture Department of Computer Science Iowa State University *I. Stoica,
Pastry Scalable, decentralized object locations and routing for large p2p systems.
A Scalable Peer-to-peer Lookup Service for Internet Applications
Controlling the Cost of Reliability in Peer-to-Peer Overlays
Issues in Peer-to-Peer Computing
DHT Routing Geometries and Chord
Chord Advanced issues.
Chord Advanced issues.
Chord Advanced issues.
EEC 688/788 Secure and Dependable Computing
A Scalable Peer-to-peer Lookup Service for Internet Applications
Presentation transcript:

One Hop Lookups for Peer-to-Peer Overlays Anjali Gupta, Barbara Liskov, Rodrigo Rodrigues Laboratory for Computer Science, MIT

Peer-to-Peer Systems Large scale dynamic network Overlay infrastructure :  Scalable  Self configuring  Fault tolerant Every node responsible for some objects Find node having desired object Challenge : Efficient Routing

Routing in Current P2P Systems Routing state size logarithmic in the number of overlay nodes Membership changes frequent Small routing table Less bookkeeping Logarithmic overlay hops per lookup, high latency Amortized cost unacceptable for storing and retrieving small objects

Our Thesis It is feasible to : Keep full routing state on every node Route in one overlay hop Achieve high lookup success rate Thus we can : Enable a large class of applications for peer-to-peer systems

Outline Structured Peer-to-Peer Overlay Goals Design Analysis Future Work

Structured Overlay Objects and nodes have identifiers A specific node is responsible for each object Different data placement algorithms, e.g., consistent hashing Successor node Node Object

Dynamic Membership Nodes join and leave Gnutella study (Saroiu et al) : average node session time 2.9 hours Rate of change Number of nodes For 100,000 nodes, approximately 20 membership changes per second

Goals

Routing Goal How does node N find object O?  Finds successor S(O) by looking in its own table  Sends a message to S(O)  If S(O) is current successor of O, responds with success message Lookup success:  Object found in first attempt Achieve high lookup success rate, e.g., 99%

Design Goals Information about a node joining or leaving the system must reach all nodes rapidly Bandwidth requirement should be feasible Should be scalable Hierarchical Scheme!

Design

Hierarchical Scheme Ring is divided statically into slices (i.e., slices are just identifier intervals) node

Hierarchical Scheme Successor of midpoint of slice is slice leader node

Hierarchical Scheme Each slice is divided statically into units Again, units are just intervals Successor of midpoint of unit is unit leader

Base Communication Node exchanges frequent keep-alive messages with predecessor and successor Detects change in successor : event Recent event log Piggyback event log on keep-alive messages Is this fast enough?

Flow of Information : Inter-slice Step 1: Event detected by node N N

Flow of Information : Inter-slice Step 2: N notifies its slice leader N Slice leader

Flow of Information : Inter-slice Step 3: N’s slice leader collects events for some time, then notifies other slice leaders N Slice leader

Is this fast enough? Slices may be large! Piggybacking on keep-alive messages can take a long time

Flow of Information : Intra-slice Step 4: S lice leaders notify their respective unit leaders periodically N Slice leader Unit leader

Analysis

Speed of Propagation Units kept small to spread information quickly Possible frequency of communications :  Slice leaders communicate with each other once every 30 seconds  Slice leader communicates with its unit leaders once every 5 seconds  Nodes exchange keep-alive messages every second

Speed of Propagation If (expected) unit size is 20, the time taken to reach farthest nodes is:  Node to Slice leader : 0 s  Slice leader to other slice leaders : 30 s  Slice leader to unit leaders : 5 s  Unit leader to edge of unit : 10 s Total time : 45 seconds!

Time Constraint Recall : Goal of 99% lookup success rate If f : acceptable lookup failure rate n : number of overlay nodes r : rate of membership change Then: All nodes must be notified about an event within e.g.: For f = 0.01, n = , r = 20 changes/second time interval = 50 seconds

Bandwidth Use 1. A slice leader tells another slice leader only about events in its own slice 2. A slice leader never sends the same notification twice to any slice leader 3. Event notification is sent by a node to its neighbor only if : 1. The neighbor is in the same unit 2. It has not already sent this notification to this neighbor

Bandwidth Use Number of slices (k) chosen to minimize total bandwidth use. We prove that For an overlay having 100,000 nodes:  Per node: 4 kbps (up, down)  Per unit leader: 4 kbps (up, down)  Per slice leader: 240 kbps (upstream), 4 kbps (down)

Slice Leaders Important to choose slice leaders correctly Super-node scheme Sufficient number of well-provisioned nodes Incentives for maintaining ring: Application dependent

Fault Tolerance Unit leader failure : easy recovery Slice leader failure  Contact unit leaders  Contact other slice leaders

Related Work Design of a Robust P2P system - R. Rodrigues et al Kelips - I. Gupta et al Controlling the Cost of Reliability in P2P Overlays - R. Mahajan et al

Future Work System implementation almost complete Experiments to see actual system behavior Systems larger than a million nodes  Two hop scheme