SCALLOP A Scalable and Load-Balanced Peer- to-Peer Lookup Protocol for High- Performance Distributed System Jerry Chou, Tai-Yi Huang & Kuang-Li Huang Embedded.

Slides:



Advertisements
Similar presentations
Ion Stoica, Robert Morris, David Karger, M. Frans Kaashoek, Hari Balakrishnan MIT and Berkeley presented by Daniel Figueiredo Chord: A Scalable Peer-to-peer.
Advertisements

Pastry Peter Druschel, Rice University Antony Rowstron, Microsoft Research UK Some slides are borrowed from the original presentation by the authors.
Scalable Content-Addressable Network Lintao Liu
1Department of Electrical Engineering and Computer Science, University of Michigan, USA. 2Department of Computer Science, National University of Singapore,
CHORD – peer to peer lookup protocol Shankar Karthik Vaithianathan & Aravind Sivaraman University of Central Florida.
CHORD: A Peer-to-Peer Lookup Service CHORD: A Peer-to-Peer Lookup Service Ion StoicaRobert Morris David R. Karger M. Frans Kaashoek Hari Balakrishnan Presented.
Chord: A scalable peer-to- peer lookup service for Internet applications Ion Stoica, Robert Morris, David Karger, M. Frans Kaashock, Hari Balakrishnan.
1 1 Chord: A scalable Peer-to-peer Lookup Service for Internet Applications Dariotaki Roula
Chord A Scalable Peer-to-peer Lookup Service for Internet Applications
Massively Distributed Database Systems Distributed Hash Spring 2014 Ki-Joune Li Pusan National University.
Sylvia Ratnasamy, Paul Francis, Mark Handley, Richard Karp, Scott Schenker Presented by Greg Nims.
SCRIBE A large-scale and decentralized application-level multicast infrastructure.
Pastry Peter Druschel, Rice University Antony Rowstron, Microsoft Research UK Some slides are borrowed from the original presentation by the authors.
1 PASTRY Partially borrowed from Gabi Kliot ’ s presentation.
Common approach 1. Define space: assign random ID (160-bit) to each node and key 2. Define a metric topology in this space,  that is, the space of keys.
Scribe: A Large-Scale and Decentralized Application-Level Multicast Infrastructure Miguel Castro, Peter Druschel, Anne-Marie Kermarrec, and Antony L. T.
Peer to Peer File Sharing Huseyin Ozgur TAN. What is Peer-to-Peer?  Every node is designed to(but may not by user choice) provide some service that helps.
Chord: A Scalable Peer-to-Peer Lookup Protocol for Internet Applications Stoica et al. Presented by Tam Chantem March 30, 2007.
Fault-tolerant Routing in Peer-to-Peer Systems James Aspnes Zoë Diamadi Gauri Shah Yale University PODC 2002.
Spring 2003CS 4611 Peer-to-Peer Networks Outline Survey Self-organizing overlay network File system on top of P2P network Contributions from Peter Druschel.
A Scalable Content-Addressable Network Authors: S. Ratnasamy, P. Francis, M. Handley, R. Karp, S. Shenker University of California, Berkeley Presenter:
A Scalable and Load-Balanced Lookup Protocol for High Performance Peer-to-Peer Distributed System Jerry Chou and Tai-Yi Huang Embedded & Operating System.
Chord-over-Chord Overlay Sudhindra Rao Ph.D Qualifier Exam Department of ECECS.
Topics in Reliable Distributed Systems Fall Dr. Idit Keidar.
Wide-area cooperative storage with CFS
Or, Providing Scalable, Decentralized Location and Routing Network Services Tapestry: Fault-tolerant Wide-area Application Infrastructure Motivation and.
1 Peer-to-Peer Networks Outline Survey Self-organizing overlay network File system on top of P2P network Contributions from Peter Druschel.
1CS 6401 Peer-to-Peer Networks Outline Overview Gnutella Structured Overlays BitTorrent.
Pastry: Scalable, decentralized object location and routing for large-scale peer-to-peer systems (Antony Rowstron and Peter Druschel) Shariq Rizvi First.
Structured P2P Network Group14: Qiwei Zhang; Shi Yan; Dawei Ouyang; Boyu Sun.
INTRODUCTION TO PEER TO PEER NETWORKS Z.M. Joseph CSE 6392 – DB Exploration Spring 2006 CSE, UT Arlington.
Roger ZimmermannCOMPSAC 2004, September 30 Spatial Data Query Support in Peer-to-Peer Systems Roger Zimmermann, Wei-Shinn Ku, and Haojun Wang Computer.
Effizientes Routing in P2P Netzwerken Chord: A Scalable Peer-to- peer Lookup Protocol for Internet Applications Dennis Schade.
1 PASTRY. 2 Pastry paper “ Pastry: Scalable, decentralized object location and routing for large- scale peer-to-peer systems ” by Antony Rowstron (Microsoft.
09/07/2004Peer-to-Peer Systems in Mobile Ad-hoc Networks 1 Lookup Service for Peer-to-Peer Systems in Mobile Ad-hoc Networks M. Tech Project Presentation.
Chord & CFS Presenter: Gang ZhouNov. 11th, University of Virginia.
Using the Small-World Model to Improve Freenet Performance Hui Zhang Ashish Goel Ramesh Govindan USC.
1 Reading Report 5 Yin Chen 2 Mar 2004 Reference: Chord: A Scalable Peer-To-Peer Lookup Service for Internet Applications, Ion Stoica, Robert Morris, david.
Chord: A Scalable Peer-to-peer Lookup Protocol for Internet Applications Xiaozhou Li COS 461: Computer Networks (precept 04/06/12) Princeton University.
1 Distributed Hash Tables (DHTs) Lars Jørgen Lillehovde Jo Grimstad Bang Distributed Hash Tables (DHTs)
Network Computing Laboratory Scalable File Sharing System Using Distributed Hash Table Idea Proposal April 14, 2005 Presentation by Jaesun Han.
Content Addressable Network CAN. The CAN is essentially a distributed Internet-scale hash table that maps file names to their location in the network.
A Scalable Content-Addressable Network (CAN) Seminar “Peer-to-peer Information Systems” Speaker Vladimir Eske Advisor Dr. Ralf Schenkel November 2003.
Presentation 1 By: Hitesh Chheda 2/2/2010. Ion Stoica, Robert Morris, David Karger, M. Frans Kaashoek, Hari Balakrishnan MIT Laboratory for Computer Science.
An IP Address Based Caching Scheme for Peer-to-Peer Networks Ronaldo Alves Ferreira Joint work with Ananth Grama and Suresh Jagannathan Department of Computer.
Chord: A Scalable Peer-to-peer Lookup Service for Internet Applications.
Chord: A Scalable Peer-to-peer Lookup Service for Internet Applications Ion Stoica, Robert Morris, David Karger, M. Frans Kaashoek, Hari Balakrishnan Presented.
SIGCOMM 2001 Lecture slides by Dr. Yingwu Zhu Chord: A Scalable Peer-to-peer Lookup Service for Internet Applications.
Scalable Content- Addressable Networks Prepared by Kuhan Paramsothy March 5, 2007.
Peer to Peer A Survey and comparison of peer-to-peer overlay network schemes And so on… Chulhyun Park
15 November 2005LCN Collision Detection and Resolution in Hierarchical Peer-to-Peer Systems Verdi March 1, Yong Meng Teo 1,2, Hock Beng Lim 2, Peter.
1 Secure Peer-to-Peer File Sharing Frans Kaashoek, David Karger, Robert Morris, Ion Stoica, Hari Balakrishnan MIT Laboratory.
1 Distributed Hash Table CS780-3 Lecture Notes In courtesy of Heng Yin.
Chord Fay Chang, Jeffrey Dean, Sanjay Ghemawat, Wilson C. Hsieh, Deborah A. Wallach, Mike Burrows, Tushar Chandra, Andrew Fikes, Robert E. Gruber Google,
Dynamic Networks for Peer-to-Peer Systems Pierre Fraigniaud CNRS Lab. de Recherche en Informatique (LRI) Univ. Paris-Sud, Orsay Joint work with Philippe.
BATON A Balanced Tree Structure for Peer-to-Peer Networks H. V. Jagadish, Beng Chin Ooi, Quang Hieu Vu.
Algorithms and Techniques in Structured Scalable Peer-to-Peer Networks
LOOKING UP DATA IN P2P SYSTEMS Hari Balakrishnan M. Frans Kaashoek David Karger Robert Morris Ion Stoica MIT LCS.
CS 347Notes081 CS 347: Parallel and Distributed Data Management Notes 08: P2P Systems.
Malugo – a scalable peer-to-peer storage system..
CS694 - DHT1 Distributed Hash Table Systems Hui Zhang University of Southern California.
Incrementally Improving Lookup Latency in Distributed Hash Table Systems Hui Zhang 1, Ashish Goel 2, Ramesh Govindan 1 1 University of Southern California.
Chord: A Scalable Peer-to-Peer Lookup Service for Internet Applications * CS587x Lecture Department of Computer Science Iowa State University *I. Stoica,
Peer-to-Peer Information Systems Week 12: Naming
A Scalable Peer-to-peer Lookup Service for Internet Applications
Early Measurements of a Cluster-based Architecture for P2P Systems
SCOPE: Scalable Consistency in Structured P2P Systems
DHT Routing Geometries and Chord
MIT LCS Proceedings of the 2001 ACM SIGCOMM Conference
Peer-to-Peer Information Systems Week 12: Naming
Presentation transcript:

SCALLOP A Scalable and Load-Balanced Peer- to-Peer Lookup Protocol for High- Performance Distributed System Jerry Chou, Tai-Yi Huang & Kuang-Li Huang Embedded & Operating System (EOS) Lab Computer Science Department National Tsing Hua University, Taiwan

CCGRID 2004 EOS Lab, National Tsing Hua University Outline Contributions Methodology Experimental Results Conclusions Future work

CCGRID 2004 EOS Lab, National Tsing Hua University Motivations and Purposes (1/2) Many large-scaled servers are implemented in a peer-to-peer distributed system due to:  Low cost of workstations  Availability of high-speed network Two main concerns of the system  Performance  response time  Stability  hot spot problem

CCGRID 2004 EOS Lab, National Tsing Hua University Motivations and Purposes (2/2) Response time = Lookup time + Service time  Shorter lookup forwarding path  smaller lookup time  Balanced-load on nodes  reduce service time Hot spot problems  An overloaded node becomes disable or malfunction  The network traffic around hot spots increases  Resulting in an unstable system  Decreasing system performance

CCGRID 2004 EOS Lab, National Tsing Hua University Related Work CAN: A scalable content addressable network. In ACM SIGCOMM 2001 Pastry: Scalable, decentralized object location, and routing for large-scale P2P system. In Lecture Notes in Computer Science 2001 Chord: A Scalable P2P lookup service for Internet applications. In ACM SIGCOMM 2001 Tapestry: An infrastructure for fault-tolerant wide-area location and routing. Technical Report 2001  All of them do not address the hot spot problem

CCGRID 2004 EOS Lab, National Tsing Hua University Our Contributions (Stability) Reduce or eliminate hot spots  Balanced-load Both data items and lookup requests are evenly distributed  Service time is also reduced  The number of children nodes is limited Each node is at most accessed by other d nodes Decentralized  Without the bottleneck on centralized servers

CCGRID 2004 EOS Lab, National Tsing Hua University Our Contributions (Performance) Scale with the number of nodes  Each node is only aware of other O(d log d N) nodes Provide an upper bound on lookup paths  The lookup path for any request is O(log d N) Allow a tradeoff between space and time  If d becomes larger  More routing information required  Shorter lookup path

CCGRID 2004 EOS Lab, National Tsing Hua University Methodology Data partition Structure of the balanced lookup tree Construction of the lookup table Self-organized mechanism

CCGRID 2004 EOS Lab, National Tsing Hua University Data Partition Use a variant of consistent hashing to partition the data set among all nodes  key k is stored in node n where |k-n| is minimum Two proven properties of consistent hashing  Each node stores similar amount of data items  Data movement is minimum when system changes Our protocol assigns each node a number, called SID, between 0 to N-1

CCGRID 2004 EOS Lab, National Tsing Hua University Balanced Lookup Tree We construct a balanced lookup tree for each node The root of a lookup tree of a node is the node itself Lookup path is bounded by O(log d N) k-2 k-dk-d k-d-1k-2d-1 … k-d 2 -1k-d-2k-2d-2 … k-d 2 -2 k-2d k-3d k-d 2 -d-1 ………………………………… k … k-1 ……… … k-d2-dk-d2-d

CCGRID 2004 EOS Lab, National Tsing Hua University Construction of Lookup Table Lookup table is a collection of lookup pairs Get lookup pairs (target, forwarder) from lookup trees  forwarder is the next node in the lookup path to the target node Group lookup pairs to reduce the number of entries in a lookup table

CCGRID 2004 EOS Lab, National Tsing Hua University Fig 3.(b) Fig 3.(a) Node 0 Target 7 Forwarder 4 Node 0 Target 1 Forwarder 1 (1) There is a lookup pair (7,4) for node 0 (2) There is a lookup pair (1,1) for node 0

CCGRID 2004 EOS Lab, National Tsing Hua University Node 0 Lookup Table Take the lookup table of node 0 as an example Collect and group lookup pairs for node 0  Reduce the entries from 16 to 7 TargetForwarder TargetForwarder EntryTarget range forwarder 0(15,0]0 1(0,1]1 2(1,2]2 3(2,3]3 4(3,8]4 5(8,12]8 6(12,15]12

CCGRID 2004 EOS Lab, National Tsing Hua University Generic Lookup Table (1/3) All lookup tables can be constructed by the generic lookup table  without examining any lookup tree Distance tree  Used to show the lookup pairs in distance relationship  Constructed by replacing each node in balanced lookup tree with a pair (D2T,D2F) D2T : the distance to target node (root – node) D2F : the distance to forwarder node (parent – node)

CCGRID 2004 EOS Lab, National Tsing Hua University (0,0) … (1,1) (2,2) (d,d)(d,d) (d 2 +d, d 2 ) … (d 2 +1,d 2 ) (d+1,d) (2d+1,2d) ………… (d+2,d)(2d+2,2d)(d 2 +2, d 2 ) (2d,d) (3d,2d) … ………………………………… (d 2 +d+1, d 2 ) Level Distance Tree (2/3) Any pair of (D2T, D2F) is independent of k  there is only one unique distance tree  there are N different lookup trees

CCGRID 2004 EOS Lab, National Tsing Hua University Generic Lookup Tree (3/3) Group D2T and D2F to get generic lookup table A node adds its SID to generic lookup table to establish its own lookup table i = 1, 2, 3, …, d total entries = l * i = O(log d N) * d = O(d log d N)

CCGRID 2004 EOS Lab, National Tsing Hua University Self-organized Mechanism The lookup tables need to be updated each time a node joins, leaves or fails All lookup tables are aware of the system change in O(log N) hops  The updating message is forwarded along a balanced lookup tree A new lookup table only relies on the old lookup table of its own & its adjacent nodes  The lookup tables of adjacent nodes are offset by 1  SID to VID mapping is resolved locally

CCGRID 2004 EOS Lab, National Tsing Hua University Experimental Results Compared SCALLOP with a binomial-tree lookup protocol (e.g., Chord, Pastry, etc)  Hot spot avoidance  Balanced-load distribution  Customized performance  Self-organized mechanism

CCGRID 2004 EOS Lab, National Tsing Hua University Structure Comparison Our lookup tree distributes lookup requests evenly Eliminate the hot spot problem on node 15 binomial-tree structurebalanced-tree structure Hot Spot

CCGRID 2004 EOS Lab, National Tsing Hua University Balanced-Load Distribution Environment  10,000 nodes  10 6 lookup requests  Hot spot scenario: 30% of the requests ask for a data item stored in node 0

CCGRID 2004 EOS Lab, National Tsing Hua University Balanced-Load Distribution (Cont.) Results  Our protocol reduces about 50% load on node 9999 (a hot-spot node in binomial lookup protocol)  Flatter distribution with or without hot spots (a) Without hot spots(b) Hot spots with 30% concentration

CCGRID 2004 EOS Lab, National Tsing Hua University Customized Performance  SCALLOP allows customizable performance by changing its degree of the lookup tree

CCGRID 2004 EOS Lab, National Tsing Hua University Environment  Assume each hop takes one time unit  1,000 nodes  1,000 lookup requests per time unit  Nodes fail scenario: 5% nodes fail per 10 time unit from time 50 to 100 in a 200 time unit simulation Self-organized Mechanism

CCGRID 2004 EOS Lab, National Tsing Hua University Self-organized Mechanism Results  Average lookup path length grows as fault rate increases  No impact on path length once fault rate is below 20%  System is stabilized quickly and eventually

CCGRID 2004 EOS Lab, National Tsing Hua University Conclusions & Future Work We develop a lookup protocol for peer-to-peer distributed system  Scalable: O(d log d N) lookup table  High-performance: O(log d N) lookup path  Balanced-load: even data and lookup distribution Future work  More experimental results  Implementation on PlanetLab distributed system.

CCGRID 2004 EOS Lab, National Tsing Hua University Question?