Ahmad Al-Shishtawy 1,2,Tareq Jamal Khan 1, and Vladimir Vlassov 1 1 - KTH Royal Institute of Technology, Stockholm, Sweden {ahmadas, tareqjk,

Slides:



Advertisements
Similar presentations
Efficient Event-based Resource Discovery Wei Yan*, Songlin Hu*, Vinod Muthusamy +, Hans-Arno Jacobsen +, Li Zha* * Chinese Academy of Sciences, Beijing.
Advertisements

P2P data retrieval DHT (Distributed Hash Tables) Partially based on Hellerstein’s presentation at VLDB2004.
Ion Stoica, Robert Morris, David Karger, M. Frans Kaashoek, Hari Balakrishnan MIT and Berkeley presented by Daniel Figueiredo Chord: A Scalable Peer-to-peer.
Peer to Peer and Distributed Hash Tables
Dynamo: Amazon’s Highly Available Key-value Store ID2210-VT13 Slides by Tallat M. Shafaat.
Scalable Content-Addressable Network Lintao Liu
PNUTS: Yahoo!’s Hosted Data Serving Platform Brian F. Cooper, Raghu Ramakrishnan, Utkarsh Srivastava, Adam Silberstein, Philip Bohannon, HansArno Jacobsen,
Replication. Topics r Why Replication? r System Model r Consistency Models r One approach to consistency management and dealing with failures.
Consistency and Replication Chapter 7 Part II Replica Management & Consistency Protocols.
CHORD – peer to peer lookup protocol Shankar Karthik Vaithianathan & Aravind Sivaraman University of Central Florida.
Chord: A scalable peer-to- peer lookup service for Internet applications Ion Stoica, Robert Morris, David Karger, M. Frans Kaashock, Hari Balakrishnan.
Dynamo: Amazon's Highly Available Key-value Store Distributed Storage Systems CS presented by: Hussam Abu-Libdeh.
PNUTS: Yahoo’s Hosted Data Serving Platform Jonathan Danaparamita jdanap at umich dot edu University of Michigan EECS 584, Fall Some slides/illustrations.
PNUTS: Yahoo!’s Hosted Data Serving Platform Yahoo! Research present by Liyan & Fang.
SPANStore: Cost-Effective Geo-Replicated Storage Spanning Multiple Cloud Services Zhe Wu, Michael Butkiewicz, Dorian Perkins, Ethan Katz-Bassett, Harsha.
Small-world Overlay P2P Network
Shivkumar Kalyanaraman Rensselaer Polytechnic Institute 1 ECSE-6600: Internet Protocols Informal Quiz #13: P2P and Sensor Networks Shivkumar Kalyanaraman:
Distributed Systems Fall 2010 Replication Fall 20105DV0203 Outline Group communication Fault-tolerant services –Passive and active replication Highly.
Peer to Peer File Sharing Huseyin Ozgur TAN. What is Peer-to-Peer?  Every node is designed to(but may not by user choice) provide some service that helps.
P2P: Advanced Topics Filesystems over DHTs and P2P research Vyas Sekar.
Distributed Systems Fall 2009 Replication Fall 20095DV0203 Outline Group communication Fault-tolerant services –Passive and active replication Highly.
Chord-over-Chord Overlay Sudhindra Rao Ph.D Qualifier Exam Department of ECECS.
SkipNet: A Scaleable Overlay Network With Practical Locality Properties Presented by Rachel Rubin CS294-4: Peer-to-Peer Systems By Nicholas Harvey, Michael.
1 CS 194: Distributed Systems Distributed Hash Tables Scott Shenker and Ion Stoica Computer Science Division Department of Electrical Engineering and Computer.
Inexpensive Scalable Information Access Many Internet applications need to access data for millions of concurrent users Relational DBMS technology cannot.
PNUTS: YAHOO!’S HOSTED DATA SERVING PLATFORM FENGLI ZHANG.
Dynamo: Amazon’s Highly Available Key-value Store Giuseppe DeCandia, et.al., SOSP ‘07.
Data Consistency in the Structured Peer-to-Peer Network Cheng-Ying Ou, Polly Huang Network and Systems Lab 台灣大學電機資訊學院電機所.
CSE 486/586, Spring 2012 CSE 486/586 Distributed Systems Case Study: Amazon Dynamo Steve Ko Computer Sciences and Engineering University at Buffalo.
Peer-to-Peer in the Datacenter: Amazon Dynamo Aaron Blankstein COS 461: Computer Networks Lectures: MW 10-10:50am in Architecture N101
INTRODUCTION TO PEER TO PEER NETWORKS Z.M. Joseph CSE 6392 – DB Exploration Spring 2006 CSE, UT Arlington.
Distributed Systems Tutorial 11 – Yahoo! PNUTS written by Alex Libov Based on OSCON 2011 presentation winter semester,
Thesis Proposal Data Consistency in DHTs. Background Peer-to-peer systems have become increasingly popular Lots of P2P applications around us –File sharing,
Fault Tolerance via the State Machine Replication Approach Favian Contreras.
On P2P Collaboration Infrastructures Manfred Hauswirth, Ivana Podnar, Stefan Decker Infrastructure for Collaborative Enterprise, th IEEE International.
A Distributed Architecture for Multi-dimensional Indexing and Data Retrieval in Grid Environments Athanasia Asiki, Katerina Doka, Ioannis Konstantinou,
High Throughput Computing on P2P Networks Carlos Pérez Miguel
Chord: A Scalable Peer-to-peer Lookup Protocol for Internet Applications Xiaozhou Li COS 461: Computer Networks (precept 04/06/12) Princeton University.
VICTORIA UNIVERSITY OF WELLINGTON Te Whare Wananga o te Upoko o te Ika a Maui SWEN 432 Advanced Database Design and Implementation Trade-offs in Cloud.
Alireza Angabini Advanced DB class Dr. M.Rahgozar Fall 88.
1 Distributed Hash Tables (DHTs) Lars Jørgen Lillehovde Jo Grimstad Bang Distributed Hash Tables (DHTs)
VICTORIA UNIVERSITY OF WELLINGTON Te Whare Wananga o te Upoko o te Ika a Maui SWEN 432 Advanced Database Design and Implementation Data Versioning Lecturer.
Chord: A Scalable Peer-to-peer Lookup Service for Internet Applications.
1 ACTIVE FAULT TOLERANT SYSTEM for OPEN DISTRIBUTED COMPUTING (Autonomic and Trusted Computing 2006) Giray Kömürcü.
Enabling Self-management Of Component Based Distributed Applications Ahmad Al-Shishtawy 1, Joel Höglund 2, Konstantin Popov 2, Nikos Parlavantzas 3, Vladimir.
Chord: A Scalable Peer-to-peer Lookup Service for Internet Applications Ion Stoica, Robert Morris, David Karger, M. Frans Kaashoek, Hari Balakrishnan Presented.
Scalable Content- Addressable Networks Prepared by Kuhan Paramsothy March 5, 2007.
Rendezvous Regions: A Scalable Architecture for Service Location and Data-Centric Storage in Large-Scale Wireless Sensor Networks Karim Seada, Ahmed Helmy.
MobiQuitous 2007 Towards Scalable and Robust Service Discovery in Ubiquitous Computing Environments via Multi-hop Clustering Wei Gao.
PROP: A Scalable and Reliable P2P Assisted Proxy Streaming System Computer Science Department College of William and Mary Lei Guo, Songqing Chen, and Xiaodong.
Authors Brian F. Cooper, Raghu Ramakrishnan, Utkarsh Srivastava, Adam Silberstein, Philip Bohannon, Hans-Arno Jacobsen, Nick Puz, Daniel Weaver, Ramana.
1 Distributed Hash Table CS780-3 Lecture Notes In courtesy of Heng Yin.
Chord Advanced issues. Analysis Search takes O(log(N)) time –Proof 1 (intuition): At each step, distance between query and peer hosting the object reduces.
Plethora: Infrastructure and System Design. Introduction Peer-to-Peer (P2P) networks: –Self-organizing distributed systems –Nodes receive and provide.
Replication (1). Topics r Why Replication? r System Model r Consistency Models r One approach to consistency management and dealing with failures.
Chapter 4 Wenbing Zhao Department of Electrical and Computer Engineering Cleveland State University Building Dependable Distributed Systems.
Stefanos Antaris A Socio-Aware Decentralized Topology Construction Protocol Stefanos Antaris *, Despina Stasi *, Mikael Högqvist † George Pallis *, Marios.
Click to edit Master title style Multi-Destination Routing and the Design of Peer-to-Peer Overlays Authors John Buford Panasonic Princeton Lab, USA. Alan.
Bruce Hammer, Steve Wallis, Raymond Ho
Content Delivery Networks: Status and Trends Speaker: Shao-Fen Chou Advisor: Dr. Ho-Ting Wu 5/8/
NCLAB 1 Supporting complex queries in a distributed manner without using DHT NodeWiz: Peer-to-Peer Resource Discovery for Grids Sujoy Basu, Sujata Banerjee,
An Efficient Implementation of File Sharing Systems on the Basis of WiMAX and Wi-Fi Jingyuan Li, Liusheng Huang, Weijia Jia, Mingjun Xiao and Peng Du Joint.
Malugo – a scalable peer-to-peer storage system..
TRUST Self-Organizing Systems Emin G ü n Sirer, Cornell University.
Chord: A Scalable Peer-to-Peer Lookup Service for Internet Applications * CS587x Lecture Department of Computer Science Iowa State University *I. Stoica,
TensorFlow– A system for large-scale machine learning
Lecturer : Dr. Pavle Mogin
Plethora: Infrastructure and System Design
Providing Secure Storage on the Internet
Global Distribution.
Presentation transcript:

Ahmad Al-Shishtawy 1,2,Tareq Jamal Khan 1, and Vladimir Vlassov KTH Royal Institute of Technology, Stockholm, Sweden {ahmadas, tareqjk, 2 - Swedish Institute of Computer Science, Stockholm, Sweden ICPADS 2011 Dec 7-9, Tainan, Taiwan

 Introduction  Majority-Based Key-Value Store  Evaluation  Conclusions

 Web 2.0 applications  WiKis, social networks, media sharing  Rapidly growing number of users  User generated data  New challenges for storage infrastructure  Scalable storage system (growing number of users)  Lower response time (uneven load, user geographically scattered)  Highly available (partial failure, large number of concurrent requests)  Some guarantee of data consistency  Privacy of user generated data

 Consistency  Availability  Partition-tolerance

Network partition is unavoidable

Availability & Performance Data Consistency

 Web 2.0 applications can tolerate relaxed consistency  Instead focuse on higher availability and lower response time

 Distributed data stores  Help developers to predict the results of read/write operations  Some consistency models:  Sequential consistency: all reads and writes appear as if executed sequentially  Eventual consistency: writes can be seen in different order for different readers.  Timeline consistency: writes are ordered but reads might return stale value  NoSQL Data Stores:  Amazon’s Dynamo  Yahoo!’s PNUTS

 Master based approach (all updates are sent to master replica)  Mastership assigned on per record basis  Asynchronous propagation of updates using pub-sub system Yahoo! Message Broker (YMB)  Per-record timeline consistency V. 2.1 indicates Generation = 2, Version =1

 Distributed network formed between nodes on the edge of the internet  Scalable and robust  Structured P2P  Structure of Overlay links  Efficient lookup service  Symmetric Replication Scheme  Used to calculate the IDs of replicas  Routing layer vs. Data layer

 Key-Value Object Store  Based on P2P network  Owned and controlled by the users  Large-scale  Fault-tolerant  Read/Write operations  Various consistency guaranties  API similar to Yahoo! PNUTS  Decentralized implementation  Based on majority voting  Better consistency guaranties than classical DHTs  Not as expensive as Paxos based replication

 Introduction  Majority-Based Key-Value Store  Evaluation  Conclusions

 We provide API similar to Yahoo!’s PNUTS  Read Any (key)  Read Critical (key, version)  Read Latest (key)  Write (key, data)  Test and Set Write (key, data, version)  We provide decentralized implementation  Tolerate high level of churn Applications can choose operations according to their performance, availability and consistency requirements.

 Avoids master-based approach adopted in PNUTS  Majority-based quorum technique to ensure data consistency WR Write to a majority of nodes will always overlap with a read from a majority

Classical DHT Majority- Based PAXOS … + P2P = works more than 99% This is because lookup inconsistencies WR  Stronger consistency than what DHTs offer  But weaker than PAXOS based systems  Guarantees data consistency with a probability of more than 99%  Suitable for Web 2.0 apps

PNUTS  Master-based  Targets a more stable & controlled environment comprising of handful data centers  If an entire region goes down, availability suffers due to master based approach  User data prone to company manipulation and privacy violation Majority-Based K-V Store  Majority-based  Works in highly dynamic environment.  Benefits from scalability, fault-tolerance and self- management properties inherent in P2P systems  Completely distributed, doesn’t depend on master availability  Relieves a company from hosting and managing data

 Consistency Layer  Implements the five read/write operations  DHT Layer  Get data/version  Put data/lock/unlock  Data relocation/recovery during node join, leave, failure  Chord Layer  Lookup key  Enables nodes to join, leave  Periodic stabilization

ReadLatest(x) Symmetric Replication & Lookup to find Responsible nodes Send Read Requests Receive Responses Call returns largest version after receiving from a majority x = x 6 x2x2 x6x6 x6x6 x6x6 x5x5

 Introduction  Majority-Based Key-Value Store  Evaluation  Conclusions

 Initially 100 nodes join the system  Default replication degree 5  Experiment run for 24 simulation hours (not including node joining time during warm up)  Request originating node and requested data identifiers are drawn from uniform distribution.

 5 experiements ExperimentProperty to Test Varying ChurnDynamism Varying Request RateScalability/Availability Varying Network SizeScalability Varying Replication degreeFault Tolerance/Availability Varying Read-Write RatioRequest Inter-dependency

 Introduction  Majority-Based Key-Value Store  Evaluation  Conclusions

 presented a majority-based key-value store  architecture, algorithms, and evaluation  P2P to reduce costs and improve data privacy  provides a number of read/write operations with multiple consistency levels  achieve robustness and withstand churn in a dynamic environment  Evaluation by simulation has shown that the system performs rather well in terms of latency and operation success ratio in the presence of churn