Presentation is loading. Please wait.

Presentation is loading. Please wait.

Unstructured P2P Networks Achieving Robustness and Scalability in Data Dissemination Scenarios Michael Mirold Seminar on Advanced Topics in Distributed.

Similar presentations

Presentation on theme: "Unstructured P2P Networks Achieving Robustness and Scalability in Data Dissemination Scenarios Michael Mirold Seminar on Advanced Topics in Distributed."— Presentation transcript:

1 Unstructured P2P Networks Achieving Robustness and Scalability in Data Dissemination Scenarios Michael Mirold Seminar on Advanced Topics in Distributed Computing 2007/2008

2 2 Introduction Structured P2P  Node graph with predefined structure  Examples –Chord –Pastry –CAN Unstructured P2P  Random node graph  Examples –Gnutella –BitTorrent

3 3 Part I Do Ut Des, Tit-for-Tat or “How Leech Proof is BitTorrent Really?” Bases on papers: “Incentives Build Robustness in BitTorrent”, Bram Cohen “Do Incentives Build Robustness in BitTorrent?”, Michael Piatek, Tomas Isdal, Thomas Anderson, Arvind Krishnamurthy, Arun Venkataramani

4 4 What is BitTorrent?  Used for distribution of single files  File is split into pieces (32kB – 2MB)  Pieces are distributed within a swarm –denotes all nodes that are interested in the file  Downloaded pieces are redistributed  No single “server”: True peer-to-peer (except perhaps of tracker)

5 5 How Does BitTorrent Work? (1) Want really badly to make everyone enjoy my new holiday pictures (HDR  some GB) 1.2 Put Torrent-File onto web server Initial Seed Tracker “Ordinary” Web Server 1.3 Register as “downloader” 1.1 Create Torrent-File Torrent File - url of tracker - name: name of file - piece length - pieces (conc. of SHA1 hashes of all pieces) - file length

6 6 How Does BitTorrent Work? (2) Initial Seed Tracker 2.1 Request Torrent-File “Ordinary” Web Server 2.3 Send peer set (“Local Neighborhood”) 2.2 Register as downloader 2.4 Open connection 2.7 Send pieces 2.5 Handshake 2.6 Request

7 7 B choke At some Unspecial Moment… seed swarm non-seed our node active set peer set active set statically sized

8 8 BitTorrent in Action

9 9 Achieving Fairness in BitTorrent  Basis: Tit-for-Tat strategy –Famous in game theory (considered “best” strategy for winning Prisoners’ Dilemma) –Idea: Be cooperative to others Penalize defective behaviour but don’t be too unforgiving  Put in BitTorrent context –Grant upload capacity to n best uploaders n: size of active set + # of optimistic unchokes –“Choke”, i.e. stop uploading to, peers that don’t perform well recompute choking every 10 seconds –However: “Optimistically unchoke” peers twice every 30 seconds  Reward good uploaders with capacity

10 10 Sounds good in theory, but…  what if you have LOTS of upload capacity? –most of your peers are slower (see later) –nevertheless you must choose a few fastest peers probably already done –AND you have split your upload capacity equally!  you give your capacity away for free  this is called “ALTRUISM”  Altruism THE source of unfairness in BitTorrent  Unfair peers only need slightly “better” than average 2.prefer rich girls, i.e. highly altruistic peers remember: active set has static size

11 11 Unfairness/Altruism Illustrated Measure of altruism

12 12 Unfairness/Altruism Illustrated (2) Altruism as wasted upload capacity slow clients never get reciprocated  every byte is wasted fast clients contribute more than necessary

13 13 Real World Observations Reference implementation uses active set size of This is what you are competing against  be a bit faster! If you can upload > 14kB/s reciprocation prob > 99% !!

14 14 The Optimal Active Set Size * * for a peer with 300 kB/s UL capacity Optimal Set Size

15 15 BitTyrant – A Selfish BitTorrent Client  Based on Azureus Client –publicly available at  Exploiting the unfairness in BitTorrent  Minimizing altruism to improve efficiency  Mechanisms: –choose only “best” peers with respect to UL/DL ratio (see next slide) –deviate from equal split –optimize active set size

16 16 Peer Choice Optimization Algorithm  Step invariant: Maintain u p, d p of peer p –d p : Estimated download performance from p –u p : Estimated upload needed for reciprocation with p  Initially: set according to theor. distribution  Each step: Rank order peers according to and choose best peers for unchoking  After each round: –Has p unchoked us?  –Has p not unchoked us?  –Has p unchoked us for the last r rounds? 

17 17 Experiences using BitTyrant  Multiple BitTyrant peers: –depends on a few factors and not easily comparable –strategic, i.e. peers use adapted choking algorithm  swarm performance improved compared to BitTorrent –strategic & selfish, i.e. peer doesn’t give back excess capacity  swarm performance decreases dramatically  One BitTyrant peer: –median: 72%

18 18 Personal Opinion  Paper shows nice “hack”  Paper shows that there is no perfect fairness  Paper shows a sensible optimization  But: I think, model is too restricted –People’s goals unconsidered Altruistic people are often just that: altruistic (they don’t mind performing suboptimal) Everyone glad with BitTorrent, why optimize?  And: will this paper make the world a better place?

19 19 Part II Getting Almost Reliable Broadcast with Almost no Pain: “Lightweight Probabilistic Broadcast” Bases on papers: “Lightweight Probabilistic Broadcast”, 2003, Eugster, Guerraoui, Handurukande & Kouznetsov

20 20 Background & Motivation  Large scale event dissemination –Processes p 1, …, p n subscribe for topic t –Event e with topic t is delivered to p 1, …, p n  Reliable Broadcast –scales poorly  Network level Broadcast/Multicast –lacks reliability guarantees –also scalability problems  Complete view on the network  leads to unsustainable memory demands

21 21 The lpbcast Protocol  System contains n processes Π = {p 1, …, p n } –dynamically joining and leaving  Processes subscribe to a single topic –easily extendible to multiple topics –joining/leaving == subscribing/unsubscribing  Gossip sent to F random nodes in view i of process p i –F is “fanout” of process –view i is subset of procs currently known by p i  Gossips sent out periodically (non-synchronized)

22 22 Gossips  gossip: all-in-one record containing – event notifications –gossip.subs: subscriptions, –gossip.unsubs: unsubscriptions, –gossip.eventIds: history of events ids received so far  Containers don’t contain duplicates, i.e. they are set-like lists

23 23 Processes  Every process p has several buffers: –view (fixed maximum size l): contains processes that are known to / seen by p –subs (fixed maximum size |subs| M ): contains subscriptions received by p –unsubs (fixed maximum size |unsubs| M ): contains unsubscriptions received by p –events (fixed maximum size |events| M ): contains event notifications since last gossip emission –eventIds (fixed maximum size |eventIds| M ): ids of events seen so far –retrieveBuf: contains ids of events seen in gossip.eventids but not known

24 24 Example Fanout: 3 |view| M = 8 Process p i view (p i )

25 25 lpbcast procedures  Upon receiving a gossip message 1.Update view and unSubs with unsubscriptions 2.Update view with new subscriptions 3.Update events with new notifications 1.deliver notifications 2.update retrieveBuf for later retrieval of notifications 4.Perform housekeeping (truncate containers)  When sending a gossip message –fill gossip message accordingly

26 26 Example: Zeit T0 12 view2 subs2 unSubs events eventId subs1, 2 unSubs events eventId GOSSIP 3 view2 Subs2 unSubs eventse3 eventId view3 subs3 unSubs eventse3, e4 eventIde3,e4 subs3, 2 unSubs eventse3 eventIde3 GOSSIP subs3, 2 unSubs eventse3, e4 eventIde3, e4 GOSSIP

27 27 Example: Zeit T1 1 2 view2 subs2 unSubs events eventId subs1, 2 unSubs eventse18 eventId GOSSIP 3 view2 Subs2 unSubs events eventIde3, e4 view1,3 subs1,3 unSubs eventse5 eventIde3,e4,e5 subs3, 2 unSubs events eventIde3, e4 GOSSIP subs1,2,3 unSubs eventse5 eventIde3, e4,e5 GOSSIP subs1,2,3 unSubs eventse5 eventIde3,e4, e5 GOSSIP

28 28 Analytical Evaluation of lpbcast  Assumptions: –Π constant during evaluation –synchronous rounds –upper bound on latency –identical fanout F for all processes –probability of message loss ε –probability of crash τ –random view ind. uniformly distributed

29 29 Analytical Evaluation  Turns out that throughput is independent from view size l –provided that views are uniformly distributed  Membership stability (no partitioning) –increases with growing view size and/or system size –partitioning probability increases slowly with rounds: rounds for n=50, l=3

30 30 Practical Observations  Throughput does depend a bit on l  Explanation: –views not as uniformly distributed as assumed

31 31 Time to Infect all Processes  Simulation meets measurements pretty well  Fanout 3  1 msg injected  System size varies

32 32 Thank you for listening!

Download ppt "Unstructured P2P Networks Achieving Robustness and Scalability in Data Dissemination Scenarios Michael Mirold Seminar on Advanced Topics in Distributed."

Similar presentations

Ads by Google