Presentation is loading. Please wait.

Presentation is loading. Please wait.

Contribution to the Design & Implementation of the Highly Available Scalable and Distributed Data Structure: LH* RS Rim Moussa

Similar presentations


Presentation on theme: "Contribution to the Design & Implementation of the Highly Available Scalable and Distributed Data Structure: LH* RS Rim Moussa"— Presentation transcript:

1 Contribution to the Design & Implementation of the Highly Available Scalable and Distributed Data Structure: LH* RS Rim Moussa Rim.Moussa@dauphine.frhttp://ceria.dauphine.fr/rim/rim.html Thesis Presentation in Computer Science *Distributed Databases Thesis Supervisor: Pr. Witold Litwin Examinators: Pr. Thomas J.E. Schwarz Pr. Toré Risch Jury President: Pr. Gérard Lévy Paris Dauphine University *CERIA Lab. *04th October 2004

2 04 Oct. 04 * Présentation de Thèse R. Moussa, U. Paris Dauphine 2 Outline 1. Issue 2. State of the Art 3. LH* RS Scheme 4. LH* RS Manager 5. Experimentations 6. LH* RS File Creation 7. Bucket Recovery 8. Parity Bucket Creation 9. Conclusion & Future Work

3 04 Oct. 04 * Présentation de Thèse R. Moussa, U. Paris Dauphine 3 Facts …  Volume of Information of 30% /year  Technology  Network Infrastructure >> Gilder Law, bandwidth triples every year.  Evolution of PCs storage & computing capacities >> Moore Law, the latters double every 18 months.  Bottleneck of Disks Accesses & CPUs Need of Distributed Data Storage Systems SDDSs: LH*, RP* …  High Throughput

4 04 Oct. 04 * Présentation de Thèse R. Moussa, U. Paris Dauphine 4 Facts … Network  Frequent & Costly Failures >> Stat. Published by the Contingency Planning Research in 1996: the cost of service interruption/h case of brokerage application is $6,45 million. Need of Distributed & Highly-Available Data Storage Systems  Multicomputers >> Modular Architecture >> Good Price/ Performance Tradeoff

5 04 Oct. 04 * Présentation de Thèse R. Moussa, U. Paris Dauphine 5 State of the Art Parity Calculus (+) (+) Good Response Time, Mirors are functional (-) High Storage Overhead (  n if n repliquas) Data Replication Criteria to evaluate Erasure-resilient Codes: Encoding Rate (Parity Volume/ Data Volume) Update Penality (Parity Volumes) Group Size used for Data Reconstruction Encoding & Decoding Complexity Recovery Capabilitties

6 04 Oct. 04 * Présentation de Thèse R. Moussa, U. Paris Dauphine 6 Parity Schemes 1-Available Schemes k-Available Schemes Binary Linear Codes: [H94]  Tolerate max. 3 failures Array Codes: EVENODD [B94 ], X-code [XB99], RDP [C+04]  Tolerate max. 2 failures Reed Solomon Codes : IDA [R89], RAID X [W91], FEC [B95], Tutorial [P97], LH* RS [LS00, ML02, MS04, LMS04]  Tolerate k failures (k > 3) … XOR Parity Calculus : RAID Technology (level 3, 4, 5…) [PGK88], SDDS LH*g [L96] …

7 04 Oct. 04 * Présentation de Thèse R. Moussa, U. Paris Dauphine 7 Outline… 1. Issue 2. State of the Art 3. LH* RS Scheme LH* RS ? SDDSs? Reed Solomon Codes? Encoding/ Decoding Optimizations 4. LH* RS Manager 5. Experimentations

8 04 Oct. 04 * Présentation de Thèse R. Moussa, U. Paris Dauphine 8 LH* RS ? Distribution using Linear Hashing (LH* LH [KLR96]) LH* LH Manager[B00] Scalability & High Throughput High Availability LH*: Scalable & Distributed Data Structure Parity Calculus using Reed-Solomon Codes [RS63] LH* RS [LS00]

9 04 Oct. 04 * Présentation de Thèse R. Moussa, U. Paris Dauphine 9 SDDSs Principles (1) Dynamic File Growth Client Network Client … Data Buckets … OVERLOADED You Split Insertions … Coordinator Record Transfert

10 04 Oct. 04 * Présentation de Thèse R. Moussa, U. Paris Dauphine 10 SDDSs Principles (2) Network (2) No Centralized Directory Access Cases de Données …… Client Query Query Forward Client Image Adjustment Message … File Image

11 04 Oct. 04 * Présentation de Thèse R. Moussa, U. Paris Dauphine 11 Reed-Solomon Codes  Encoding From m Data Symbols  Calculus of n Parity Symbols  Data Representation  Galois Field Fields with finite size: q Closure Propoerty: Addition, Substraction, Multiplication, Division. In GF(2 w ), (1) Addition (XOR) (2) Multiplication (Tables: gflog and antigflog) e 1 * e 2 = antigflog[ gflog[e 1 ] + gflog[e 2 ] ]

12 04 Oct. 04 * Présentation de Thèse R. Moussa, U. Paris Dauphine 12 10000…0 C 1,1 … C 1,j … C 1,n-m 01000 …0 C 2,1 … C 2,j … C 2,n-m 00100 …0 C 3,1 … C 3,j … C 3,n-m … … … … … 0 0000 …1 C m,1 … C m,j … C m,n-m RS Encoding S1S2S3:Si:SmS1S2S3:Si:Sm  S 1 : S m P 1 P 2 : P j : P n-m = C 1,j C 2,j C 3,j : C m,j PjPj (S 1  C 1,j )  (S 2  C 2,j )  …  (S m  C m,j ) m-1 XORs GF m Multiplications GF S1S2S3:Si:SmS1S2S3:Si:Sm ImIm P(m  (n-m)) (1) Systematic Encoding: Matrix (Im|P) (2) Any m columns are linearly independent Parity Matrix

13 04 Oct. 04 * Présentation de Thèse R. Moussa, U. Paris Dauphine 13 Optimized Decoding Multiply the ‘‘m OK symbols’’ By columns of H -1 corresponding to lost symbols m OK symbols H m : m corresponding columns  H -1 = [ S 1 S 2 S 3 S 4 ….. S m ] Gauss Transformatiom 10000…0C 1,1 C 1,2 C 1,3 … C 1,n-m 01000 …0 C 2,1 C 2,2 C 2,3 … C 2,n-m 00100 …0 C 3,1 C 3,2 C 3,3 … C 3,n-m … … … … … 0 00 0 0 …1 C m,1 C m,2 C m,3 … C m,n-m RS Decoding S 1 S 2 S 3 S 4 : S m P 1 P 2 P 3 : P n-m

14 04 Oct. 04 * Présentation de Thèse R. Moussa, U. Paris Dauphine 14 Galois Field Parity Matrix Optimizations GF Multiplication (+) GF(2 16 ) vs. GF(2 8 ) reduces the #Symbols by 1/2  #Operations in the GF. GF(2 8 )  1 symbol = 1 Byte GF(2 16 )  1 symbol = 2 Bytes (-) Multiplication Tables Size GF(2 8 ): 0,768 Ko GF(2 16 ): 393,216 Ko (512  0,768)

15 04 Oct. 04 * Présentation de Thèse R. Moussa, U. Paris Dauphine 15 Galois Field Parity Matrix Optimizations (2) GF Multiplication 1st Column of ‘1’s Encoding of the 1st PB along XOR Calculus  Gain in encoding & decoding 1st Row of ‘1’s Any update from 1st DB is processed with XOR Calculus  Gain in Performance of 4% (case PB creation, m =4) 0001 0001 0001 … 0001 eb9b 2284 … 0001 2284 9é74 … 0001 9e44 d7f1 … … …

16 04 Oct. 04 * Présentation de Thèse R. Moussa, U. Paris Dauphine 16 Galois Field Parity Matrix Optimizations (3) GF Multiplication Encoding Log Pre-calculus of the Coef. of P Matrix  Improvement of 3,5% 0000 0000 0000 … 0000 5ab5 e267 … 0000 e267 0dce … 0000 784d 2b66 … … … Decoding Log Pre-calculus of coef. of H -1 matrix and OK symbols vector  Improvement of 4% to 8% depending on the #buckets to recover Goal: Reduce GF Multiplication Complexity e 1 * e 2 = antigflog[ gflog[e 1 ] + gflog[e 2 ] ]

17 04 Oct. 04 * Présentation de Thèse R. Moussa, U. Paris Dauphine 17 LH* RS -Parity Groups   Data Buckets Parity Buckets      : Key; Data Insert Rank r  : Rank; [Key-list ]; Parity Key r 210 210 210 210 A k-Acvailable Group survive to the failure of k buckets Grouping Concept  m: #data buckets  k: #parity buckets

18 04 Oct. 04 * Présentation de Thèse R. Moussa, U. Paris Dauphine 18 Outline… 1. Issue 2. State of the Art 3. LH* RS Scheme 4. LH* RS Manager Communication Gross Architecture 5. Experimentations 6. File Creation 7. Bucket Recovery …

19 04 Oct. 04 * Présentation de Thèse R. Moussa, U. Paris Dauphine 19 Communication TCP/IPUDP“Multicast”  Individual Operations (Insert, Update, Delete, Search)  Record Recovery  Control Messages Performance

20 04 Oct. 04 * Présentation de Thèse R. Moussa, U. Paris Dauphine 20 Communication TCP/IPUDP“Multicast” Large Buffers Transfert  New Parity Buckets  Transfer Parity Update & Record (Bucket Split)  Bucket Recovery Performance & Reliability

21 04 Oct. 04 * Présentation de Thèse R. Moussa, U. Paris Dauphine 21 Communication TCP/IPUDP“Multicast” Looking for New Data/Parity Buckets Communication Multipoints

22 04 Oct. 04 * Présentation de Thèse R. Moussa, U. Paris Dauphine 22 Architecture (1)TCP/IP Connection Handler Principle of “Sending Credit & Message Conservation until delivery” [J88, GRS97, D01] 1 Bucket Recovery (3,125 MB ): SDDS 2000: 6,7 s SDDS2000-TCP: 2,6 s (Hardware Config.: CPU 733MhZ machines, network 100Mbps) Before  Improvement of 60% TCP/IP Connections are passive OPEN, RFC 793 – [ISI81], TCP/IP under Win2K Server OS [MB00] (2)Flow Control & Message Acknowledgement (FCMA) Enhancements to SDDS2000 Architecture:

23 04 Oct. 04 * Présentation de Thèse R. Moussa, U. Paris Dauphine 23 Architecture (2) Before To tag new servers (data or parity) using Multicast: (3)Dynamic IP Addressing Structure Pre-defined and Static IP@s Table Multicast Group of Blank Data Buckets Multicast Group of Blank Parity Buckets Coordinator Created Buckets

24 04 Oct. 04 * Présentation de Thèse R. Moussa, U. Paris Dauphine 24 Architecture (3) Multicast Listening Port UDP Sending Port TCP/IP Port UDP Listening Port UDP Listening Thread Messages Queue TCP Listening Thread Multicast listening Thread Message Queue Pool of Working Threads Network ACK Mgmt Threads  Free Zones Messages waiting for ACK. Not acquitted Messages … ACK Structure Multicast Working Thread

25 04 Oct. 04 * Présentation de Thèse R. Moussa, U. Paris Dauphine 25 Experimentation  Performance Evaluation * CPU Time *Communication Time  Experimental Environment *5 Machines (Pentium IV: 1.8 GHz, RAM: 512 Mb) *Ethernet Network 1 Gbps *O.S.: Win2K Server *Tested Configuration: 1 Client, A group of 4 Data Buckets, k Parity Buckets (k = 0,1,2,3).

26 04 Oct. 04 * Présentation de Thèse R. Moussa, U. Paris Dauphine 26 Outline… 1. Issue 2. State of the Art 3. LH* RS Scheme 4. LH* RS Manager 5. Experimentations 6. File Creation Parity Update Performance 7. Bucket Recovery 8. Parity Bucket Creation

27 04 Oct. 04 * Présentation de Thèse R. Moussa, U. Paris Dauphine 27 File Creation  Client Operations Propagation of Data Record Inserts/ Updates/ Deletes to Parity Buckets.  Update: Send only  –record.  Deletes: Management of Free Ranks within Data Buckets.  Data Bucket Split N1: #renaining records N2: #leaving records Parity Group of the Splitting Data Bucket N1+N2 Deletes + N1 Inserts Parity Group of the New Data Bucket N2 Inserts

28 04 Oct. 04 * Présentation de Thèse R. Moussa, U. Paris Dauphine 28 Performances Config. Client Window = 1 Client Window = 5 Max Bucket Size = 10 000 records File of 25 000 records 1 record = 104 Bytes No difference GF(2 8 ) et GF(2 16 ) (we don’t wait for ACKs between DBs and PBs)

29 04 Oct. 04 * Présentation de Thèse R. Moussa, U. Paris Dauphine 29 Performances Config. Client Window = 1 Client Window = 5 k = 0 ** k = 1  Perf. Degradation of 20% k = 1 ** k = 2  Perf. Degradation of 8%

30 04 Oct. 04 * Présentation de Thèse R. Moussa, U. Paris Dauphine 30 Performances Config. Client Window = 1 Client Window = 5 k = 0 ** k = 1  Perf. Degradation of 37% k = 1 ** k = 2  Perf. Degradation of 10%

31 04 Oct. 04 * Présentation de Thèse R. Moussa, U. Paris Dauphine 31 Outline… 1. Issue 2. State of the Art 3. LH* RS Scheme 4. LH* RS Manager 5. Experimentations 6. File Creation 7. Bucket Recovery Scenario Performances 8. Parity Bucket Creation

32 04 Oct. 04 * Présentation de Thèse R. Moussa, U. Paris Dauphine 32 Failure Detection Are you Alive? Data Buckets Parity Buckets  Scenario Coordinator 

33 04 Oct. 04 * Présentation de Thèse R. Moussa, U. Paris Dauphine 33 Waiting for Responses … OK Data Buckets Parity Buckets Scenario (2) OK Coordinator  

34 04 Oct. 04 * Présentation de Thèse R. Moussa, U. Paris Dauphine 34 Searching Spare Buckets … Wanna be Spare ? Scenario (3) Multicast Group of Blank Data Buckets Coordinator

35 04 Oct. 04 * Présentation de Thèse R. Moussa, U. Paris Dauphine 35 Waiting for Replies … Launch UDP Listening Launch TCP Listening, Launch Working Thredsl *Waiting for Confirmation* If Time-out elapsed  cancel everything I would Scenario (4) Multicast Group of Blank Data Buckets Coordinator I would

36 04 Oct. 04 * Présentation de Thèse R. Moussa, U. Paris Dauphine 36 Spare Selection Scenario (5) Multicast Group of Blank Data Buckets Confirmed Cancellation Confirmed You are Hired Coordinator

37 04 Oct. 04 * Présentation de Thèse R. Moussa, U. Paris Dauphine 37 Parity Buckets Recover Failed Buckets Scenario (6) Recovery Manager Selection Coordinator

38 04 Oct. 04 * Présentation de Thèse R. Moussa, U. Paris Dauphine 38 Data Buckets Parity Buckets Recovery Manager Spare Buckets Buckets participating to Recovery Send me Records of rank in [r, r+slice-1] … Scenario (7) Query Phase

39 04 Oct. 04 * Présentation de Thèse R. Moussa, U. Paris Dauphine 39 Decoding Phase Recovered Slices Data Buckets Parity Buckets Spare Buckets Buckets participating to Recovery Requested Buffers … Scenario (8) Reconstruction Phase Recovery Manager In // with Query Phase

40 04 Oct. 04 * Présentation de Thèse R. Moussa, U. Paris Dauphine 40 2 DBs1 DB XORConfig. 1 DB RS XOR vs. RS Performances  File Info File of 125 000 records Record Size = 100 bytes Bucket Size = 31250 records  3.125 MB Group of 4 Data Buckets (m = 4), k-Available with k = 1,2,3  Decoding * GF(2 16 ) * RS + Decoding (RS + log Pre-calculus of H -1 and OK Symboles Vector)  Recovery per Slice (adaptative to PCs storage & computing capacities)

41 04 Oct. 04 * Présentation de Thèse R. Moussa, U. Paris Dauphine 41 2 DBs1 DB XORConfig. 1 DB RS XOR vs. RS Performances Slice Total Time (sec) CPU Time (sec) Com. Time (sec) 12500,6250,2660,348 31250,5880,2550,323 62500,5520,2400,312 156250,5620,2550,302 312500,5780,2500,328  Slice (from 4% to 100% of a bucket content)  Total Time is almost constant 0,58

42 04 Oct. 04 * Présentation de Thèse R. Moussa, U. Paris Dauphine 42 2 DBs1 DB XORConfig. 1 DB RS XOR vs. RS Performances Slice Total Time (sec) CPU Time (sec) Com. Time (sec) 12500,7340,3490,365 31250,6880,3590,323 62500,6560,3540,297 156250,6670,3600,297 312500,6880,3600,328 0,67  Slice (from 4% to 100% of a bucket content)  Total Time is almost constant

43 04 Oct. 04 * Présentation de Thèse R. Moussa, U. Paris Dauphine 43 2 DBs1 DB XORConfig. Performances Time to Recover 1DB -XOR : 0,58 sec XOR in GF(2 16 ) realizes a gain of 13% in Total Time (and 30% in CPU Time) Time to Recover 1DB –RS : 0,67 sec 1 DB RS XOR vs. RS

44 04 Oct. 04 * Présentation de Thèse R. Moussa, U. Paris Dauphine 44 3 DBs2 DBsSummaryXOR vs. RS1 DB RS Performances Slice Total Time (sec) CPU Time (sec) Com. Time (sec) 12500,9760,5770,375 31250,9320,5890,338 62500,8830,5620,321 156250,8750,5620,281 312500,8750,5620,313 0,9  Slice (from 4% to 100% of a bucket content)  Total Time is almost constant

45 04 Oct. 04 * Présentation de Thèse R. Moussa, U. Paris Dauphine 45 3 DBs2 DBsSummaryXOR vs. RS1 DB RS Performances SliceTotal Time (sec) CPU Time (sec) Com. Time (sec) 12501,2810,8280,406 31251,2500,8280,390 62501,2110,8520,352 156251,1880,8230,361 312501,2030,8280,375 1,23  Slice (from 4% to 100% of a bucket content)  Total Time is almost constant

46 04 Oct. 04 * Présentation de Thèse R. Moussa, U. Paris Dauphine 46 Performances 3 DBs2 DBsSummaryXOR vs. RS1 DB RS f Bucket Size (MB) Total Time (sec) Recovery Speed ( MB/sec ) 1 (XOR) 1 (RS) 3,125 0,585.38 0,674.66 26,2500,96.94 39,3751,237,62 Time to Recover f Buckets  f  Time to Recover 1 Bucket Factorized Query Phase  The + is Decoding Time & Time to send Recovered Buffers

47 04 Oct. 04 * Présentation de Thèse R. Moussa, U. Paris Dauphine 47 Performances GF(2 8 )  XOR in GF(2 8 ) improves decoding perf. of 60% compared to RS in GF(2 8 ).  RS/RS+ decoding in GF(2 16 ) realize a gain of 50% compared to decoding in GF(2 8 ). 3 DBs2 DBsSummaryXOR vs. RS

48 04 Oct. 04 * Présentation de Thèse R. Moussa, U. Paris Dauphine 48 Outline… 1. Issue 2. State of the Art 3. LH* RS Scheme 4. LH* RS Manager 5. Experimentations 6. File Creation 7. Bucket Recovery 8. Parity Bucket Creation Scenario Performances

49 04 Oct. 04 * Présentation de Thèse R. Moussa, U. Paris Dauphine 49 Scenario Multicast Group of Blank Parity Buckets Wanna Join Group g ? Searching for a new Parity Bucket Coordinator

50 04 Oct. 04 * Présentation de Thèse R. Moussa, U. Paris Dauphine 50 Scenario (2) Coordinator I Would Launch UDP Listening Launch TCP Listening, Launch Working Thredsl *Waiting for Confirmation* If Time-out elapsed  cancel everything Waiting for Replies … Multicast Group of Blank Parity Buckets I Would

51 04 Oct. 04 * Présentation de Thèse R. Moussa, U. Paris Dauphine 51 Scenario (3) You are Hired Confirmed Cancellation New Parity Bucket Selection Multicast Group of Blank Parity Buckets Coordinator

52 04 Oct. 04 * Présentation de Thèse R. Moussa, U. Paris Dauphine 52 Send me your contents ! … Scenario (4) Group of Data Buckets New Parity Bucket … Auto-creation *Query Phase

53 04 Oct. 04 * Présentation de Thèse R. Moussa, U. Paris Dauphine 53 Requested Buffers … Scenario (5) Group of Data Buckets Buffer Processing … Auto-creation *Encoding Phase New Parity Bucket

54 04 Oct. 04 * Présentation de Thèse R. Moussa, U. Paris Dauphine 54 Performances Max Bucket Size : 5000.. 50000 records Bucket Load Factor: 62,5% Record Size: 100 octets Group of 4 Data Buckets Encoding  GF(2 16 )  RS++ ( Log Pre-calculus & Row ‘1’s  XOR encoding to Process 1st DB buffer) XORRS XOR vs. RS Config.GF(2 8 )

55 04 Oct. 04 * Présentation de Thèse R. Moussa, U. Paris Dauphine 55 Performances Bucket Size Total Time (sec) CPU Time (sec) Com. Time (sec) 50000.1900.1400.029 100000.4290.3040.066 250001.0070.7380.144 500002.0621.4840.322 XORRS XOR vs. RS Config.GF(2 8 ) Same Encoding Rate  Bucket Size: CPU Time  74% Total Time

56 04 Oct. 04 * Présentation de Thèse R. Moussa, U. Paris Dauphine 56 Performances Bucket Size Total Time (sec) CPU Time (sec) Com. Time (sec) 50000.1930.1490.035 100000.4460.3280.059 250001.0530.7660.153 500002.1031.5310.322 XORRS XOR vs. RS Config.GF(2 8 ) Same Encoding Rate  Bucket Size: CPU Time  74% Total Time

57 04 Oct. 04 * Présentation de Thèse R. Moussa, U. Paris Dauphine 57 Performances XOR encoding speed : 2.062 sec RS encoding speed: 2.103 sec XOR realizes a performance gain in CPU time of 5% (  only 0,02% on Total Time) For Bucket Size = 50000 records XORRS XOR vs. RS Config.GF(2 8 )

58 04 Oct. 04 * Présentation de Thèse R. Moussa, U. Paris Dauphine 58 XORRS XOR vs. RS Config.GF(2 8 ) Performances  Idem GF(2 16 ), CPU Time = 3/4 Total Time  XOR in GF(2 8 ) improves CPU Time by 22%

59 04 Oct. 04 * Présentation de Thèse R. Moussa, U. Paris Dauphine 59 Performance File Creation Rate 0.33MB/s for k = 0 0.25MB/s for k = 1 0.23MB/s for k = 2 Record Insert Time 0.29ms for k = 0 0.33ms for k = 1 0.36ms for k = 2 Bucket Recovery Rate 4.66MB/s from 1-unavailability 6.94MB/s from 2-unavailability 7.62MB/s from 3-unavailability Record Recovery Time About 1.3ms Key Search Time Individual> 0.24ms Bulk> 0.056ms Wintel P4, 1.8GHz, 1Gbps

60 04 Oct. 04 * Présentation de Thèse R. Moussa, U. Paris Dauphine 60 Conclusion Experiments prove: Optimizations Encoding/ Decoding Architecture  Impact on Performance Good Recovery Performances

61 04 Oct. 04 * Présentation de Thèse R. Moussa, U. Paris Dauphine 61 Future Work Update Propagation to Parity Buckets Reliability Performance Reduce Coordinator Tasks « Parity Declustering » Investigation of New Erausure-Resilient Codes

62 04 Oct. 04 * Présentation de Thèse R. Moussa, U. Paris Dauphine 62 References [PGK88] D. A. Patterson, G. Gibson & R. H. Katz, A Case for Redundant Arrays of Inexpensive Disks, Proc. of ACM SIGMOD Conf, pp.109-106, June 1988. [ISI81] Information Sciences Institute, RFC 793: Transmission Control Protocol (TCP) – Specification, Sept. 1981, http://www.faqs.org/rfcs/rfc793.htmlhttp://www.faqs.org/rfcs/rfc793.html [MB 00] D. MacDonal, W. Barkley, MS Windows 2000 TCP/IP Implementation Details, http://secinf.net/info/nt/2000ip/tcpipimp.html http://secinf.net/info/nt/2000ip/tcpipimp.html [J88] V. Jacobson, M. J. Karels, Congestion Avoidance and Control, Computer Communication Review, Vol. 18, No 4, pp. 314-329. [XB99] L. Xu & J. Bruck, X-Code: MDS Array Codes with Optimal Encoding, IEEE Trans. on Information Theory, 45(1), p.272-276, 1999. [CEG+ 04] P. Corbett, B. English, A. Goel, T. Grcanac, S. Kleiman, J. Leong, S. Sankar, Row-Diagonal Parity for Double Disk Failure Correction, Proc. of the 3 rd USENIX –Conf. On File and Storage Technologies, Avril 2004. [R89] M. O. Rabin, Efficient Dispersal of Information for Security, Load Balancing and Fault Tolerance, Journal of ACM, Vol. 26, N° 2, April 1989, pp. 335-348. [W91] P.E. White, RAID X tackles design problems with existing design RAID schemes, ECC Technologies, ftp://members.aol.com.mnecctek.ctr1991.pdfftp://members.aol.com.mnecctek.ctr1991.pdf [GRS97] J. C. Gomez, V. Redo, V. S. Sunderam, Efficient Multithreaded User-Space Transport for Network Computing, Design & Test of the TRAP protocol, Journal of Parallel & Distributed Computing, 40 (1) 1997.

63 04 Oct. 04 * Présentation de Thèse R. Moussa, U. Paris Dauphine 63 References (2) [BK+ 95] J. Blomer, M. Kalfane, R. Karp, M. Karpinski, M. Luby & D. Zuckerman, An XOR-Based Erasure- Resilient Coding Scheme, ICSI Tech. Rep. TR-95-048, 1995. [LS00] W. Litwin & T. Schwarz, LH* RS : A High-Availability Scalable Distributed Data Structure using Reed Solomon Codes, p.237-248, Proceedings of the ACM SIGMOD 2000. [KLR96] J. Karlson, W. Litwin & T. Risch, LH*LH: A Scalable high performance data structure for switched multicomputers, EDBT 96, Springer Verlag. [RS60] I. Reed & G. Solomon, Polynomial codes over certain Finite Fields, Journal of the society for industrial and applied mathematics, 1960. [P97] J. S. Plank, A Tutorial on Reed-Solomon Coding for fault-Tolerance in RAID-like Systems, Software– Practise & Experience, 27(9), Sept. 1997, pp 995- 1012, [D01] A.W. Diène, Contribution à la Gestion de Structures de Données Distribuées et Scalables, PhD Thesis, Nov. 2001, Université Paris Dauphine. [B00] F. Sahli Bennour, Contribution à la Gestion de Structures de Données Distribuées et Scalables, PhD Thesis, Juin 2000, Université Paris Dauphine. + Références: http://ceria.dauphine.fr/rim/theserim.pdf

64 04 Oct. 04 * Présentation de Thèse R. Moussa, U. Paris Dauphine 64 Publications [ML02] R. Moussa, W. Litwin, Experimental Performance Analysis of LH* RS Parity Management, Carleton Scientific Records of the 4th International Workshop on Distributed Data & Structure : WDAS 2002, p.87-97. [MS04] R. Moussa, T. Schwarz, Design and Implementation of LH* RS – A Highly- Available Scalable Distributed Data Structure, Carleton Scientific Records of the 6th International Workshop on Distributed Data & Structure: WDAS 2004. [LMS04] W. Litwin, R. Moussa, T. Schwarz, Prototype Demonstration of LH* RS : A Highly Available Distributed Storage System, Proc. of VLDB 2004 (Demo Session ) p.1289- 1292. [LMS04-a] W. Litwin, R. Moussa, T. Schwarz, LH* RS : A Highly Available Distributed Storage System, journal version submitted, under revision.

65 Thank You For Your Attention Questions ?


Download ppt "Contribution to the Design & Implementation of the Highly Available Scalable and Distributed Data Structure: LH* RS Rim Moussa"

Similar presentations


Ads by Google