Presentation is loading. Please wait.

Presentation is loading. Please wait.

Berkeley Data Analytics Stack (BDAS) Overview

Similar presentations


Presentation on theme: "Berkeley Data Analytics Stack (BDAS) Overview"— Presentation transcript:

1 Berkeley Data Analytics Stack (BDAS) Overview
Ion Stoica UC Berkeley March 7, 2013 UC BERKELEY

2 What is Big Data used For?
Reports, e.g., Track business processes, transactions Diagnosis, e.g., Why is user engagement dropping? Why is the system slow? Detect spam, worms, viruses, DDoS attacks Decisions, e.g., Decide what feature to add Decide what ad to show Block worms, viruses, … What do people use big data for? First, they use it to generate reports to track and better understand business processes, ransactions Second, they use it to diagnose and answer questions such as Why the user engagement dropping?, why is the system slow? Or to detect spam, worms, or DDoS attacks But most importantly they use it to make decisions, such us improving the business process, deciding what features to add to the product, deciding what ad to show, or, once it identifies a spam, to block it. Thus, the development of the BDAS stack is driven by the believe that “data is as useful as the decisions you can take based on that data” Data is only as useful as the decisions it enables

3 8/26/2018 Data Processing Goals Low latency (interactive) queries on historical data: enable faster decisions E.g., identify why a site is slow and fix it Low latency queries on live data (streaming): enable decisions on real-time data E.g., detect & block worms in real-time (a worm may infect 1mil hosts in 1.3sec) Sophisticated data processing: enable “better” decisions E.g., anomaly detection, trend analysis So what does this mean? Well, this means that we want low response-time on historical data since the faster we can make a decision the better. We want the ability to perform queries on live data since decisions on real-time data are better than on stale data. Finally, we want to perform sophisticated processing on massive data as, in principle, processing more data will lead to better decisions.

4 Today’s Open Analytics Stack…
8/26/2018 Today’s Open Analytics Stack… ..mostly focused on large on-disk datasets: great for batch but slow Data Processing Application Today’s open analytics stack is mostly focused on large on-disk datasets. As a result, it provides sophisticated processing on massive data but it is slow.. This stack consists of roughly four layers… Except Google, Microsoft, and at some extend Amazon, this is the de-facto standarrd stack for data analysis today, and it is used by Yahoo!, Facebook, Twitter, to name a few. Storage Infrastructure

5 Goals Batch Interactive Streaming One stack to rule them all!
Easy to combine batch, streaming, and interactive computations Easy to develop sophisticated algorithms Compatible with existing open source ecosystem (Hadoop/HDFS)

6 Our Approach: Support Interactive and Streaming Comp.
High end datacenter node 10Gbps Aggressive use of memory Why? Memory transfer rates >> disk or even SSDs Gap is growing especially w.r.t. disk Many datasets already fit into memory The inputs of over 90% of jobs in Facebook, Yahoo!, and Bing clusters fit into memory E.g., 1TB = 1 billion 1 KB each Memory density (still) grows with Moore’s law RAM/SSD hybrid memories at horizon GB What do people use big data for? First, they use it to generate reports to track and better understand business processes, ransactions Second, they use it to diagnose and answer questions such as Why the user engagement dropping?, why is the system slow? Or to detect spam, worms, or DDoS attacks But most importantly they use it to make decisions, such us improving the business process, deciding what features to add to the product, deciding what ad to show, or, once it identifies a spam, to block it. Thus, the development of the BDAS stack is driven by the believe that “data is as useful as the decisions you can take based on that data” 40-60GB/s 16 cores 0.2-1GB/s (x10 disks) 1-4GB/s (x4 disks) 10-30TB 1-4TB

7 Our Approach: Support Interactive and Streaming Comp.
result T Increase parallelism Why? Reduce work per node  improve latency Techniques: Low latency parallel scheduler that achieve high locality Optimized parallel communication patterns (e.g., shuffle, broadcast) Efficient recovery from failures and straggler mitigation What do people use big data for? First, they use it to generate reports to track and better understand business processes, ransactions Second, they use it to diagnose and answer questions such as Why the user engagement dropping?, why is the system slow? Or to detect spam, worms, or DDoS attacks But most importantly they use it to make decisions, such us improving the business process, deciding what features to add to the product, deciding what ad to show, or, once it identifies a spam, to block it. Thus, the development of the BDAS stack is driven by the believe that “data is as useful as the decisions you can take based on that data” result Tnew (< T)

8 Our Approach: Support Interactive and Streaming Comp.
GB 16 cores 40-60GB/s Trade between result accuracy and response times Why? In-memory processing does not guarantee interactive query processing E.g., ~10’s sec just to scan 512 GB RAM! Gap between memory capacity and transfer rate increasing Challenges: accurately estimate error and running time for… … arbitrary computations doubles every 18 months What do people use big data for? First, they use it to generate reports to track and better understand business processes, ransactions Second, they use it to diagnose and answer questions such as Why the user engagement dropping?, why is the system slow? Or to detect spam, worms, or DDoS attacks But most importantly they use it to make decisions, such us improving the business process, deciding what features to add to the product, deciding what ad to show, or, once it identifies a spam, to block it. Thus, the development of the BDAS stack is driven by the believe that “data is as useful as the decisions you can take based on that data” doubles every 36 months

9 Our Approach Easy to combine batch, streaming, and interactive computations Single execution model that supports all computation models Easy to develop sophisticated algorithms Powerful Python and Scala shells High level abstractions for graph based, and ML algorithms Compatible with existing open source ecosystem (Hadoop/HDFS) Interoperate with existing storage and input formats (e.g., HDFS, Hive, Flume, ..) Support existing execution models (e.g., Hive, GraphLab)

10 Berkeley Data Analytics Stack (BDAS)
New apps: AMP-Genomics, Carat, … Application Application … in two fundamental aspects… At the application layer we build new, real applications such a AMP-Genomics a genomics pipeline, and Carat, an application I’m going to talk about soon. Building real applications allows us to drive the features and design of the lower layers. in-memory processing trade between time, quality, and cost Data Processing Data Processing Data Management Storage Efficient data sharing across frameworks Infrastructure Resource Management Share infrastructure across frameworks (multi-programming for datacenters)

11 The Berkeley AMPLab “Launched” January 2011: 6 Year Plan AMP
Algorithms Machines People AMP “Launched” January 2011: 6 Year Plan 8 CS Faculty ~40 students 3 software engineers Organized for collaboration:

12 Berkeley Data Analytics Stack (BDAS)
The Berkeley AMPLab Funding: XData, CISE Expedition Grant Industrial, founding sponsors 18 other sponsors, including Goal: next Generation of open source analytics stack for industry & academia: Berkeley Data Analytics Stack (BDAS)

13 Berkeley Data Analytics Stack (BDAS)
Application … in two fundamental aspects… At the application layer we build new, real applications such a AMP-Genomics a genomics pipeline, and Carat, an application I’m going to talk about soon. Building real applications allows us to drive the features and design of the lower layers. Data Management Data Processing Resource Management Resource Management Data Management Data Processing Resource Management Data Management Data Processing Infrastructure

14 Berkeley Data Analytics Stack (BDAS)
Existing stack components…. Our instantiation of the resource management is Mesos Data Management Data Processing Resource Management HDFS MPI Resource Mgmnt. Data Processing Hadoop HIVE Pig HBase Storm

15 Mesos [Released, v0.9] Management platform that allows multiple framework to share cluster Compatible with existing open analytics stack Deployed in production at Twitter on 3,500+ servers Our instantiation of the resource management is Mesos HIVE Pig HBase Storm MPI Data Processing Hadoop HDFS Data Mgmnt. Resource Mgmnt. Mesos

16 One Framework Per Cluster Challenges
Inefficient resource usage E.g., Hadoop cannot use available resources from MPI’s cluster No opportunity for stat. multiplexing Hard to share data Copy or access remotely, expensive Hard to cooperate E.g., Not easy for MPI to use data generated by Hadoop Hadoop MPI Hadoop MPI Need to run multiple frameworks on same cluster

17 Mesos Uniprograming Multiprograming Solution: Mesos Hadoop MPI Hadoop
Common resource sharing layer abstracts (“virtualizes”) resources to frameworks enable diverse frameworks to share cluster Our solution is Mesos, a resource sharing layer that abstracts resources to framework, and this way allows over which diverse frameworks cab run, by abstracting resources to framework By doing so we go from uniprograming or one framework per cluster to multiprogramming where we share entire cluster among diverse frameworks. Hadoop MPI Hadoop MPI Mesos Uniprograming Multiprograming

18 Dynamic Resource Sharing
100 node cluster This graph shows the number of CPUS allocated to each of the four frameworks in the macrobenchmark over time for the duration of the 25 minute experiment. You can see that some frameworks take advantage of the cluster when other frameworks have a lull in activity. For instance, around time 1050, the Torque framework finishes running the MPI job and no more MPI jobs have been queued up to run, so it releases its shares (until another MPI job is queued around time 1250 seconds). The two hadoop jobs are able to take advantage of the resources Torque is not using and get more work done and increase the average percent of the cluster that is allocated at any time.

19 Spark [Release, v0.7] In-memory framework for interactive and iterative computations Resilient Distributed Dataset (RDD): fault-tolerance, in-memory storage abstraction Scala interface, Java and Python APIs Our instantiation of the resource management is Mesos HIVE Pig Storm MPI Data Processing Spark Hadoop HDFS Data Mgmnt. Resource Mgmnt. Mesos

20 Our Solution Resilient Distributed Data Sets (RDD)
Partitioned collection of records Immutable Can be created only through deterministic operations from other RDDs Handle of each RDD stores its lineage: Lineage: sequence of operations that created the RDD Recovery: use lineage information to rebuild RDD

21 RDD Example Two-partition RDD A={A1, A2} stored on disk
Apply f() and cache  RDD B Shuffle, and apply g()  RDD C Aggregate using h()  D B = A.f() f() B2 B1 C = B.g() g() C2 C1 A1 h() D = C.h() D RDD A A2

22 RDD Example C1 lost due to node failure before h() is computed
B = A.f() f() B2 B1 A1 g() C1 h() RDD A D D = C.h() A2 g() C2 C = B.g()

23 RDD Example C1 lost due to node failure before h() is computed
Reconstruct C1, eventually, on a different node B = A.f() f() B2 B1 A1 g() C1 h() D h() RDD A D D = C.h() A2 g() C2 C = B.g()

24 PageRank Performance 54 GB Wikipedia dataset

25 Other Iterative Algorithms
Time per Iteration (s) 100 GB datasets

26 Spark Community 3000 people attended online training in August
500+ meetup members 14 companies contributing spark-project.org

27 Spark Streaming [Alpha Release]
Large scale streaming computation Ensure exactly one semantics Integrated with Spark  unifies batch, interactive, and streaming computations! Our instantiation of the resource management is Mesos Spark Streaming HIVE Pig Storm MPI Data Processing Spark Hadoop HDFS Data Mgmnt. Resource Mgmnt. Mesos

28 Existing Streaming Systems
Traditional streaming systems have a event-driven record-at-a-time processing model Each node has mutable state For each record, update state & send new records State is lost if node dies! Making stateful stream processing be fault-tolerant is challenging mutable state node 1 node 3 input records node 2 Traditional streaming systems have what we call a “record-at-a-time” processing model. Each node in the cluster processing a stream has a mutable state. As records arrive one at a time, the mutable state is updated, and a new generated record is pushed to downstream nodes. Now making this mutable state fault-tolerant is hard.

29 Spark: Discretized Stream Processing
Run a streaming computation as a series of very small, deterministic batch jobs live data stream Spark Streaming Chop up the live stream into batches of X seconds Spark treats each batch of data as RDDs and processes them using RDD operations Finally, the processed results of the RDD operations are returned in batches batches of X seconds Spark processed results

30 Spark: Discretized Stream Processing
Run a streaming computation as a series of very small, deterministic batch jobs live data stream Spark Streaming Batch sizes as low as ½ second, latency ~ 1 second Potential for combining batch processing and streaming processing in the same system batches of X seconds Spark processed results

31 Performance Can process 6 GB/sec (60M records/sec) of data on 100 nodes at sub-second latency Tested with 100 streams of data on 100 EC2 instances with 4 cores each

32 Comparison with Storm and S4
Higher throughput than Storm Spark Streaming: 670k records/second/node Storm: 115k records/second/node Apache S4: 7.5k records/second/node Streaming Spark offers similar speed while providing FT and consistency guarantees that these systems lack

33 Fast Fault Recovery Recovers from faults/stragglers within 1 sec

34 Shark [Release, v0.2] HIVE over Spark: SQL-like interface (supports Hive 0.9) up to 100x faster for in-memory data, and 5-10x for disk In tests on hundreds node cluster at Our instantiation of the resource management is Mesos Spark Streaming HIVE Pig Storm MPI Data Processing Shark Spark Hadoop HDFS Data Mgmnt. Resource Mgmnt. Mesos

35 Conviva Warehouse Queries (1.7 TB)
1.1 0.8 0.7 1.0

36 Spark & Shark available now on EMR!
Our instantiation of the resource management is Mesos

37 Tachyon [Alpha Release, this Spring]
High-throughput, fault-tolerant in-memory storage Interface compatible to HDFS Support for Spark and Hadoop Our instantiation of the resource management is Mesos Spark Streaming HIVE Pig Storm MPI Data Processing Shark Hadoop Spark HDFS Tachyon Data Mgmnt. Resource Mgmnt. Mesos

38 BlinkDB [Alpha Release, this Spring]
Large scale approximate query engine Allow users to specify error or time bounds Preliminary prototype starting being tested at Facebook Our instantiation of the resource management is Mesos Spark Streaming BlinkDB Pig Storm MPI Data Processing Shark HIVE Spark Hadoop HDFS Tachyon Data Mgmnt. Resource Mgmnt. Mesos

39 SparkGraph [Alpha Release, this Spring]
GraphLab API and Toolkits on top of Spark Fault tolerance by leveraging Spark Our instantiation of the resource management is Mesos Spark Streaming Spark Graph BlinkDB Pig Storm MPI Data Processing Shark HIVE Spark Hadoop HDFS Tachyon Data Mgmnt. Resource Mgmnt. Mesos

40 MLbase [In development]
Declarative approach to ML Develop scalable ML algorithms Make ML accessible to non-experts Our instantiation of the resource management is Mesos Spark Streaming Spark Graph MLbase BlinkDB Pig Storm MPI Data Processing Shark HIVE Spark Hadoop HDFS Tachyon Data Mgmnt. Resource Mgmnt. Mesos

41 Compatible with Open Source Ecosystem
Support existing interfaces whenever possible GraphLab API Our instantiation of the resource management is Mesos Hive Interface and Shell Spark Streaming Spark Graph MLbase BlinkDB Pig Storm MPI Data Processing Shark HIVE Spark Hadoop Compatibility layer for Hadoop, Storm, MPI, etc to run over Mesos HDFS Tachyon Data Mgmnt. HDFS API Resource Mgmnt. Mesos

42 Compatible with Open Source Ecosystem
Accept inputs from Kafka, Flume, Twitter, TCP Sockets, … Use existing interfaces whenever possible Support Hive API Our instantiation of the resource management is Mesos Spark Streaming Spark Graph MLbase BlinkDB Pig Storm MPI Data Processing Shark HIVE Support HDFS API, S3 API, and Hive metadata Spark Hadoop HDFS Tachyon Data Mgmnt. Resource Mgmnt. Mesos

43 Holistic approach to address next generation of Big Data challenges!
Summary Holistic approach to address next generation of Big Data challenges! Support interactive and streaming computations In-memory, fault-tolerant storage abstraction, low-latency scheduling,... Easy to combine batch, streaming, and interactive computations Spark execution engine supports all comp. models Easy to develop sophisticated algorithms Scala interface, APIs for Java, Python, Hive QL, … New frameworks targeted to graph based and ML algorithms Compatible with existing open source ecosystem Open source (Apache/BSD) and fully committed to release high quality software Three-person software engineering team lead by Matt Massie (creator of Ganglia, 5th Cloudera engineer) Batch Interactive Streaming Spark

44 Thanks!

45 Hive Architecture

46 Shark Architecture


Download ppt "Berkeley Data Analytics Stack (BDAS) Overview"

Similar presentations


Ads by Google