Chapter 2 Lin and Dyer & http://developer.yahoo.com/hadoop/tutorial/ MapReduce Basics Chapter 2 Lin and Dyer & http://developer.yahoo.com/hadoop/tutorial/

Slides:



Advertisements
Similar presentations
MAP REDUCE PROGRAMMING Dr G Sudha Sadasivam. Map - reduce sort/merge based distributed processing Best for batch- oriented processing Sort/merge is primitive.
Advertisements

MapReduce.
Overview of MapReduce and Hadoop
Based on the text by Jimmy Lin and Chris Dryer; and on the yahoo tutorial on mapreduce at index.html
Cloud Computing Lecture #3 More MapReduce Jimmy Lin The iSchool University of Maryland Wednesday, September 10, 2008 This work is licensed under a Creative.
Based on Lin and Dryer’s text: Chapter 3.  Figure 2.6.
CS 345A Data Mining MapReduce. Single-node architecture Memory Disk CPU Machine Learning, Statistics “Classical” Data Mining.
Jimmy Lin The iSchool University of Maryland Wednesday, April 15, 2009
Homework 2 In the docs folder of your Berkeley DB, have a careful look at documentation on how to configure BDB in main memory. In the docs folder of your.
Google’s Map Reduce. Commodity Clusters Web data sets can be very large – Tens to hundreds of terabytes Standard architecture emerging: – Cluster of commodity.
Parallel K-Means Clustering Based on MapReduce The Key Laboratory of Intelligent Information Processing, Chinese Academy of Sciences Weizhong Zhao, Huifang.
Design Patterns for Efficient Graph Algorithms in MapReduce Jimmy Lin and Michael Schatz University of Maryland MLG, January, 2014 Jaehwan Lee.
CS506/606: Problem Solving with Large Clusters Zak Shafran, Richard Sproat Spring 2011 Introduction URL:
MapReduce.
Map Reduce for data-intensive computing (Some of the content is adapted from the original authors’ talk at OSDI 04)
1 The Map-Reduce Framework Compiled by Mark Silberstein, using slides from Dan Weld’s class at U. Washington, Yaniv Carmeli and some other.
Hadoop/MapReduce Computing Paradigm 1 Shirish Agale.
Vyassa Baratham, Stony Brook University April 20, 2013, 1:05-2:05pm cSplash 2013.
Tutorial: Big Data Algorithms and Applications Under Hadoop KUNPENG ZHANG SIDDHARTHA BHATTACHARYYA
MAP REDUCE BASICS CHAPTER 2. Basics Divide and conquer – Partition large problem into smaller subproblems – Worker work on subproblems in parallel Threads.
MapReduce Algorithm Design Based on Jimmy Lin’s slides
MapReduce and the New Software Stack CHAPTER 2 1.
Chapter 5 Ranking with Indexes 1. 2 More Indexing Techniques n Indexing techniques:  Inverted files - best choice for most applications  Suffix trees.
IBM Research ® © 2007 IBM Corporation Introduction to Map-Reduce and Join Processing.
C-Store: MapReduce Jianlin Feng School of Software SUN YAT-SEN UNIVERSITY May. 22, 2009.
Big Data Infrastructure Week 2: MapReduce Algorithm Design (2/2) This work is licensed under a Creative Commons Attribution-Noncommercial-Share Alike 3.0.
MapReduce Basics Chapter 2 Lin and Dyer & /tutorial/
COMP7330/7336 Advanced Parallel and Distributed Computing MapReduce - Introduction Dr. Xiao Qin Auburn University
Jimmy Lin and Michael Schatz Design Patterns for Efficient Graph Algorithms in MapReduce Michele Iovino Facoltà di Ingegneria dell’Informazione, Informatica.
CS350 - MAPREDUCE USING HADOOP Spring PARALLELIZATION: BASIC IDEA Parallelization is “easy” if processing can be cleanly split into n units:
Image taken from: slideshare
Auburn University
Hadoop Aakash Kag What Why How 1.
Map-Reduce framework.
Large-scale file systems and Map-Reduce
Hadoop MapReduce Framework
MapReduce Types, Formats and Features
Auburn University COMP7330/7336 Advanced Parallel and Distributed Computing MapReduce - Introduction Dr. Xiao Qin Auburn.
Auburn University COMP7330/7336 Advanced Parallel and Distributed Computing Data Partition Dr. Xiao Qin Auburn University.
PREGEL Data Management in the Cloud
Lecture 3: Bringing it all together
MapReduce Computing Paradigm Basics Fall 2013 Elke A. Rundensteiner
Database Applications (15-415) Hadoop Lecture 26, April 19, 2016
Hadoop MapReduce Types
COS 418: Distributed Systems Lecture 1 Mike Freedman
Cloud Distributed Computing Environment Hadoop
湖南大学-信息科学与工程学院-计算机与科学系
MapReduce: Data Distribution for Reduce
MapReduce Algorithm Design
Cse 344 May 4th – Map/Reduce.
MapReduce Algorithm Design Adapted from Jimmy Lin’s slides.
CS110: Discussion about Spark
Word Co-occurrence Chapter 3, Lin and Dyer.
Distributed System Gang Wu Spring,2018.
Cloud Computing: Project Tutorial Hadoop Map-Reduce Programming
Lecture 16 (Intro to MapReduce and Hadoop)
Introduction to MapReduce
MAPREDUCE TYPES, FORMATS AND FEATURES
Midterm Review CSE4/587 B.Ramamurthy 4/4/2019 4/4/2019 B.Ramamurthy
MapReduce Algorithm Design
Midterm Review CSE4/587 B.Ramamurthy 4/8/2019 4/8/2019 B.Ramamurthy
CS639: Data Management for Data Science
5/7/2019 Map Reduce Map reduce.
Chapter 2 Lin and Dyer & MapReduce Basics Chapter 2 Lin and Dyer &
COS 518: Distributed Systems Lecture 11 Mike Freedman
MapReduce: Simplified Data Processing on Large Clusters
Midterm Exam Review.
Word Co-occurrence Chapter 3, Lin and Dryer.
Distributed Systems and Concurrency: Map Reduce
Map Reduce, Types, Formats and Features
Presentation transcript:

Chapter 2 Lin and Dyer & http://developer.yahoo.com/hadoop/tutorial/ MapReduce Basics Chapter 2 Lin and Dyer & http://developer.yahoo.com/hadoop/tutorial/

Lin and Dyer’s text Chapter 1: Please read: sets the context for MR Chapter 2: MR Basics: analysis of a sample problem analysis/walkthrough Chapter 3: MR Algorithm Design (up to p.60) Chapter 4: Inverted index for text retrieval Chapter 5: Graph algorithms (pagerank and other classical algorithms)

Divide and Conquer Algorithm How do we break up a large problem into smaller tasks? More specifically, how do we decompose the problem so that the smaller tasks can be executed in parallel? How do we assign tasks to workers distributed across a potentially large number of machines (while keeping in mind that some workers are better suited to running some tasks than others, e.g., due to available resources, locality constraints, etc.)? How do we ensure that the workers get the data they need? How do we coordinate synchronization among the different workers? How do we share partial results from one worker that is needed by another? How do we accomplish all of the above in the face of software errors and hardware faults?

Basics map: (k1,v1) --> [(k2,v2)] reduce: (k2,[v2])-->[(k3,v3)] Implicit between map and reduce is the implicit "group by" operation of intermediate keys Hadoop does not have the GFS restriction that reducer's input key and output key type should be same.

Architecture http://developer.yahoo.com/hadoop/tutorial

Combiners and Partitioners Combiner: combines the same “keys” at the output of a mapper Runs after the Mapper and before the Reducer. Usage of the Combiner is optional. The Combiner will receive as input all data emitted by the Mapper instances on a given node. The output from the Combiner is then sent to the Reducers, instead of the output from the Mappers. The Combiner is a "mini-reduce" process which operates only on data generated by one machine. This process of moving map outputs to the reducers is known as shuffling. A different subset of the intermediate key space is assigned to each reduce node; these subsets (known as "partitions") are the inputs to the reduce tasks. Default partition function is a hash function. Each map task may emit (key, value) pairs to any partition; all values for the same key are always reduced together regardless of which mapper is its origin.

MR Workflow

Complete Example Figure 2.2: Simplified MR process (picture) Figure 2.3: MR Algorithm (pseudo code) Figure 2.4: MR process with combiners and partitioners (picture) Your answers to midterm questions should be in this format. Lets go through couple of examples: (i) word count (ii) max value for each key