760 likes | 779 Views
CS 63016 & CS 73016 Big Data Analytics. Chapter 4: MapReduce. Xiang Lian Department of Computer Science Kent State University Email: xlian@kent.edu Homepage: http ://www.cs.kent.edu/~xlian/. Objectives. In this chapter, you will: Learn the background of MapReduce
E N D
CS 63016 & CS 73016 Big Data Analytics Chapter 4: MapReduce Xiang Lian Department of Computer Science Kent State University Email: xlian@kent.edu Homepage: http://www.cs.kent.edu/~xlian/
Objectives • In this chapter, you will: • Learn the background of MapReduce • Understand the MapReduce programming model • Know the properties of MapReduce • Get familiar with the idea of MapReduce with a "Hello world!" example • Get to know big data platforms • Google File Systems • Hadoop • Amazon Web Service (AWS)
Outline • Introduction • Challenges • Background of MapReduce • MapReduce • Characteristics of MapReduce • Google File Systems • Hadoop • Amazon Web Services
Introduction • Lots of demands for large-scale data processing • Big spatial data • Big Web data • Big stream data • Big graph data • … • Applications • Location-based services • Semantic Web • Social media studies • Transportation systems • Sensor data analysis • …
Typical Big Data Problems • Iterate over a large number of records • Extract something of interest from each • Shuffle and sort intermediate results • Aggregate intermediate results • Generate final output
Typical Big Data Problems (cont'd) • The problem • Diverse input format (data diversity and heterogeneity) • Large Scale: Terabytes, Petabytes • Parallelization
How to Leverage a Number of Cheap Off-the-Shelf Computers? Image from http://wiki.apache.org/hadoop-data/attachments/HadoopPresentations/attachments/aw-apachecon-eu-2009.pdf
Divide and Conquer “Work” Partition w1 w2 w3 “worker” “worker” “worker” r1 r2 r3 Combine “Result”
grep matches Split data grep Split data matches All matches grep cat Split data matches grep Split data matches Distributed Grep grep is a command-line utility for searching plain-text data sets for lines matching a regular expression. very big data
count Split data count count Split data count merged count merge Split data count count count Split data count Distributed Word Count Word countis to count the number of words in documents very big data
Parallelization Challenges • How do we assign work units to workers? • What if we have more work units than workers? • What if workers need to share partial results? • How do we aggregate partial results? • How do we know all the workers have finished? • What if workers die? What is the common theme of all of these problems?
Common Theme? • Parallelization problems arise from: • Communication between workers (e.g., to exchange state) • Access to shared resources (e.g., data) • Thus, we need a synchronization mechanism
Managing Multiple Workers • Difficult because • We don’t know the order in which workers run • We don’t know when workers interrupt each other • We don’t know the order in which workers access shared data • Thus, we need: • Semaphores (lock, unlock) • Conditional variables (wait, notify, broadcast) • Barriers • Still, lots of problems: • Deadlock, livelock, race conditions... • Dining philosophers, sleeping barbers, cigarette smokers... • Moral of the story: be careful!
Current Tools Message Passing Shared Memory • Programming models • Shared memory (pthreads) • Message passing (MPI) • Design Patterns • Master-slaves • Producer-consumer flows • Shared work queues Memory P1 P2 P3 P4 P5 P1 P2 P3 P4 P5 producer consumer master work queue slaves producer consumer
Concurrency Challenge! • Concurrency is difficult to reason about • Concurrency is even more difficult to reason about • At the scale of datacenters (even across datacenters) • In the presence of failures • In terms of multiple interacting services • Not to mention debugging… • The reality: • Lots of one-off solutions, custom code • Write you own dedicated library, then program with it • Burden on the programmer to explicitly manage everything
What's the Point? • It's all about the right level of abstraction • The von Neumann architecture has served us well, but is no longer appropriate for the multi-core/cluster environment • Hide system-level details from the developers • No more race conditions, lock contention, etc. • Separating the what from how • Developer specifies the computation that needs to be performed • Execution framework (“runtime”) handles actual execution
Key Ideas • Scale "out", not "up" • Limits of SMP and large shared-memory machines • Move processing to the data • Cluster have limited bandwidth • Process data sequentially, avoid random access • Seeks are expensive, disk throughput is reasonable • Seamless scalability • From the mythical man-month to the tradable machine-hour
The Datacenter is the Computer! Image from http://wiki.apache.org/hadoop-data/attachments/HadoopPresentations/attachments/aw-apachecon-eu-2009.pdf
What is MapReduce? • Origin from Google [OSDI’04] • J. Dean and S. Ghemawat. MapReduce: Simplified Data Processing on Large Clusters. In OSDI, 2004. • https://research.google.com/archive/mapreduce.html • A simple programming model • For large-scale data processing • Exploits a large set of commodity computers • Executes process in a distributed manner • Offers high availability
Large Scale Data Processing • Many tasks: process lots of data to produce other data • Want to use hundreds or thousands of CPUs • ... but this needs to be easy • MapReduce provides: • Automatic parallelization and distribution • Fault-tolerance • I/O scheduling • Status and monitoring
Typical Large-Data Problem Map • Iterate over a large number of records • Extract something of interest from each • Shuffle and sort intermediate results • Aggregate intermediate results • Generate final output Reduce Key idea: provide a functional abstraction for these two operations (Dean and Ghemawat, OSDI 2004)
Map: Accepts input key/value pair Emits intermediate key/value pair Reduce : Accepts intermediate key/value* pair Emits output key/value pair Partitioning Function Map+Reduce R E D U C E M A P very big data Result
MapReduce – A Programming Model • Input and Output of MapReduce • each a set of key/value pairs • Programmers specify two functions: map(k, v) → [(k’, v’)] • Processes input key/value pair • Produces set of intermediate pairs reduce(k’, [v’]) → [(k’, v’)] • Combines all intermediate values for a particular key (i.e., all values with the same key are sent to the same reducer) • Produces a set of merged output values (usually just one)
Example of MapReduce (Word Count) k1 v1 k2 v2 k3 v3 k4 v4 k5 v5 k6 v6 map map map map a 1 b 2 c 3 c 6 a 5 c 2 b 7 c 8 Shuffle and Sort: aggregate values by keys a 1 5 b 2 7 c 2 3 6 8 reduce reduce reduce r1 s1 r2 s2 r3 s3
Word Count Execution Input Map Shuffle & Sort Reduce Output the, 1 brown, 1 fox, 1 the quick brown fox brown, 2 fox, 2 how, 1 now, 1 the, 3 Map Reduce the, 1 fox, 1 the, 1 the fox ate the mouse Map quick, 1 how, 1 now, 1 brown, 1 ate, 1 cow, 1 mouse, 1 quick, 1 ate, 1 mouse, 1 Reduce how now brown cow Map cow, 1
"Hello World" Example: Count Word Occurrences • map (String input_key, String input_value): • // input_key: document name • // input_value: document contents • for each word w in input_value: • EmitIntermediate(w, "1"); • reduce (String output_key, Iterator intermediate_values): • // output_key: a word • // output_values: a list of counts • int result = 0; • for each v in intermediate_values: • result += ParseInt(v); • Emit(AsString(result));
MapReduce – A Programming Model (cont'd) • Programmers specify two functions: map(k, v) → [(k', v')] reduce(k', [v']) → [(k', v')] • The execution framework handles everything else… What’s "everything else"?
MapReduce "Runtime" • Handles scheduling • Assigns workers to map and reduce tasks • Handles "data distribution" • Moves processes to data • Handles synchronization • Gathers, sorts, and shuffles intermediate data • Handles errors and faults • Detects worker failures and restarts • Everything happens on top of a distributed FS (later)
MapReduce • Programmers specify two functions: map(k, v) → [(k', v')] reduce(k', [v']) → [(k', v')] • All values with the same key are reduced together • The execution framework handles everything else… • Not quite…usually, programmers also specify: partition(k', number of partitions)→ partition for k' • Often a simple hash of the key, e.g., hash(k') mod n • Divides up key space for parallel reduce operations combine(k', [v']) → [(k', v")] • Mini-reducers that run in memory after the map phase • Used as an optimization to reduce network traffic
k1 v1 k2 v2 k3 v3 k4 v4 k5 v5 k6 v6 map map map map a 1 b 2 c 3 c 6 a 5 c 2 b 7 c 8 combine combine combine combine a 1 b 2 c 9 a 5 c 2 b 7 c 8 partition partition partition partition Shuffle and Sort: aggregate values by keys a 1 5 b 2 7 c c 2 2 3 9 6 8 8 reduce reduce reduce r1 s1 r2 s2 r3 s3
Architecture Overview Master node user Job tracker Slave node N Slave node 2 Slave node 1 Task tracker Task tracker Task tracker Workers Workers Workers
Distributed File System • Don't move data to workers … move workers to the data! • Store data on the local disks of nodes in the cluster • Start up the workers on the node that has the data local • Why? • Not enough RAM to hold all the data in memory • Disk access is slow, but disk throughput is reasonable • A distributed file system is the answer • GFS (Google File System) for Google’s MapReduce • HDFS (Hadoop Distributed File System) for Hadoop
GFS: Assumptions • Commodity hardware over "exotic" hardware • Scale "out", not "up" • High component failure rates • Inexpensive commodity components fail all the time • "Modest" number of huge files • Multi-gigabyte files are common, if not encouraged • Files are write-once, mostly appended to • Perhaps concurrently • Large streaming reads over random access • High sustained throughput over low latency GFS slides adapted from material by (Ghemawat et al., SOSP 2003)
GFS: Underlying Storage System • Goal • Global view • Make huge files available in the face of node failures • Master node (meta server) • Centralized, index all chunks on data servers • Chunk server (data server) • File is split into contiguous chunks, typically 16-64MB • Each chunk replicated (usually 2x or 3x) • Try to keep replicas in different racks
GFS: Design Decisions • Files stored as chunks • Fixed size (64MB) • Reliability through replication • Each chunk replicated across 3+ chunk servers • Single master to coordinate access, keep metadata • Simple centralized management • No data caching • Little benefit due to large datasets, streaming reads • Simplify the API • Push some of the issues onto the client (e.g., data layout) HDFS = GFS clone (same basic ideas)
GFS Master C1 C0 C1 C0 C5 C5 C2 C2 C3 C5 GFS Architecture Client … ChunkserverN Chunkserver 2 Chunkserver 1
Functions in the Model • Map • Process a key/value pair to generate intermediate key/value pairs • Reduce • Merge all intermediate values associated with the same key • Partition • By default : hash(key) mod R • Well balanced
Fault Tolerance • Fault tolerance handled via re-execution • Worker failure • Heartbeat, Workers are periodically pinged by master • NO response = failed worker • If the processor of a worker fails, the tasks of that worker are reassigned to another worker • Master failure • Master writes periodic checkpoints • Another master can be started from the last checkpointed state • If eventually the master dies, the job will be aborted Robust: lost 1600 of 1800 machines once, but finished fine
Fault Tolerance (cont'd) • Refinement: Redundant Execution • The problem of “stragglers” (slow workers) • Other jobs consuming resources on machine • Bad disks with soft errors transfer data very slowly • Weird things: processor caches disabled (!!) • When the computation is almost done, reschedule in-progress tasks • Whenever either the primary or the backup execution finishes, mark it as completed Effect: Dramatically shortens job completion time
Usage: MapReduce Jobs Run in August 2004 • Number of jobs 29,423 • Average job completion time 634 secs • Machine days used 79,186 days • Input data read 3,288 TB • Intermediate data produced 758 TB • Output data written 193 TB • Average worker machines per job 157 • Average worker deaths per job 1.2 • Average map tasks per job 3,351 • Average reduce tasks per job 55 • Unique map implementations 395 • Unique reduce implementations 269 • Unique map/reduce combinations 426
The MapReduce Model is Widely Applicable • MapReduce Programs in Google Source Tree Examples as follows
Applications • String Match, such as grep • Inverted index • Count URL access frequency • Lots of examples in data mining • Mining frequent itemsets • Clustering • …
Detailed Example: Word Count (2) • Reduce