Presentation is loading. Please wait.

Presentation is loading. Please wait.

Cloud Computing Systems

Similar presentations


Presentation on theme: "Cloud Computing Systems"— Presentation transcript:

1 Cloud Computing Systems
Distributed Systems Topic 11: Cloud Computing Systems Dr. Michael R. Lyu Computer Science & Engineering Department The Chinese University of Hong Kong This lecture is a high level introduction of Cloud Computing.

2 Outline 1. Cloud Computing Introduction 2. Virtualization Concept
3. The Hadoop Distributed File System (HDFS) 4. MapReduce Computing Paradigm 5. Summary First we will brief introduce the main idea of cloud computing Then we will see how virtualization concept enables cloud computing Then the popular MapReduce computing thought And talking about cloud computing in practice

3 1 What Is Cloud Computing?
Cloud computing is a hot buzzword recently More and more companies have claimed their products are cloud computing based So, what is cloud computing A new name of an old technology created by marketing people? A new technology focusing on higher-speed computing? A new programming paradigm? Everyone may have heard about cloud computing…it has became a hot buzzword in recent years. More and more companies have advertised that their products are cloud-based…for desktop applications, examples even include the software that is used everyday. For instances, the Chinese input methods software and the anti-virus software. So what is cloud computing? If you don’t know the technical things of cloud computing, when seeing so much “cloud” things over the media, the commercials especially, you may probably think that cloud computing is nothing but a marketing word. Or if you are less cynical, you may think cloud computing is, well, actually a new thing, but what is it? A new technology for achieving higher speed computing, or a new programming paradigm… Well, there are actually many cloud computing definitions if you search it from the Internet. In this course, I will not define this buzzword, since its meaning is actually evolving…This is very common for an emerging IT technique. I will try to let you know the underlying philosophy of cloud computing….

4 1 What Is Cloud Computing?
Let us see what cloud computing is by its economic drive See an example on hosting an Internet service e.g., a website Example is the best way for you to understand the fundamental idea of cloud computing. Well, cloud computing is coming from the industrial community. The industrial ppl value economic drive most…So let’s see what cloud computing is by its economic drive. Suppose you are running a small Internet-based business...You want to build, say, a website for your clients to access some sort of information... Server CUHK co. ltd.

5 1.1 Hosting Computing Systems
CUHK co. ltd. A single server is enough for small business Your company is small, the number of your clients is small too... To say money, actually you just need to buy a low-end desktop PC to host your web server. Everyone is feeling good in this scenario… A small amount of budget

6 1.1 Hosting Computing Systems
CUHK co. ltd. The fortunately, your company is getting bigger, the number of your clients is getting larger too... So we need to buy more servers for the Internet service. Surely it will cost more manpower and money to provide the service. And so far so good…everyone is still satisfactory More servers are needed if you have more clients Cost more manpower and money to provide the service

7 1.1 Hosting Computing Systems
A huge data center is needed if you have a huge number of clients CUHK co. ltd. Then suppose your company is getting larger and larger…now you have a large amount of clients. You will need a huge data center… Surely the data center and its maintenance will cost you a huge amount of money now… You would wish your printer can print cash for you…  Cost a huge amount of money to provide the service

8 1.1 Hosting Computing Systems
A huge data center is needed if you have a huge number of clients CUHK co. ltd. Furthermore, during peak hours, many of your clients will still feel unsatisfied …. They will meet such 503 message more frequently. You have two choices: 1) let them be unsatisfied or.2). Buy more servers….but most servers will be used only during the peak hours…during the rest of the time, the servers will be idle….this is really wasting money. Cost a huge amount of money to provide the service

9 1.1 Hosting Computing Systems
The number of concurrent users is dynamic You need to prepare for the peak number Otherwise, bad user experience! So you have to pay for the servers no matter whether you are using them or not The money you spend to buy the hardware The salary for the maintainers The electricity (including that for air conditioning) Other maintenance costs, e.g., repairing, upgrading, renting a room for the servers We can see that the problem here is that the number of concurrent users is dynamic. The Internet service needs to prepare for the peak number (if you not, it will cause bad user experience during peak hours). So you have to buy servers for the peak hours, however, the servers will cause your money no matter whether it is in peak hours or not… The money include: The money you spend to buy the hardware. The salary for the maintainers. The electricity (including that for air conditioning. Other maintenance costs, e.g., repairing, upgrading, renting a room for the servers

10 1.2 Cloud Computing Motivations
How to make it more economic when hosting computing systems I wish this cloud can really host my system so that I can provide a scalable, on-demand service So the problem is: how to make it more economic when hosting internet based computing systems You may wish that the cloud in your design sketch will really host your system automatically…you don’t need to care much about the hardware implementation…. But you can still get a scalable on demand service ……that is , no matter how many users you get, the users can all get a good experience using your system…and on the other hand, you don’t need to pay all the time like for the peak hours…. Server CUHK co. ltd.

11 1.2 Cloud Computing Introduction
Hosting computing systems in a scalable, on-demand way To be economic! Cloud computing An emerging computing concept that approaches such a goal Based on a lot of distributed system techniques Service-oriented system architecture Virtualization technology Distributed computing paradigm Cloud computing is an emerging computing concept that can approach such a goal…. Cloud computing can actually base itself on a lot of current distributed techniques….like soa: service oriented architecture…virtualization technology and contemporary distributed computing paradigm

12 1.3 Pay As You Go The core of cloud computing is computing/ storage outsourcing To another company To a dedicating unit in the same organization vs The core of cloud computing is computing/ storage outsourcing By outsourcing, we mean a more general term than just finding another company to handle your internet-based services…you can outsourcing within your organization to a dedicating unit, say some IT department dedicating in service hosting…. Cloud computing concept is kind of like using a cab rather than buying a car for transportation…. You pay for exactly what you’ve used….this is definitely more economic Pay for exactly what you’ve used Pay as you go!! Cost Down!!!

13 1.3 Outsourcing Cost down!!! Electronics industry Cloud computing
Service vendor Services for end users Brand vendor Cloud computing provider Manufacturer Manufacturing outsourcing Computing and storage outsourcing Different service requirements in different times Different manufacturing requirements in low season and high season Outsourcing is not a new idea…actually it seems like a successful concept for electronics industry…This slide compares OEM manufacturing and cloud computing… # of hardware machines # of maintainers # of product lines # of workers Cost down!!!

14 1.4 Cloud Computing System Architecture
User software, API Application OS, middleware Platform Now we know the basic concept of cloud computing….let’s start with discuss the typical architecture of a cloud computing system A computer system will typically contains three tiers. The hardware, the OS/middleware, and over which the user software are running Hardware Infrastructure

15 1.4 Cloud Computing System Architecture
Three cloud computing stacks A cloud user’s own application A cloud user’s own application and OS Application SaaS: Software as a service Platform Infrastructure PaaS: Platform as a service If the whole thing is provides by the cloud computing provider, we say this is SaaS: software as a service. In this architecture, you may use the cloud-based application directly. Examples are Gmail and Google map... You may also call the api within your own application… If the platform and the hardware infrastructure is provided by the cloud computing provider, you can build your own application over the platform…we say this is PaaS: platform as a service…. If only the infrastructure is provided by the cloud computing provider, we say this is IaaS: Infrastructure as a service. Provided by the cloud computing providers IaaS: Infrastructure as a service

16 1.5 Cloud Computing Industry
Amazon Elastic Compute Cloud (EC2) Microsoft's Windows Azure Platform Google App Engine Other small startups: Heroku & Engine Yard And more (and more)… Now we introduce some typical cloud computing providers…

17 1.5 Cloud Computing Industry
Amazon Elastic Compute Cloud (EC2) Xen-based virtual computing environment A user can run Linux-based applications IaaS: A user can control nearly the entire software stack, from the kernel upwards. Provides low level of virtualization Raw CPU cycles, block-device storage, IP-level connectivity Microsoft's Windows Azure Platform Google App Engine Other small startups: Heroku & Engine Yard And more (and more)… Now we introduce some typical cloud computing providers… The first example is the Amazon EC2. EC2 provides a Xen-based virtual computing environment where a user can run Linux-based applications. It is an IaaS: A user can control nearly the entire software stack, from the kernel upwards. It provides low level of virtualization: Raw CPU cycles, block-device storage, IP-level connectivity. In summary, EC2 instances are something like remote PCs, you can login, run your application there… (PLEASE DO NO DISCUSS MICROSOFT AND THE REST OF THE COMPANIES, THEY WILL BE DISCUSSED LATER)

18 1.5 Cloud Computing Industry
Amazon Elastic Compute Cloud (EC2) Microsoft's Windows Azure Platform PaaS: Use Windows Azure Hypervisor as the infrastructure, and use the .NET framework as the application container Support general-purpose computing Users can choose language, but cannot control the underlying operating system or runtime Google App Engine Other small startups: Heroku & Engine Yard And more (and more)… Microsoft also provides a PaaS cloud computing solution called azure When using this cloud computing system, you can use Windows Azure Hypervisor as the infrastructure, and use the .NET framework as the application container That means your application should be dot-Net-based…But you have your flexibility to choose your programming language, but you cannot control the underlying operating system or runtime….Azure supports general-purpose computing..

19 1.5 Cloud Computing Industry
Amazon Elastic Compute Cloud (EC2) Microsoft's Windows Azure Platform Google App Engine Support APIs for the Google Datastore, Google Accounts, URL fetch, image manipulation, and services, etc. SaaS: Application domain-specific platforms Not suitable for general-purpose computing Other small startups: Heroku & Engine Yard And more (and more)… Google App Engine is a cloud computing solution provided by Google. It supports APIs for the Google Datastore, Google Accounts, URL fetch, image manipulation, and services, etc. It is easy to see that Google App Engine is a SaaS cloud computing system: it is application domain-specific platforms.. So it is not suitable for general-purpose computing, but very convenient to program google business related applications, like using google map.

20 1.5 Cloud Computing Industry
Amazon Elastic Compute Cloud (EC2) Microsoft's Windows Azure Platform Google App Engine Other small startups: Heroku & Engine Yard Based on Ruby on Rails PaaS And more (and more)… Finally, there are a lot of small startup companies that provide elegant cloud computing solutions. Examples are Heroku and Engine Yard, both are Silicon-valley based startups. They provide PaaS solutions which are based on Ruby on Rails… you can search the technical details from the Internet if interested

21 2 Virtualization Concept
Virtualization refers to the act of creating a virtual (rather than actual) version of computer hardware platform operating system (OS) storage device database server computer network resources We can virtualize physical servers to support all applications running over them Virtualization of the key to Cloud Computing Virtualization refers to the act of creating a virtual (rather than actual) version of computer hardware platform operating system (OS) storage device database server computer network resources We can virtualize physical servers to support all applications running over them Virtualization of the key to Cloud Computing

22 2.1 The Traditional Server Concept
We have discussed so much basic concept of cloud computing. Now let’s discuss its key enabling technology: virtualization… First, let’s see the traditional server concept, as shown in the slide. For example, if you want to build a database server, you need to install OS on your computer. Then install the database application over the OS. If your service need a web server, which is a IIS running over Windows , an app server which is a glassfish running over Linux, a database server which is MySQL running over linux, and finally an server which is an exchange server running over Windows. In total, you will need 4 computers. Web Server Windows IIS App Server Linux Glassfish DB Server Linux MySQL Windows Exchange

23 2.1 The Traditional Server Concept
System administrators often talk about servers as a whole unit Including the hardware, the OS, the storage, and the applications Servers are often referred to by their functions e.g., the Web servers, the SQL servers If the SQL servers are overloaded, the administrator must add in a new one System administrators often talk about servers as a whole unit including the hardware, the OS, the storage, and the applications Servers are often referred to by their functions. For example, the Web servers, the SQL servers…. So we can see that if the SQL servers are overloaded, the administrator must add in a new one …

24 2.1 If Something Going Wrong…
Now if the App server is down…. Well since the whole service is depended on the app server, the whole service will be down too… Web Server Windows IIS App Server DOWN! DB Server Linux MySQL Windows Exchange

25 2.1 The Traditional Server Concept
If server failure is experienced Unless there are redundant servers, the whole service is down The administrators can implement clusters of servers Make the service more fault tolerant However, even clusters are not scalable And, resource utilization is typically low The example in the previous slide shows that if a server failure is experienced, unless there are redundant servers, the whole service is down No one can guarantee no servers will be down, right? Since the software may have bugs, the hardware, unfortunately, is unreliable too.. Moreover, there may be a lot of concurrent users, making the serves overload…. To increase the availability of the whole service, the administrators can implement clusters of servers…That is, use a set of identical servers to provide redundant services. … This can make the service more reliable, since if one server is down, the other servers of the same functionality Is still available. But this solution is not good enough….clusters are not scalable, you use more servers, resource utilization will be typically low

26 2.2 The Virtual Server Concept
A better solution is to virtualize the physical servers in the cluster, so that all applications can run over the same cluster In this way, it would be easier to dynamically assign the number of servers for each service application And even split a hardware server (a computer) for multiple usages (for hosting multiple services) To this end, we need a Virtual Machine Monitor (VMM) layer between Guest OS and hardware Virtual Machine Monitor (VMM) layer between Guest OS and hardware

27 2.2 The Virtual Server Concept
Virtual servers encapsulate the service software away from the hardware A virtual server can be hosted in one or more hardware machines One hardware machine can host more than one virtual server Clustering Service Console Server 1 Guest OS Server 2 Guest OS Virtual servers encapsulate the service software away from the hardware A virtual server can be hosted in one or more hardware machines One hardware machine can host more than one virtual server The VMM can intercept hardware request form the hosted applications, and translate them to real hardware request. VMM (Virtual Machine Monitor) x86 Architecture Intercepts hardware requests

28 2.2 Virtualized Vs. Traditional
App App App App App App OS OS OS Operating System Hypervisor This slide compare the traditional way to host Internet-based application with the virtualized way to do so We can see that, there is a hypervisor in between the hardware and the guest OS…so that several operating systems can run over one computer, Hardware Hardware Traditional Stack Virtualized Stack

29 2.2 The Virtual Server Concept
A flexible way to host computing systems Virtual servers are just files in the storage They can be migrated from one machine to another easily Hardware machines can be removed and introduced conveniently Administrators can adjust the amount of resources allocated to each virtual server Accommodate to the service requirements Virtualization technique provides a flexible way to host computing systems…..Actually, each virtual server is just a file in the storage..It can be read and loaded conveniently. A virtual server can also be migrated from one machine to another easily In this way, hardware machines can be removed and introduced conveniently Administrators (or a sophisticated program) can adjust the amount of resources allocated to each virtual server…this provides a good way to accommodate the service requirements of hardware resources

30 2.3 Virtualization Technique Supports
Software implementations Commercial software e.g., VMWare Open-source software e.g., XEN, VirtualBox Hardware support Intel VT (Intel Virtualization Technology) AMD-V (AMD-Virtualization) Virtualization is a well-established technique Virtualization technique is implemented by a lot of software product, from commercial ones (like VMWare) to open-source ones (like XEN and virtual box)…. Modern CPUs provide supports to virtualization too, to make virtualization faster… Intel provides Intel VT And AMD provides AMD-V Virtualization technique is now a well-established technique that can be conveniently used by cloud computing service provider.

31 3 Large-scale Computing in Cloud
Large-scale computing for data mining problems on commodity hardware Challenges: How do you distribute computation? How can we make it easy to write distributed programs? How can you handle machine failures? One server may stay up 3 years (1,000 days) If you have 1,000 servers, expect to lose 1/day It is estimated that Google had 2M machines in 2013 2,000 machine fails every day!

32 3.1 Idea and Solution Issue: Copying data over a network takes time
Bring computation close to the data Store files multiple times for reliability MapReduce addresses these problems Google’s computational/data manipulation model Elegant way to work with big data Storage Infrastructure – File system Google: GFS; Hadoop: HDFS Programming model Map-reduce

33 Relationship MapReduce HDFS Hardware & OS HBase Hive Mahout Pig
ZooKeeper Hadoop Projects Hadoop

34 3.3 The Hadoop Distributed File System (HDFS)
With hundreds of machines at hand, failure is the norm rather than exception Traditional file storage system cannot cope with the scale and failure faced by large clusters The Hadoop Distributed File System (HDFS) is a natural solution to this problem Distributed File System Provides global file namespace Replica to ensure data recovery

35 3.3 The Hadoop Distributed File System (HDFS)
A HDFS instance may consist of thousands of server machines, each storing part of the file system’s data. Since we have huge number of components, and each component has non-trivial probability of failure, it means that there is always some component that is non-functional. Detection of faults and quick, automatic recovery from them are a core architectural goal of HDFS.

36 3.3 Data Characteristics Streaming data access
Batch processing rather than interactive user access Write-once-read-many: a file, once created, written and closed, need not be changed – this assumption simplifies coherency Other characteristics: Applications need streaming access to data Large data sets and files: gigabytes to terabytes size High aggregate data bandwidth Scale to hundreds of nodes in a cluster Tens of millions of files in a single instance A map-reduce application or web-crawler application fits perfectly with this model.

37 3.4 HDFS Architecture Master/slave architecture
Master: NameNode Slave: DataNode HDFS exposes a file system namespace (NameNode) and allows user data to be stored in files. A file is split into one or more blocks and set of blocks are stored in DataNodes. HDFS cluster consists of a single NameNode, a master server that manages the file system namespace and regulates access to files by clients. There are a number of DataNodes usually one per node in a cluster. The DataNodes manage storage attached to the nodes that they run on. DataNodes: serves read, write requests, performs block creation, deletion, and replication upon instruction from NameNode.

38 HDFS Architecture Namenode Metadata ops Client Block ops Read
Metadata(Name, replicas..) (/home/foo/data,6. .. Metadata ops Client Block ops Read Datanodes Datanodes B replication Blocks Rack2 Rack1 Write Client

39 3.5 File System Namespace Namenode maintains the file system.
Hierarchical file system with directories and files. Create, remove, move, rename, etc. Any meta information changes to the file system is recorded by the Namenode. An application can specify the number of replicas of the file needed: replication factor of the file. This information is stored in the Namenode.

40 3.5 Data Replication HDFS is designed to store very large files across machines in a large cluster. Each file is a sequence of blocks. All blocks in the file except the last are of the same size. Blocks are replicated for fault tolerance. Block size and replicas are configurable per file. The NameNode receives a Heartbeat and a BlockReport from each DataNode in the cluster. BlockReport contains all the blocks on a DataNode.

41 3.6 Replica Selection Replica selection for read operation: HDFS tries to minimize the bandwidth consumption and latency. If there is a replica on the Reader node then that is preferred. HDFS cluster may span multiple data centers: replica in the local data center is preferred over the remote one.

42 3.6 Safemode Startup Each DataNode checks in with Heartbeat and BlockReport. NameNode verifies that each block has acceptable number of replicas After a configurable percentage of safely replicated blocks check in with the NameNode, NameNode exits Safemode. It then makes the list of blocks that need to be replicated. NameNode then proceeds to replicate these blocks to other DataNodes. On startup NameNode enters Safemode. Replication of data blocks do not occur in Safemode.

43 3.7 Filesystem Metadata The HDFS namespace is stored by NameNode.
NameNode uses a transaction log called the EditLog to record every change that occurs to the filesystem meta data. For example, creating a new file Change replication factor of a file EditLog is stored in the NameNode’s local filesystem Entire filesystem namespace including mapping of blocks to files and file system properties is stored in a file FsImage. Stored in Namenode’s local filesystem.

44 3.7 NameNode Keeps image of entire file system namespace.
When the Namenode starts up Gets the FsImage and Editlog. Update FsImage with EditLog information. Stores a copy of the FsImage as a checkpoint. In case of crash Last checkpoint is recovered. 4GB of local RAM is sufficient to support the above data structures that represent the huge number of files and directories.

45 3.7 DataNode A DataNode stores data in files in its local file system.
Each block of HDFS is a separate file. These files are placed in different directories. Creation of new directory is determined by heuristics. When the filesystem starts up: Generates Blockreport. Sends this report to NameNode. DataNode has no knowledge about HDFS filesystem DataNode does not create all files in the same directory.

46 HDFS Summary Reliable distributed file system
Data kept in “chunks” spread across machines Each chunk replicated on different machine and racks Seamless recovery from disk or machine failure C0 C1 D0 D1 C2 C1 C0 E1 …. C2 D6 C3 C6 D0 C6 A2 B6

47 4 MapReduce Warm-up task Sample application
We have a huge text document Count the number of times each distinct word appears in the file Sample application Analyze web server logs to find popular URLs

48 4.1 Task: Word Count Using Unix tool chain, we can count the occurrences of words: Words (doc.txt) | sort | uniq –c Where words takes a file and outputs the words in it, one per line This way of counting captures the essence of MapReduce Mapper (done by words) Group by keys and sort (done by sort) Reducer (done by uniq) Hadoop handles the partition and parallelization

49 4.1 MapReduce: Overview Sequentially read a lot of data Map:
Extract something you care about Group by key: Sort and Shuffle Reduce: Aggregate, summarize, filter or transform Write the result

50 4.1 MapReduce Input: a set of key-value paris
Programmer must specifies two methods: Map(k,v) -> <k’, v’> Takes a key-value pair and outputs a set of key-value pairs There is one Map call for every (k,v) pair Reduce (k’, <v’>) -> <k’, v’’> All values v’ with the same key k’ are reduced together and processed in v’ order There is one Reduce function call per unique key k’

51 4.1 MapReduce: Word Count Example
Now that one document changes to a large corpus of documents

52 4.2 MapReduce: Word Count Example
// key: document name; value: text of the document Map(key, value): for each word w in value: emit(w, 1) // key: a word; value: an iterator over counts Reduce(key, values): result = 0 for each count v in values: results += v emit(key, result) In-class Practice

53 4.3 When To Use MapReduce? You have a cluster of computers
You are working with large dataset You are working with independent data (or assumed) The data processing can be casted into Map and Reduce operations We now know what MapReduce is doing… We will see when to use MapReduce. First of all, You should have a cluster of computers, since we need them to work in parallel Second, our input is a large dataset.. Otherwise, we don’t need a cluster computing. Third, we are working with independent data. That is to say, the data can be divided into many parts and processed individually…The results are the same as we process them together as a whole. Finally, the data processing can be casted into two operations: Map and Reduce…The answer of whether a data processing task can be casted into Map and Reduce is actually problem-specific.

54 4.3 MapReduce Implementations
MapReduce by Google Inc. Centric to Google’s searching engine Generate Google's index of the World Wide Web Open-source implementations Hadoop Yahoo! Search Webmap is a Hadoop application to produce data used in every Yahoo! Web search query And many others Disco, Skynet, … The MapReduce paradigm is initially implemented by Google and it is centric to its searching engine core algorithms. Google resorts to MapReduce to generate its index of the WWW There are many open source implementations. The most popular is Hadoop, which is used in Yahoo! Search Webmap used in every Yahoo! Web search query. And also there are many others, for example, Disco and Skynet…You can search them from the Internet if you are interested.

55 4.4 MapReduce: Environment
MapReduce environment takes care of: Partitioning the input data Scheduling the program’s execution across a set of machines Performing the group by key step Handling machine failures Managing required inter-machine communication

56 4.4 MapReduce Map: Read input and produces a set of key-value pairs
Group by key: Collect all pairs with the same key Reduce: Collect all values belonging to the key and output

57 4.4 MapReduce Move computation to the data C0 C1 D0 D1 C2 C1 C3 E1 ….
B6 DataNode DataNode DataNode DataNode TaskTracker TaskTracker TaskTracker TaskTracker Bring computation directly to the data! DataNode also serve as compute servers

58 4.5 Data Flow Input and final output are stored on a distributed file system (FS): Scheduler tries to schedule map tasks “close” to physical storage location of input data Intermediate results are stored on local FS of Map and Reduce workers Output is often input to another MapReduce task

59 4.5 Coordination: Master Master node takes care of coordination:
Task status: (idle, in-progress, completed) Idle tasks get scheduled as workers become available When a map task completes, it sends the master the location and sizes of its R intermediate files, one for each reducer Master pushes this info to reducers Master pings workers periodically to detect failures

60 4.6 Dealing with Failures Map worker failure Reduce worker failure
Map tasks completed or in-progress at worker are reset to idle Reduce workers are notified when task is rescheduled on another worker Reduce worker failure Only in-progress tasks are reset to idle Reduce task is restarted Master failure MapReduce task is aborted and client is notified.

61 4.6 How Many Map and Reduce Jobs?
M map tasks, R reduce tasks Rule of a thumb: Make M much larger than the number of nodes in the cluster One chunk per map is common Improves dynamic load balancing and speeds up recovery from worker failures Usually R is smaller than M Output is spread across R files

62 5 Summary Cloud computing can provide elastic Internet-based services Host computing systems in an on-demand, scalable manner For cost-down purpose Cloud computing systems are typically implemented as service-based systems: IaaS, PaaS, SaaS. Virtualization technology is key to implement cloud computing systems Hadoop Distributed File System handles failures systematically MapReduce is an emerging computing paradigm that can be implemented with cloud computing systems This is the summary of the lecture.


Download ppt "Cloud Computing Systems"

Similar presentations


Ads by Google