Tài liệu hạn chế xem trước, để xem đầy đủ mời bạn chọn Tải xuống
1
/ 34 trang
THÔNG TIN TÀI LIỆU
Thông tin cơ bản
Định dạng
Số trang
34
Dung lượng
841,05 KB
Nội dung
xxxiv Cloud Computing this Act can be punished if the offense is committed for purposes of com- mercial advantage, malicious destruction or damage, or private commercial gain, or in furtherance of any criminal or tortious act in violation of the Constitution or laws of the United States or any state by a fine or imprison- ment or both for not more than five years in the case of a first offense. For a second or subsequent offense, the penalties stiffen to fine or imprisonment for not more than 10 years, or both. What Are the Key Characteristics of Cloud Computing? There are several key characteristics of a cloud computing environment. Service offerings are most often made available to specific consumers and small businesses that see the benefit of use because their capital expenditure is minimized. This serves to lower barriers to entry in the marketplace, since the infrastructure used to provide these offerings is owned by the cloud ser- vice provider and need not be purchased by the customer. Because users are not tied to a specific device (they need only the ability to access the Inter- net) and because the Internet allows for location independence, use of the cloud enables cloud computing service providers’ customers to access cloud- enabled systems regardless of where they may be located or what device they choose to use. Multitenancy 9 enables sharing of resources and costs among a large pool of users. Chief benefits to a multitenancy approach include: Centralization of infrastructure and lower costs Increased peak-load capacity Efficiency improvements for systems that are often underutilized Dynamic allocation of CPU, storage, and network bandwidth Consistent performance that is monitored by the provider of the service Reliability is often enhanced in cloud computing environments because service providers utilize multiple redundant sites. This is attractive to enter- 9. http://en.wikipedia.org/wiki/Multitenancy, retrieved 5 Jan 2009. Multitenancy refers to a principle in software architecture where a single instance of the software runs on a SaaS vendor’s servers, serving multiple client organizations (tenants). Intro.fm Page xxxiv Friday, May 22, 2009 11:24 AM What Are the Key Characteristics of Cloud Computing? xxxv prises for business continuity and disaster recovery reasons. The drawback, however, is that IT managers can do very little when an outage occurs. Another benefit that makes cloud services more reliable is that scalabil- ity can vary dynamically based on changing user demands. Because the ser- vice provider manages the necessary infrastructure, security often is vastly improved. As a result of data centralization, there is an increased focus on protecting customer resources maintained by the service provider. To assure customers that their data is safe, cloud providers are quick to invest in dedi- cated security staff. This is largely seen as beneficial but has also raised con- cerns about a user’s loss of control over sensitive data. Access to data is usually logged, but accessing the audit logs can be difficult or even impossi- ble for the customer. Data centers, computers, and the entire associated infrastructure needed to support cloud computing are major consumers of energy. Sus- tainability of the cloud computing model is achieved by leveraging improve- ments in resource utilization and implementation of more energy-efficient systems. In 2007, Google, IBM, and a number of universities began work- ing on a large-scale cloud computing research project. By the summer of 2008, quite a few cloud computing events had been scheduled. The first annual conference on cloud computing was scheduled to be hosted online April 20–24, 2009. According to the official web site: This conference is the world’s premier cloud computing event, cov- ering research, development and innovations in the world of cloud computing. The program reflects the highest level of accomplish- ments in the cloud computing community, while the invited pre- sentations feature an exceptional lineup of speakers. The panels, workshops, and tutorials are selected to cover a range of the hottest topics in cloud computing. 10 It may seem that all the world is raving about the potential of the cloud computing model, but most business leaders are likely asking: “What is the market opportunity for this technology and what is the future potential for long-term utilization of it?” Meaningful research and data are difficult to find at this point, but the potential uses for cloud computing models are wide. Ultimately, cloud computing is likely to bring supercomputing capa- 10. http://cloudslam09.com, retireved 5 Jan 09. Intro.fm Page xxxv Friday, May 22, 2009 11:24 AM xxxvi Cloud Computing bilities to the masses. Yahoo, Google, Microsoft, IBM, and others are engaged in the creation of online services to give their users even better access to data to aid in daily life issues such as health care, finance, insur- ance, etc. Challenges for the Cloud The biggest challenges these companies face are secure data storage, high- speed access to the Internet, and standardization. Storing large amounts of data that is oriented around user privacy, identity, and application-specific preferences in centralized locations raises many concerns about data protec- tion. These concerns, in turn, give rise to questions regarding the legal framework that should be implemented for a cloud-oriented environment. Another challenge to the cloud computing model is the fact that broadband penetration in the United States remains far behind that of many other countries in Europe and Asia. Cloud computing is untenable without high- speed connections (both wired and wireless). Unless broadband speeds are available, cloud computing services cannot be made widely accessible. Finally, technical standards used for implementation of the various com- puter systems and applications necessary to make cloud computing work have still not been completely defined, publicly reviewed, and ratified by an oversight body. Even the consortiums that are forming need to get past that hurdle at some point, and until that happens, progress on new products will likely move at a snail’s pace. Aside from the challenges discussed in the previous paragraph, the reli- ability of cloud computing has recently been a controversial topic in tech- nology circles. Because of the public availability of a cloud environment, problems that occur in the cloud tend to receive lots of public exposure. Unlike problems that occur in enterprise environments, which often can be contained without publicity, even when only a few cloud computing users have problems, it makes headlines. In October 2008, Google published an article online that discussed the lessons learned from hosting over a million business customers in the cloud computing model. 11 Google‘s personnel measure availability as the average uptime per user based on server-side error rates. They believe this reliability metric allows a true side-by-side comparison with other solutions. Their 11. Matthew Glotzbach, Product Management Director, Google Enterprise, “What We Learned from 1 Million Businesses in the Cloud,” http://googleblog.blogspot.com/2008/10/what- we-learned-from-1-million.html, 30 Oct 2008. Intro.fm Page xxxvi Friday, May 22, 2009 11:24 AM Challenges for the Cloud xxxvii measurements are made for every server request for every user, every moment of every day, and even a single millisecond delay is logged. Google analyzed data collected over the previous year and discovered that their Gmail application was available to everyone more than 99.9% of the time. One might ask how a 99.9% reliability metric compares to conven- tional approaches used for business email. According to the research firm Radicati Group, 12 companies with on-premises email solutions averaged from 30 to 60 minutes of unscheduled downtime and an additional 36 to 90 minutes of planned downtime per month, compared to 10 to 15 min- utes of downtime with Gmail. Based on analysis of these findings, Google claims that for unplanned outages, Gmail is twice as reliable as a Novell GroupWise solution and four times more reliable than a Microsoft Exchange-based solution, both of which require companies to maintain an internal infrastructure themselves. It stands to reason that higher reliability will translate to higher employee productivity. Google discovered that Gmail is more than four times as reliable as the Novell GroupWise solution and 10 times more reliable than an Exchange-based solution when you fac- tor in planned outages inherent in on-premises messaging platforms. Based on these findings, Google was confident enough to announce publicly in October 2008 that the 99.9% service-level agreement offered to their Premier Edition customers using Gmail would be extended to Google Calendar, Google Docs, Google Sites, and Google Talk. Since more than a million businesses use Google Apps to run their businesses, Google has made a series of commitments to improve communications with customers during any outages and to make all issues visible and transparent through open user groups. Since Google itself runs on its Google Apps platform, the commitment they have made has teeth, and I am a strong advocate of “eat- ing your own dog food.” Google leads the industry in evolving the cloud computing model to become a part of what is being called Web 3.0—the next generation of Internet. 13 In the following chapters, we will discuss the evolution of computing from a historical perspective, focusing primarily on those advances that led to the development of cloud computing. We will discuss in detail some of the more critical components that are necessary to make the cloud com- 12. The Radicati Group, 2008, “Corporate IT Survey—Messaging & Collaboration, 2008– 2009,” http://www.marketwatch.com/news/story/The-Radicati-Group-Releases-New/ story.aspx?guid=%7B80D6388A-731C-457F-9156-F783B3E3C720%7D, retrieved 12 Feb 2009. 13. http://en.wikipedia.org/wiki/Web_3.0, retrieved 5 Jan 2009. Intro.fm Page xxxvii Friday, May 22, 2009 11:24 AM xxxviii Cloud Computing puting paradigm feasible. Standardization is a crucial factor in gaining widespread adoption of the cloud computing model, and there are many different standards that need to be finalized before cloud computing becomes a mainstream method of computing for the masses. This book will look at those various standards based on the use and implementation issues surrounding cloud computing. Management of the infrastructure that is maintained by cloud computing service providers will also be dis- cussed. As with any IT, there are legal considerations that must be addressed to properly protect user data and mitigate corporate liability, and we will cover some of the more significant legal issues and even some of the philosophical issues that will most likely not be resolved without adoption of a legal framework. Finally, this book will take a hard look at some of the cloud computing vendors that have had significant success and examine what they have done and how their achievements have helped to shape cloud computing. Intro.fm Page xxxviii Friday, May 22, 2009 11:24 AM 1 Chapter 1 The Evolution of Cloud Computing 1.1 Chapter Overview It is important to understand the evolution of computing in order to get an appreciation of how we got into the cloud environment. Looking at the evo- lution of the computing hardware itself, from the first generation to the cur- rent (fourth) generation of computers, shows how we got from there to here. The hardware, however, was only part of the evolutionary process. As hardware evolved, so did software. As networking evolved, so did the rules for how computers communicate. The development of such rules, or proto- cols, also helped drive the evolution of Internet software. Establishing a common protocol for the Internet led directly to rapid growth in the number of users online. This has driven technologists to make even more changes in current protocols and to create new ones. Today, we talk about the use of IPv6 (Internet Protocol version 6) to mitigate address- ing concerns and for improving the methods we use to communicate over the Internet. Over time, our ability to build a common interface to the Internet has evolved with the improvements in hardware and software. Using web browsers has led to a steady migration away from the traditional data center model to a cloud-based model. Using technologies such as server virtualization, parallel processing, vector processing, symmetric multipro- cessing, and massively parallel processing has fueled radical change. Let’s take a look at how this happened, so we can begin to understand more about the cloud. In order to discuss some of the issues of the cloud concept, it is impor- tant to place the development of computational technology in a historical context. Looking at the Internet cloud’s evolutionary development, 1 and the problems encountered along the way, provides some key reference points to help us understand the challenges that had to be overcome to develop the Internet and the World Wide Web (WWW) today. These challenges fell Chap1.fm Page 1 Friday, May 22, 2009 11:24 AM 2 Cloud Computing into two primary areas, hardware and software. We will look first at the hardware side. 1.2 Hardware Evolution Our lives today would be different, and probably difficult, without the ben- efits of modern computers. Computerization has permeated nearly every facet of our personal and professional lives. Computer evolution has been both rapid and fascinating. The first step along the evolutionary path of computers occurred in 1930, when binary arithmetic was developed and became the foundation of computer processing technology, terminology, and programming languages. Calculating devices date back to at least as early as 1642, when a device that could mechanically add numbers was invented. Adding devices evolved from the abacus. It was a significant mile- stone in the history of computers. In 1939, the Berry brothers invented an electronic computer capable of operating digitally. Computations were per- formed using vacuum-tube technology. In 1941, the introduction of Konrad Zuse’s Z3 at the German Labora- tory for Aviation in Berlin was one of the most significant events in the evo- lution of computers because this machine supported both floating-point and binary arithmetic. Because it was a “Turing-complete” device, 2 it is con- sidered to be the very first computer that was fully operational. A program- ming language is considered Turing-complete if it falls into the same computational class as a Turing machine, meaning that it can perform any calculation a universal Turing machine can perform. This is especially sig- nificant because, under the Church-Turing thesis, 3 a Turing machine is the embodiment of the intuitive notion of an algorithm. Over the course of the next two years, computer prototypes were built to decode secret German messages by the U.S. Army. 1. Paul Wallis, “A Brief History of Cloud Computing: Is the Cloud There Yet? A Look at the Cloud’s Forerunners and the Problems They Encountered,” http://soa.sys-con.com/node/ 581838, 22 Aug 2008, retrieved 7 Jan 2009. 2. According to the online encyclopedia Wikipedia, “A computational system that can com- pute every Turing-computable function is called Turing-complete (or Turing-powerful). Alternatively, such a system is one that can simulate a universal Turing machine.” http://en.wikipedia.org/wiki/Turing_complete, retrieved 17 Mar 2009. 3. http://esolangs.org/wiki/Church-Turing_thesis, retrieved 10 Jan 2009. Chap1.fm Page 2 Friday, May 22, 2009 11:24 AM Hardware Evolution 3 1.2.1 First-Generation Computers The first generation of modern computers can be traced to 1943, when the Mark I and Colossus computers (see Figures 1.1 and 1.2) were developed, 4 albeit for quite different purposes. With financial backing from IBM (then International Business Machines Corporation), the Mark I was designed and developed at Harvard University. It was a general-purpose electrome- chanical programmable computer. Colossus, on the other hand, was an elec- tronic computer built in Britain at the end 1943. Colossus was the world’s first programmable, digital, electronic, computing device. First-generation computers were built using hard-wired circuits and vacuum tubes (thermi- onic valves). Data was stored using paper punch cards. Colossus was used in secret during World War II to help decipher teleprinter messages encrypted by German forces using the Lorenz SZ40/42 machine. British code breakers referred to encrypted German teleprinter traffic as “Fish” and called the SZ40/42 machine and its traffic “Tunny.” 5 To accomplish its deciphering task, Colossus compared two data streams read at high speed from a paper tape. Colossus evaluated one data stream representing the encrypted “Tunny,” counting each match that was discovered based on a programmable Boolean function. A comparison with the other data stream was then made. The second data stream was generated internally and designed to be an electronic simulation of the 4. http://trillian.randomstuff.org.uk/~stephen/history, retrieved 5 Jan 2009. 5. http://en.wikipedia.org/wiki/Colossus_computer, retrieved 7 Jan 2009. Figure 1.1 The Harvard Mark I computer. (Image from www.columbia.edu/acis/ history/mark1.html, retrieved 9 Jan 2009.) Chap1.fm Page 3 Friday, May 22, 2009 11:24 AM 4 Cloud Computing Lorenz SZ40/42 as it ranged through various trial settings. If the match count for a setting was above a predetermined threshold, that data match would be sent as character output to an electric typewriter. 1.2.2 Second-Generation Computers Another general-purpose computer of this era was ENIAC (Electronic Numerical Integrator and Computer, shown in Figure 1.3), which was built in 1946. This was the first Turing-complete, digital computer capable of being reprogrammed to solve a full range of computing problems, 6 although earlier machines had been built with some of these properties. ENIAC’s original purpose was to calculate artillery firing tables for the U.S. Army’s Ballistic Research Laboratory. ENIAC contained 18,000 thermionic valves, weighed over 60,000 pounds, and consumed 25 kilowatts of electri- cal power per hour. ENIAC was capable of performing 100,000 calculations a second. Within a year after its completion, however , the invention of the transistor meant that the inefficient thermionic valves could be replaced with smaller, more reliable components, thus marking another major step in the history of computing. Figure 1.2 The British-developed Colossus computer. (Image from www.com- puterhistory.org, retrieved 9 Jan 2009.) 6. Joel Shurkin, Engines of the Mind: The Evolution of the Computer from Mainframes to Microprocessors, New York: W. W. Norton, 1996. Chap1.fm Page 4 Friday, May 22, 2009 11:24 AM Hardware Evolution 5 Transistorized computers marked the advent of second-generation computers, which dominated in the late 1950s and early 1960s. Despite using transistors and printed circuits, these computers were still bulky and expensive. They were therefore used mainly by universities and govern- ment agencies. The integrated circuit or microchip was developed by Jack St. Claire Kilby, an achievement for which he received the Nobel Prize in Physics in 2000. 7 In congratulating him, U.S. President Bill Clinton wrote, “You can take pride in the knowledge that your work will help to improve lives for generations to come.” It was a relatively simple device that Mr. Kilby showed to a handful of co-workers gathered in the semiconductor lab at Texas Instruments more than half a century ago. It was just a transistor and a few other components on a slice of germanium. Little did this group real- ize that Kilby’s invention was about to revolutionize the electronics industry. 1.2.3 Third-Generation Computers Kilby’s invention started an explosion in third-generation computers. Even though the first integrated circuit was produced in September 1958, Figure 1.3 The ENIAC computer. (Image from www.mrsec.wisc.edu/ /computer/ eniac.html, retrieved 9 Jan 2009.) 7. http://www.ti.com/corp/docs/kilbyctr/jackstclair.shtml, retrieved 7 Jan 2009. Chap1.fm Page 5 Friday, May 22, 2009 11:24 AM [...]... and watered with different volumes of data depending on the tasks they are processing Running a dataintensive process with disparate data sources can create a bottleneck in the I/O, causing the CPU to run inefficiently, and affecting economic viability .20 19 20 Paul Wallis, “Keystones and Rivets,” http://it.toolbox.com/blogs/keystones -and- rivets/understanding -cloud- computing- 22 611, retrieved 2 Jan 20 09... http://it.toolbox.com/blogs/keystones -and- rivets/understanding -cloud- computing- 22 611, retrieved 2 Jan 20 09 Ibid 22 Cloud Computing The issues of storage management, migration of data, and security provisioning were key to any proposed solution in order for a grid model to succeed A toolkit called Globus21 was created to solve these issues, but the infrastructure hardware available still has not progressed to a level where true grid computing can be wholly achieved The Globus... Internet users has also evolved and changed the way business is done Chapter 2 Web Services Delivered from the Cloud 2. 1 Chapter Overview In this chapter we will examine some of the web services delivered from the cloud We will take a look at Communication-as-a-Service (CaaS) and explain some of the advantages of using CaaS Infrastructure is also a service in cloud land, and there are many variants on... for building grid systems and applications It is being developed and maintained by the Globus Alliance 22 and many others all over the world The Globus Alliance has grown into community of organizations and individuals developing fundamental technologies to support the grid model The toolkit provided by Globus allows people to share computing power, databases, instruments, and other online tools securely... importance of knowing about the evolution of computing in order to get an appreciation of how we got into the cloud environment Examining the history of computing hardware and software helps us to understand why we are standing on the shoulders of giants We discussed how the rules computers use to communicate came about, and how the development of networking and communications protocols has helped drive... books, records, and communications, and which is mechanized so that it may be consulted with exceeding speed and flexibility.”9 Figure 1.5 Vannevar Bush’s MEMEX (Image from www.icesi.edu.co/ blogs_estudiantes/luisaulestia, retrieved 9 Jan 20 09.) 8 9 http://en.wikipedia.org/wiki/Vannevar_Bush, retrieved 7 Jan 20 09 http://www.livinginternet.com/i/ii_summary.htm, retrieved 7 Jan 20 09 8 Cloud Computing After... infrastructure is managed in cloud environments When vendors outsource Infrastructure-as-a-Service (IaaS), it relies heavily on modern on-demand computing technology and high-speed networking We will look at some vendors who provide Software-as-a-Service (SaaS), such as Amazon.com with their elastic cloud platform, and foray into the implementation issues, the characteristics, benefits, and architectural maturity... Environment. 12 SAGE was the most ambitious computer project ever undertaken at the time, and it required over 800 programmers and the technical resources of some of America’s largest corporations SAGE was started in the 1950s and became operational by 1963 It remained in continuous operation for over 20 years, until 1983 10 11 12 http://www.theatlantic.com/doc/194507/bush, retrieved 7 Jan 20 09 http://en.wikipedia.org/wiki/Norbert_Wiener,... after the expiration of the timer, and then dispatches the next process in the queue The process whose time has expired is placed at the end of the queue If a process is still running at the end of a time slice, the CPU is interrupted and the process goes to the end 25 26 http://searchdatacenter.techtarget.com/sDefinition/0,,sid80_gci2 127 47,00.html, retrieved 10 Jan 20 09 http://choices.cs.uiuc.edu/~f-kon/RoundRobin/node1.html,... Machine, and the Ultracomputer In data mining, there is a need to perform multiple searches of a static database The earliest massively parallel 27 http://en.wikipedia.org/wiki/Massive_parallel_processing, retrieved 10 Jan 20 09 28 Cloud Computing processing systems all used serial computers as individual processing units in order to maximize the number of units available for a given size and cost . Ultimately, cloud computing is likely to bring supercomputing capa- 10. http://cloudslam09.com, retireved 5 Jan 09. Intro.fm Page xxxv Friday, May 22 , 20 09 11 :24 AM xxxvi Cloud Computing . http://www.marketwatch.com/news/story/The-Radicati-Group-Releases-New/ story.aspx?guid=%7B80D6388A-731C-457F-9156-F783B3E3C 720 %7D, retrieved 12 Feb 20 09. 13. http://en.wikipedia.org/wiki/Web_3.0, retrieved 5 Jan 20 09. Intro.fm Page xxxvii Friday, May 22 , 20 09 11 :24 AM xxxviii Cloud Computing puting. of Cloud Computing: Is the Cloud There Yet? A Look at the Cloud s Forerunners and the Problems They Encountered,” http://soa.sys-con.com/node/ 581838, 22 Aug 20 08, retrieved 7 Jan 20 09. 2. According