Grid Computing 1
Grid Computing 1
INDEX
1. Introduction 1.1Grid Concept and Grid Vision 2. What is Grid? 3. The Grid problem 4. Why Grids and Why now? 5. The Grid Architecture Description 5.1 Fabric Layer: Interfaces to Local Control 5.2 Connectivity Layer: Communicating Easily and Securely 5.3 Resource Layer: Sharing Single Resource 5.4 Collective: Coordinating Multiple Resources 6. Cluster Computing Grid Computing Internet 7. The Future: All Software is Network- Centric 8. Benefits of Grid Computing 9. Summary
technology
base
that
addresses
significant aspects of this problem. In the process, we develop a detailed architecture and roadmap for current and 1. INTRODUCTION: Grid computing is a method of harnessing the power of many computers in a network to solve problems requiring a large number of processing cyc1es and involving huge amounts of data. The grid computing helps in exploiting underutilized resources research for resources, achieving and have parallel CPU capacity; provide virtual collaboration might reliability. Although commercial and organizations collaborative or monetary reasons to share resources, they are unlikely to adopt such a distributed infrastructure until they can rely on the confidentiality of the communication, the integrity of their data and resources, and the privacy of the user information. In other words, large-scale deployment of grids will occur when users can count on their security. Our purpose in this article is to argue that the Grid concept is indeed motivated by a real and specific problem and that there is an emerging, well-defined Grid future Grid technologies. Furthermore, we assert that while Grid technologies are currently distinct from other major technology trends, such as Internet, enterprise, distributed, and peer-to-peer computing, these other trends can benefit significantly from growing into the problem space addressed by Grid technologies. 1.1 GRID CONCEPT AND GRID VISION The following are points to be noted when comparing our Grid with that of a power grid. On-demand access to ubiquitous distributed computing Transparent access to multipetabyte distributed data bases Easy to plug resources into Complexity of the infrastructure is hidden When the network
is as fast as the computer's internal links, the machine disintegrates across the net into a set of special purpose appliances E-Science and information utilities (Taylor) Science through increasingly distributed done global
access to computers, software, data, and other resources, as is required by a range of collaborative problem-solving and resource-brokering strategies emerging in industry, science, and engineering. This sharing is, necessarily, highly controlled, with resource providers and consumers defining clearly and carefully just what is shared, who is allowed to share, and the conditions under which sharing occurs. A set of individuals and/or institutions defined by such sharing rules form what we call a virtual organization (VO). VOs vary tremendously in their purpose, scope, size, duration, structure, community, Nevertheless, underlying and careful technology sociology. study of requirements
collaborations between people, enabled by the Internet Using very large data collections, terascale computing resources, and high performance visualisation Derived from instruments and facilities controlled and shared via the infrastructure Scaling x1000 in processing power, data, bandwidth 2. What is Grid? Resource sharing & coordinated problem solving in dynamic, multiinstitutional virtual organizations. 3. THE GRID PROBLEM The real and specific problem that underlies coordinated the Grid concept sharing is and resource
leads us to identify a broad set of common concerns and requirements. In particular, we see a need for highly flexible sharing relationships, ranging from client-server to peer-to-peer; for sophisticated and precise levels of control over how shared resources are used, including fine-grained and multistakeholder access control, delegation, and application of local and global policies; for sharing of varied resources, ranging from programs, files, and data to
problem solving in dynamic, multiinstitutional virtual organizations. The sharing that we are concerned with is not primarily file exchange but rather direct
computers, sensors, and networks; and for diverse usage modes, ranging from single user to multi-user and from performance sensitive to cost-sensitive and hence embracing issues of quality of service, scheduling, co-allocation, and accounting. Current distributed computing technologies do not address the concerns and requirements just listed. For
An
emergency
response
team
couples real time data, weather model, population data A multidisciplinary analysis in aerospace couples code and data in four companies A home user invokes architectural design functions at an application service provider Scientists new product A community group pools members PCs to analyze alternative designs for a local road Why Now? The following are the reasons why now we are concentrating on Grids: Moores law improvements in computing produce highly functional end systems The Internet and burgeoning wired and wireless provide universal connectivity Changing modes of working and problem solving emphasize teamwork, computation Network exponentials produce dramatic changes in geometry and geography working for a multinational soap company design a
example, current Internet technologies address communication and information exchange among computers but do not provide integrated approaches to the coordinated use of resources at multiple sites for computation. Business-tobusiness exchanges focus on information sharing (often via centralized servers). 4. WHY GRIDS AND WHY NOW? A biochemist exploits computers to screen compounds in an hour 1,000 physicists worldwide pool resources for petaop analyses of petabytes of data Civil engineers collaborate to design, execute, & analyze shake table experiments Climate annotate, scientists & analyze visualize, terabyte 10, 000 100,000
simulation datasets
The network potentials are as follows: Network vs. computer performance A Computer speed doubles every 18 months Network speed doubles every 9 months Difference = order of magnitude per 5 years comparison of networks vs. Computer performance from 1986 to 2000 is as follows: The Computers performance is increased 500 times. 5.THE GRID ARCHITECTURE DESCRIPTION Our goal in describing our Grid architecture is not to provide a complete enumeration of all required protocols (and services, APIs, and SDKs) but rather to identify requirements for general classes of component. The result is an extensible, open architectural structure within which can be placed solutions to key VO requirements. Our architecture and the subsequent discussion organize components into layers, as shown in Figure. Components times where as Network performance is increased by 340,000
within
each
layer but
share can
common build on
characteristics
capabilities and behaviors provided by any lower layer. In specifying the various layers of the Grid architecture, we follow the principles of the hourglass model. The narrow neck of the hourglass defines a small set of core abstractions and protocols (e.g., TCP and HTTP in the Internet), onto which many different high-level behaviors can be mapped (the top of the hourglass), and which themselves can be mapped onto many different underlying technologies (the base of the hourglass). pplications Diverse global services A
The Connectivity layer defines Local OS core communication and authentication protocols required for Grid-specific network transactions. Communication protocols enable the exchange of data between Fabric layer to resources. build on provide Authentication communication protocols services
cryptographically secure mechanisms for 5.1 FABRIC LAYER: INTERFACES TO LOCAL CONTROL The Grid Fabric layer provides the resources to which shared access is mediated example, storage by Grid protocols: for computational systems, catalogs, resources, network verifying the identity of users and resources. Communication requirements include transport, routing, and naming. While alternatives certainly exist, we assume here that these protocols are drawn from the TCP/IP protocol stack: specifically, the Internet (IP and ICMP), transport (TCP, UDP), and application (DNS, OSPF, RSVP, etc.) layers of the Internet layered protocol architecture 5.3 RESOURCE LAYER: SHARING SINGLE RESOURCE The Resource layer builds on Connectivity layer communication and authentication protocols to define protocols (and APIs and SDKs) for the secure negotiation, initiation, monitoring, control, accounting, and payment of sharing operations on individual resources. Resource layer implementations of these protocols call
resources, and sensors. A resource may be a logical entity, such as a distributed file system, computer cluster, or distributed computer pool; in such cases, a resource implementation may involve internal protocols. There is thus a tight and subtle interdependence between the functions implemented at the Fabric level, on the one hand, and the sharing operations supported, on the other. 5.2 CONNECTIVITYLAYER: COMMUNICATING EASILY AND SECURELY
Fabric layer functions to access and control local resources. Resource layer protocols are concerned entirely with individual resources and hence ignore issues of global state and atomic actions across distributed collections; such issues are the concern of the Collective layer discussed next. Two primary classes of Resource layer protocols can be distinguished: Information protocols are used to obtain information about the structure and state of a resource, for example, its configuration, current load, and usage policy (e.g., cost). Management protocols are used to negotiate access to a shared resource, specifying, for example, resource requirements (including advanced reservation and quality of service) and the operation(s) to be performed, such as process creation, or data access. Since management sharing protocols are responsible for instantiating relationships, they must serve as a policy application point, ensuring that the requested protocol operations are consistent 5.4
with the policy under which the resource is to be shared. Issues that must be considered include accounting protocol and may payment. also A support
monitoring the status of an operation and controlling (for example, operation. COLLECTIVE:COORDINATING MULTIPLE RESOURCES While the Resource layer is focused on interactions with a single resource, architecture the next layer in the and contains protocols terminating) the
services (and APIs and SDKs) that are not associated with any one specific resource but rather are global in nature and capture interactions across collections of resources. For this reason, we refer to the next layer of the architecture as the Collective layer. Because Collective components build on the narrow Resource and Connectivity layer neck in the protocol hourglass, they can implement a wide variety of sharing behaviors without placing new requirements on the resources being shared.
6. CLUSTER COMPUTING GRID COMPUTING INTERNET Cluster computing focuses on platforms consisting of often nodes homogeneous in a single interconnected
for
high
performance
administrative domain.
Cluster
components
can
be
shared or dedicated
7. THE FUTURE: ALL SOFTWARE IS NETWORK- CENTRIC We dont or build lease or buy we
distributed computations. Web focuses on platforms consisting of any combination which of resources and networks support naming
anymore,
required
o When I walk into a room, need to solve a problem, need to communicate A computer is a dynamically, often collaboratively constructed collection of processors, data sources, sensors, networks o Similar And Thus Reduced barriers to access mean that we do much more observations
Web consists of very diverse set of computational, and storage, other communication, number of users
computing, and more interesting computing, than today => Many more components (& services);massive parallelism All resources are owned by others => Sharing (for fun or profit) is fundamental; trust, policy, negotiation, payment All computing is performed on unfamiliar systems => Dynamic behaviors, discovery, adaptively, failure. 8. BENEFITS OF GRID COMPUTING
enables This is
companies to access and share especially beneficial to the life research communities, where enormous volumes of data are generated and analysed during any given day.
Grid computing enables widely dispersed organizations to easily collaborate everything applications on projects by creating the ability to share from and software data, to
Grid
computing to
enables aggregate
engineering blueprints. Grid computing can create a more robust and resilient IT infrastructure respond disasters.
organizations
resources within an entire IT infrastructure no matter where in the world they are located. It eliminates situations where one site is running on maximum capacity,
better minor
able or
to
to
major
A grid can harness the idle processing in various cycles that are across For would Tokyo available in desktop PCs located locations zones. that PCs multiple example, a time
while
others
have
cycles to spare. Organizations can dramatically improve the quality and speed of the products and services they deliver, while reducing IT costs by sharing. enabling and transparent resource collaboration
utilized during the day by its North American operations 9. SUMMARY . The Grid problem: Resource sharing solving in dynamic, & coordinated problem
multiGrid enable
institutional architecture: