Grid computing

Grid computing is a term for either of two broad subcategories of distributed computing:


 * Online computation or storage offered as a service supported by a pool of distributed computing resources, also known as utility computing, on-demand computing, or cloud computing. Data grids provide controlled sharing and management of large amounts of distributed data, often used in combination with computational grids.


 * The creation of a "virtual supercomputer" composed of a network of loosely-coupled computers, acting in concert to perform very large tasks. This technology has been applied to computationally-intensive scientific, mathematical, and academic problems through volunteer computing, and it is used in commercial enterprises for such diverse applications as drug discovery, economic forecasting, seismic analysis, and back-office data processing in support of e-commerce and web services.

What distinguishes grid computing from typical cluster computing systems is that grids tend to be more loosely coupled, heterogeneous, and geographically dispersed. Also, while a computing grid may be dedicated to a specialized application, it is often constructed with the aid of general purpose grid software libraries and middleware.



Grids versus conventional supercomputers
"Distributed" or "grid" computing in general is a special type of parallel computing which relies on complete computers (with onboard CPU, storage, power supply, network interface, etc.) connected to a network (private, public or the Internet) by a conventional network interface, such as Ethernet. This is in contrast to the traditional notion of a supercomputer, which has many processors connected by a local high-speed computer bus.

The primary advantage of distributed computing is that each node can be purchased as commodity hardware, which when combined can produce similar computing resources to a multiprocessor supercomputer, but at lower cost. This is due to the economies of scale of producing commodity hardware, compared to the lower efficiency of designing and constructing a small number of custom supercomputers. The primary performance disadvantage is that the various processors and local storage areas do not have high-speed connections. This arrangement is thus well-suited to applications in which multiple parallel computations can take place independently, without the need to communicate intermediate results between processors.

The high-end scalability of geographically dispersed grids is generally favorable, due to the low need for connectivity between nodes relative to the capacity of the public Internet. Conventional supercomputers also create physical challenges in supplying sufficient electricity and cooling capacity in a single location. Both supercomputers and grids can be used to run multiple parallel computations at the same time, which might be different simulations for the same project, or computations for completely different applications. The infrastructure and programming considerations needed to do this on each type of platform are different, however.

There are also some differences in programming and deployment. It can be costly and difficult to write programs so that they can be run in the environment of a supercomputer, which may have a custom operating system, or require the program to address concurrency issues. If a problem can be adequately parallelized, a "thin" layer of "grid" infrastructure can allow conventional, standalone programs to run on multiple machines (but each given a different part of the same problem). This makes it possible to write and debug programs on a single conventional machine, and eliminates complications due to multiple instances of the same program running in the same shared memory and storage space at the same time.

Design considerations and variations
One feature of distributed grids is that they can be formed from computing resources belonging to multiple individuals or organizations (known as multiple administrative domains). This can facilitate commercial transactions, as in utility computing, or make it easier to assemble volunteer computing networks.

One disadvantage of this feature is that the computers which are actually performing the calculations might not be entirely trustworthy. The designers of the system must thus introduce measures to prevent malfunctions or malicious participants from producing false, misleading, or erroneous results, and from using the system as an attack vector. This often involves assigning work randomly to different nodes (presumably with different owners) and checking that at least two different nodes report the same answer for a given work unit. Discrepancies would identify malfunctioning and malicious nodes.

Due to the lack of central control over the hardware, there is no way to guarantee that nodes will not drop out of the network at random times. Some nodes (like laptops or dialup Internet customers) may also be available for computation but not network communications for unpredictable periods. These variations can be accommodated by assigning large work units (thus reducing the need for continuous network connectivity) and reassigning work units when a given node fails to report its results as expected.

The impacts of trust and availability on performance and development difficulty can influence the choice of whether to deploy onto a dedicated computer cluster, to idle machines internal to the developing organization, or to an open external network of volunteers or contractors.

In many cases, the participating nodes must trust the central system not to abuse the access that is being granted, by interfering with the operation of other programs, mangling stored information, transmitting private data, or creating new security holes. Other systems employ measures to reduce the amount of trust "client" nodes must place in the central system such as placing applications in virtual machines.

Public systems or those crossing administrative domains (including different departments in the same organization) often result in the need to run on heterogeneous systems, using different operating systems and hardware architectures. With many languages, there is a tradeoff between investment in software development and the number of platforms that can be supported (and thus the size of the resulting network). Cross-platform languages can reduce the need to make this tradeoff, though potentially at the expense of high performance on any given node (due to run-time interpretation or lack of optimization for the particular platform).

Various middleware projects have created generic infrastructure, to allow diverse scientific and commercial projects to harness a particular associated grid, or for the purpose of setting up new grids. BOINC is a common one for academic projects seeking public volunteers; more are listed at the end of the article

CPU scavenging
CPU-scavenging, cycle-scavenging, cycle stealing, or shared computing creates a "grid" from the unused resources in a network of participants (whether worldwide or internal to an organization). Typically this technique uses desktop computer instruction cycles that would otherwise be wasted at night, during lunch, or even in the scattered seconds throughout the day when the computer is waiting for user input or slow devices.

Volunteer computing projects use the CPU scavenging model almost exclusively.

In practice, participating computers also donate some supporting amount of disk storage space, RAM, and network bandwidth, in addition to raw CPU power. Since nodes are apt to go "offline" from time to time, as their owners use their resources for their primary purpose, this model must be designed to handle such contingencies.

History
The term Grid computing originated in the early 1990s as a metaphor for making computer power as easy to access as an electric power grid in Ian Foster and Carl Kesselmans seminal work, "The Grid: Blueprint for a new computing infrastructure".

CPU scavenging and volunteer computing were popularized beginning in 1997 by distributed.net and later in 1999 by SETI@home to harness the power of networked PCs worldwide, in order to solve CPU-intensive research problems.

The ideas of the grid (including those from distributed computing, object oriented programming, cluster computing, web services and others) were brought together by Ian Foster, Carl Kesselman and Steve Tuecke, widely regarded as the "fathers of the grid ." They led the effort to create the Globus Toolkit incorporating not just computation management but also storage management, security provisioning, data movement, monitoring and a toolkit for developing additional services based on the same infrastructure including agreement negotiation, notification mechanisms, trigger services and information aggregation. While the Globus Toolkit remains the defacto standard for building grid solutions, a number of other tools have been built that answer some subset of services needed to create an enterprise or global grid.

During 2006 the term "Great Global Grid" was coined by Robert Marcus in his book "Emerging Technology Strategies". Many organizations working on grid computing name their servers "ggg. .com" as an analogy to the "www. .com" naming convention of the World Wide Web.

During 2007 the term cloud computing came into popularity. It is conceptually identical to the canonical Foster definition of grid computing below. In practice all clouds are grids, but not all grids manage a cloud.

Fastest virtual supercomputers

 * BOINC - 1,064 teraflops, as of May 03, 2008
 * Folding@Home - 2,231 teraflops, as of 03 May 2008

Seminal work done:

Current projects and applications
Grids offer a way to solve Grand Challenge problems like protein folding, financial modeling, earthquake simulation, and climate/weather modeling. Grids offer a way of using the information technology resources optimally inside an organization. They also provide a means for offering information technology as a utility for commercial and non-commercial clients, with those clients paying only for what they use, as with electricity or water.

Grid computing is presently being applied successfully by the National Science Foundation's National Technology Grid, NASA's Information Power Grid, Pratt & Whitney, Bristol-Myers Squibb, Co., and American Express.

One of the most famous cycle-scavenging networks is SETI@home, which was using more than 3 million computers to achieve 23.37 sustained teraflops (979 lifetime teraflops) as of September 2001.

As of March 2008, Folding@home had achieved peaks of 1502 teraflops on over 270,000 machines.

Another well-known project is distributed.net, which was started in 1997 and has run a number of successful projects in its history.

The NASA Advanced Supercomputing facility (NAS) has run genetic algorithms using the Condor cycle scavenger running on about 350 Sun and SGI workstations.

Until April 27, 2007, United Devices operated the United Devices Cancer Research Project based on its Grid MP product, which cycle scavenges on volunteer PCs connected to the Internet. As of June 2005, the Grid MP ran on about 3,100,000 machines.

The Enabling Grids for E-sciencE project, which is based in the European Union and includes sites in Asia and the United States, is a follow up project to the European DataGrid (EDG) and is arguably the largest computing grid on the planet. This, along with the LHC Computing Grid (LCG) have been developed to support the experiments using the CERN Large Hadron Collider. The LCG project is driven by CERN's need to handle huge amounts of data, where storage rates of several gigabytes per second (10 petabytes per year) are required. A list of active sites participating within LCG can be found online as can real time monitoring of the EGEE infrastructure. The relevant software and documentation is also publicly accessible.

Definitions
Today there are many definitions of Grid computing:
 * In his article "What is the Grid? A Three Point Checklist", Ian Foster lists these primary attributes:
 * Computing resources are not administered centrally.
 * Open standards are used.
 * Non-trivial quality of service is achieved.


 * Plaszczak/Wellner define grid technology as "the technology that enables resource virtualization, on-demand provisioning, and service (resource) sharing between organizations."
 * IBM defines grid computing as "the ability, using a set of open standards and protocols, to gain access to applications and data, processing power, storage capacity and a vast array of other computing resources over the Internet. A grid is a type of parallel and distributed system that enables the sharing, selection, and aggregation of resources distributed across 'multiple' administrative domains based on their (resources) availability, capacity, performance, cost and users' quality-of-service requirements"
 * An earlier example of the notion of computing as utility was in 1965 by MIT's Fernando Corbató. Fernando and the other designers of the Multics operating system envisioned a computer facility operating "like a power company or water company". http://www.multicians.org/fjcc3.html
 * Buyya defines a grid as "a type of parallel and distributed system that enables the sharing, selection, and aggregation of geographically distributed autonomous resources dynamically at runtime depending on their availability, capability, performance, cost, and users' quality-of-service requirements".
 * CERN, one of the largest users of grid technology, talk of The Grid: "a service for sharing computer power and data storage capacity over the Internet."

Grids can be categorized with a three stage model of departmental grids, enterprise grids and global grids. These correspond to a firm initially utilising resources within a single group i.e. an engineering department connecting desktop machines, clusters and equipment. This progresses to enterprise grids where non-technical staff's computing resources can be used for cycle-stealing and storage. A global grid is a connection of enterprise and departmental grids which can be used in a commercial or collaborative manner.

Concepts and related technology

 * Distributed computing
 * List of distributed computing projects
 * High-performance computing
 * Network Agility
 * Render farm
 * Semantic grid
 * Supercomputer
 * Computer cluster
 * Computon
 * Grid FileSystem
 * Edge computing
 * Metacomputing
 * Cloud Computing
 * Space based architecture (SBA)

Alliances and organizations

 * Open Grid Forum (Formerly Global Grid Forum)
 * Object Management Group

Production grids

 * Enabling Grids for E-sciencE
 * NorduGrid
 * Open Science Grid
 * OurGrid
 * Sun Grid
 * Xgrid
 * Distributed European Infrastructure for Supercomputing Applications DEISA
 * FusionGrid


 * INFN Production Grid
 * UC Grid

National Grid Projects

 * D-Grid (German)
 * Grid5000 (French)
 * GARUDA (Indian)
 * National Grid Service (UK)
 * Open Science Grid (USA)
 * VECC (Calcutta, India)
 * China Grid Project
 * INFN Grid (Italian)
 * KnowledgeGrid Malaysia
 * NAREGI Project
 * Singapore National Grid Project
 * Thai National Grid Project
 * LitGRID (Lithuanian)


 * Hellasgrid (Greek)
 * Swiss National Grid Association
 * Swegrid (Swedish National computational resource)
 * RDIG - Russian Data Intensive Grid
 * NorGrid - Norwegian Grid Initiative
 * Rogrid - Romanian Grid Initiative
 * Austrian Grid - Austrian Grid Initiative
 * TR-Grid - Turkish National Grid Initiative

Standards and APIs

 * A Simple API for Grid Applications (SAGA)
 * Common Object Request Broker Architecture (CORBA)
 * Distributed Resource Management Application API (DRMAA)
 * Grid Security Infrastructure (GSI)
 * Open Grid Services Architecture (OGSA)
 * Open Grid Services Infrastructure (OGSI)
 * Web Services Resource Framework (WSRF)

Software implementations and middleware

 * Advanced Resource Connector (NorduGrid's ARC)
 * Berkeley Open Infrastructure for Network Computing (BOINC)
 * Globus Toolkit
 * Load Sharing Facility (LSF)
 * Message Passing Interface (MPI)
 * Parallel Virtual Machine (PVM)
 * Simple Grid Protocol
 * Sun Grid Engine
 * ProActive
 * UNICORE
 * SDSC Storage resource broker (data grid)
 * gLite (EGEE)
 * NInf GridRPC
 * IceGrid
 * Invisionix Roaming System Remote (IRSR)
 * Java CoG Kit
 * GridWay


 * Alchemi
 * GridGain
 * gridGISTICS
 * Gridbus Middleware
 * Java Parallel Processing Framework (JPPF)
 * Vishwa
 * UGP
 * GRIA
 * iRODS (data grid)

News Sites

 * GridComputingPlanet - Part of the JupiterMedia empire
 * GridToday
 * International Science Grid This Week
 * Primeur magazine - HPC and Grid computing news

Information sites

 * Gridipedia - The European Grid Marketplace Contains reports and software components relating to Grid computing
 * Grid Computing Info Center
 * IEEE Distributed Systems Online, Grid Computing Section

Portals and grid projects

 * ASKALON
 * World Community Grid: Focuses on advancing scientific projects to benefit humanity, such as researching possible cures for cancer and muscular dystrophy, sequencing human genomes, finding better drug molecular structures to combat AIDS, etc. Open to anyone who wants to contribute idle PC processing time.
 * DistributedComputing.info Contains informations and links to mathematical, science, security, biological, rendering, economical, games and other world-wide distributed computing projects
 * Wikipedia article on the World Community Grid: Contains additional links for each project being conducted on the World Community Grid.
 * 3tera AppLogic
 * Gigaspaces
 * Appistry
 * DataSynapse
 * GridGain Systems
 * EnterTheGrid directory on Grid computing
 * EELA: E-Infrastructure shared between Europe and Latin America
 * Enabling Grids for E-sciencE (EGEE)
 * BREIN: Business objective driven reliable and intelligent grids for real business.
 * Fura GPL Ready to use grid
 * IBM Grid Computing website
 * ICEAGE: International Collaboration to Extend and Advance Grid Education
 * Java Parallel Processing Framework
 * ParadisEO, a C++ framework coupled with Globus and Condor-G for combinatorial and continuous optimization on grid support
 * GridSphere Portal Framework (JSR-168 compliant)
 * GridSummit.com
 * Gridalogy
 * BigBlueRiver
 * Grid Computing Now!: Knowledge Transfer Network
 * NICE EnginFrame: Grid computing portals for research and industry
 * Nivio: Virtual Desktop Based on Grid Computing
 * Rechenkraft.net (German)
 * gridGistics: service virtualization and grid computing.
 * myGrid: bioinformatics and eScience research project built by several UK universities and EMBL-EBI.
 * Consortsium SIRENE (Sharing Infrastructure and REsources iN Europe)
 * ECSS: European Community for Software and Software Services - Architectures, Infrastructures, Engineering
 * Vendor-independent documentation on Grid-compliant open source portals

Grid Simulators

 * SimGrid
 * GridSim
 * GangSim
 * Bricks
 * Grid Matrix Simulator

Grid Emulators

 * MicroGrid
 * Grid eXplorer

Articles

 * O'Reilly article about grid computing software
 * Grid Café, the place for everyone to learn about the Grid
 * Describing the Elephant: The Different Faces of IT as Service, positions grid in a broader context

Associations and conferences

 * Open Grid Forum
 * Israeli Association of Grid Technologies (IGT)
 * International Conference on Grid Computing (Grid XY)
 * International Symposium on Cluster Computing and the Grid (CCGrid XY)

Past events

 * GridWorld Washington 2006
 * IEEE Richmond Section Blog. Meeting — 5 October 2006:.

Grid-Computing Grid computing Grille de calcul 그리드 컴퓨팅 Komputasi grid Grid computing Grid computing Grid computing Grid computing グリッド・コンピューティング Siatka komputerowa Computação em grade Грид Ґрід мережі طنابی شمارندہ کاری 网格计算