Beowulf cluster: Difference between revisions

From Citizendium
Jump to navigation Jump to search
imported>Eric M Gearhart
(more clarification in the intro paragraph. <sigh> this needs more work)
imported>Eric M Gearhart
(Sectioning off parts, but trying to improve the 'flow' of the article)
Line 1: Line 1:
[[Image:Simple Beowulf Cluster Diagram.png|right|thumb|250px|In a Beowulf cluster, the network cables (shown in blue) that connect the eight nodes to the master serve as a [[bus]]]]
[[Image:Simple Beowulf Cluster Diagram.png|right|thumb|250px|In a Beowulf cluster, the network cables (shown in blue) that connect the eight nodes to the master serve as a [[Computer#How_computers_work:_the_stored_program_architecture|bus]]]]


A '''Beowulf cluster''' is a class of supercomputer, specifically one that utilizes "Commerical Off the Shelf" (COTS) hardware such as [[personal computer|personal computers]] and [[Computer network|Ethernet switches]] to link the machines together so that they act as one. This is known as [[distributed computing]].  
A '''Beowulf cluster''' is a class of supercomputer, specifically one that utilizes "Commerical Off the Shelf" (COTS) hardware such as [[personal computer|personal computers]] and [[Computer network|Ethernet switches]] to link the machines together so that they act as one, and the Beowulf library of software which is used to help implement a [[distributed computing|distributed application]].  


At its most basic level Beowulf is a library of software used to help implement a [[distributed computing|distributed application]]. This software includes facilities such as providing a global [[PID|process id]], methods of [[remote execution]] of [[process|processes]], etc. <ref name="Donald Becker at NYLUG">{{cite web
This software includes facilities such as providing a global [[PID|process id]], methods of [[remote execution]] of [[process|processes]], etc. <ref name="Donald Becker at NYLUG">{{cite web|url=http://ssadler.phy.bnl.gov/adler/DB/DonaldBecker.html|title="From Word Processors to Super Computers: Donald Becker Speaks about Beowulf at NYLUG"|date=Retreived 11-April-2007}}</ref>  
|url=http://ssadler.phy.bnl.gov/adler/DB/DonaldBecker.html|title="From Word Processors to Super Computers: Donald Becker Speaks about Beowulf at NYLUG"|date=Retreived 11-April-2007}}</ref>  


The concept of clustering machines together in this way is known as [[distributed computing]].
==Beowulf Development==
In early 1993, NASA scientists [[Donald Becker]] and [[Thomas Sterling]] began sketching out the details of what would become a revolutionary way to build a cheap supercomputer: link low-cost desktops together with commodity, off the shelf (COTS) hardware and combine their performance.<ref name="The inside story of the Beowulf saga">{{cite web|url=http://www.gcn.com/print/24_8/35499-1.html|title="The inside story of the Beowulf saga"|date=Retreived 11-April-2007}}</ref>
By 1994, under the sponsorship of the "High Performance Computing & Communications
for Earth & Space Sciences" (HPCC/ESS)<ref name="HPCC/ESS">{{cite web|url=http://www.lcp.nrl.navy.mil/hpcc-ess/|title="High Performance Computing & Communications for Earth & Space Sciences homepage"|date=Retreived 11-April-2007}}</ref> project, the Beowulf Parallel Workstation project at NASA's Goddard Space Flight Center had begun.<ref name="Becker Bio">{{cite web|url=http://www.beowulf.org/community/bio.html|title="Donald Becker's Bio at Beowulf.org"|date=Retreived 11-April-2007}}</ref>
<ref name="beowulf.org history">{{cite web|url=http://www.beowulf.org/overview/history.html|title="Beowulf History from beowulf.org"|date=Retreived 11-April-2007}}</ref>
==Beowulf Implementation==
This type of cluster is composed of a 'master' (which coordinates the processing power of the cluster) and usually many 'nodes' (computers that actually perform the calculations). The 'master' typically is [[server]]-class, and has more horsepower (i.e. [[RAM|Memory]] and [[CPU]] power) than the individual nodes. The [[node|nodes]] in the cluster don't have to be identical, although to simplify deployment this is usually the case.
This type of cluster is composed of a 'master' (which coordinates the processing power of the cluster) and usually many 'nodes' (computers that actually perform the calculations). The 'master' typically is [[server]]-class, and has more horsepower (i.e. [[RAM|Memory]] and [[CPU]] power) than the individual nodes. The [[node|nodes]] in the cluster don't have to be identical, although to simplify deployment this is usually the case.


Line 21: Line 30:
| date= Retreived 08-April-2007
| date= Retreived 08-April-2007
}}</ref>
}}</ref>
==Beowulf Development==
In early 1993, NASA scientists [[Donald Becker]] and [[Thomas Sterling]] began sketching out the details of what would become a revolutionary way to build a cheap supercomputer: link low-cost desktops together with commodity, off the shelf (COTS) hardware and combine their performance.<ref name="The inside story of the Beowulf saga">{{cite web|url=http://www.gcn.com/print/24_8/35499-1.html|title="The inside story of the Beowulf saga"|date=Retreived 11-April-2007}}</ref>
By 1994, under the sponsorship of the "High Performance Computing & Communications
for Earth & Space Sciences" (HPCC/ESS)<ref name="HPCC/ESS">{{cite web|url=http://www.lcp.nrl.navy.mil/hpcc-ess/|title="High Performance Computing & Communications for Earth & Space Sciences homepage"|date=Retreived 11-April-2007}}</ref> project, the Beowulf Parallel Workstation project at NASA's Goddard Space Flight Center had begun.<ref name="Becker Bio">{{cite web|url=http://www.beowulf.org/community/bio.html|title="Donald Becker's Bio at Beowulf.org"|date=Retreived 11-April-2007}}</ref>
<ref name="beowulf.org history">{{cite web|url=http://www.beowulf.org/overview/history.html|title="Beowulf History from beowulf.org"|date=Retreived 11-April-2007}}</ref>


==Popularity in High-Performance Computing==
==Popularity in High-Performance Computing==

Revision as of 13:00, 11 April 2007

In a Beowulf cluster, the network cables (shown in blue) that connect the eight nodes to the master serve as a bus

A Beowulf cluster is a class of supercomputer, specifically one that utilizes "Commerical Off the Shelf" (COTS) hardware such as personal computers and Ethernet switches to link the machines together so that they act as one, and the Beowulf library of software which is used to help implement a distributed application.

This software includes facilities such as providing a global process id, methods of remote execution of processes, etc. [1]

The concept of clustering machines together in this way is known as distributed computing.

Beowulf Development

In early 1993, NASA scientists Donald Becker and Thomas Sterling began sketching out the details of what would become a revolutionary way to build a cheap supercomputer: link low-cost desktops together with commodity, off the shelf (COTS) hardware and combine their performance.[2]

By 1994, under the sponsorship of the "High Performance Computing & Communications for Earth & Space Sciences" (HPCC/ESS)[3] project, the Beowulf Parallel Workstation project at NASA's Goddard Space Flight Center had begun.[4] [5]

Beowulf Implementation

This type of cluster is composed of a 'master' (which coordinates the processing power of the cluster) and usually many 'nodes' (computers that actually perform the calculations). The 'master' typically is server-class, and has more horsepower (i.e. Memory and CPU power) than the individual nodes. The nodes in the cluster don't have to be identical, although to simplify deployment this is usually the case.

Usually the Beowulf 'nodes' are running Linux,[6] however this is not required, as both Mac OS X and FreeBSD clusters have been created.[7][8]

Popularity in High-Performance Computing

Today Beowulf systems are deployed worldwide as both as "cheap supercomputers" and as more traditional high-performance projects, chiefly in support of number crunching and scientific computing.

It should be noted that more than 50 percent of the machines on the Top 500 List of supercomputers [9] are clusters of this sort.[2]

External links

The Linux Beowulf HOWTO, from the Linux documentation project

References