HP creates off-the-shelf supercomputer

Hewlett-Packard and a national laboratory in France add some network switches, Ethernet cabling and Linux software to some stripped-down PCs to create a new supercomputer.

CNET News staff
4 min read
By Matthew Broersma

How to build your own supercomputer: Take a few off-the-shelf, stripped-down PCs, add some network switches, a maze of Ethernet cabling and some homegrown Linux software, and you'll be well on your way.

Hewlett-Packard, together with a national laboratory in France, tried this recipe out and, to the great surprise of many scientists, it worked.

What they ended up with is the "I-Cluster," a Mandrake Linux-powered cluster of 225 PCs that has benchmarked its way into the list of the top 500 most powerful computers in the world.

At a technical session last summer, scientists from HP's labs in Grenoble, France, started talking to experts at the local INRIA Rhone-Alps (France's National Institute for Research in Computer Science) about the possibility of doing "something a little unusual." The idea was to build a supercomputer out of standard hardware components like those that might be found in the typical big business.

They started with 100 of Hewlett-Packard's e-PCs--simplified PCs with reduced expandability--and finally worked up to the present configuration of 225 nodes, which is near the cluster's physical limit.

HP and INRIA showed the system to journalists for the first time Wednesday.

The version of the e-PC used for I-Cluster is sealed, meaning no hardware tweaks can be made, and the experiment uses standard networking equipment. This means that, unlike other clustered supercomputing projects, a business could use the I-Cluster method to draw on idle computing power from around the company network to carry out computing-intensive tasks.

"These are really standard machines; we didn't even open the box," said Bruno Richard, program manager with HP Labs Grenoble.

Other clusters, like ASCI Red at Sandia National Laboratories in New Mexico, are made up of heavily modified parts.

The hard part
There were formidable obstacles in getting the cluster running as if it were one device, Richard said, such as distributing functions like storage and network caching to general-purpose devices and managing and programming the cluster.

"Our previous cluster was 12 machines," he said. "When you have 200, you have to rethink everything."

For example, making even simple software changes became a difficult task with so many machines to be altered. In the end, however, the technicians devised tools capable of reinstalling every machine in the cluster from scratch in about 12 minutes, according to Richard.

The researchers plan to release the tools they developed as open-source software for anyone who might want to build a supercomputer themselves. The whole project, minus network cabling, cost about $210,000.

The individual machines that made up the I-Cluster are now out of date, each running on 733MHz Pentium III processors with 256MB of RAM and a 15GB hard drive. HP introduced a faster version at the beginning of this month and will launch a Pentium 4 e-PC by the end of the year.

e-PC features like super-quiet cooling and low power consumption, originally designed for the corporate buyer, proved useful in the supercomputing environment too--the cluster runs surprisingly quietly and doesn't require anything more than standard air conditioning to keep it cool.

As ranked by standard benchmarks, I-Cluster is ranked 385th worldwide for supercomputing. Richard said the experiment showed that there is a linear relationship between the number of nodes and performance, meaning that it's relatively simple to add or remove computing power depending on the task.

About 60 research teams worldwide are working on the system, with half running typical supercomputing tasks and the other half exploring how I-Cluster works.

Serious computing
The project shows that standard computing power--like the unused processing power on an office network--can be harnessed for serious computing work. In the business world, CAD designers and chemists are among those who need intensive computing power, Richard said.

"You could gather the latent power from office PCs using this technique," he said. "We eventually want to scale it higher, to thousands of PCs."

Currently the hard limit for such a cluster is about 256 nodes, because of switching capacity, but that could be surpassed by linking several clusters that are physically near each other.

A more daunting task might be taking the model to a consumer environment, which, Richard pointed out, is full of often dormant processors like those in printers and DVD players.

HP imagines "clouds" of devices, or "virtual entities," which could discover and use the resources around a user. Richard said that supercomputing power could come in handy for certain tasks, like converting large video files from one format to another, that currently take a good amount of patience.

Other scientists predict that the practical difficulties of such a home network will prove difficult to solve. Brigitte Plateau, head of INRIA's Apache parallel computing project, said consumer need for such power probably wouldn't make it worth the effort that such a system would require.

"It is more likely that you would see an external service," she said.

HP's Richard said the use of Linux--version 7 of Mandrake's distribution, in this case--was important because low-level changes could be made easily to the software, and then the alterations could be shared freely with other scientists, something that would have required a special agreement with Microsoft if Windows had been used.

Plateau, whose Apache project encompasses I-Cluster, said the lab is also working with Microsoft to port parallel computing applications to Windows.

"We had to face heterogeneity by spreading it over Linux and Windows too," she said. "It's not scientific, but technically it's good experience."

Staff writer Matthew Broersma reported from London.