Linux Clusters Finally Break the TeraFLOP barrier 223
cworley submitted - several times - this well-linked submission about a slightly boring topic - fast computers. "Top500.org
has just released its latest
list of the world's fastest supercomputers (updated twice yearly). For
the first time, Linux Beowulf clusters
have joined the teraFLOP club, with six new clusters breaking the teraFLOP
barrier. Two Linux clusters now rank in the Top 10: Lawrence Livermore's "MCR" (built by Linux NetworX ) ranks #5 achieving 5.694 teraFLOP/s, and Forecast Systems Laboratory's "Jet" (built by HPTi) ranks #8 reaching
3.337 TeraFLOP/s. Other Linux clusters surpassing the teraFLOP/s barrier
include:
LSU's "SuperMike" at #17 (from Atipa
), the University at Buffalo
at #22 and Sandia National Lab at
#32 (both from Dell ), an Itanium cluster
for British Petroleum Houston at #42 (from HP
), and Argonne National Labs at
#46 (from Linux NetworX ) reached just
over the one teraFLOP/s mark with 361 processors. In the previous Top500 list compiled last June, the fastest Intel based Netfinity 1024 processor clusters from IBM were sub-teraFLOP/s and the University of Heidelberg's AMD based "HELICS" cluster (built by
Megware
) held the top tux rank at #35 with 825 GFLOP/s."
Have another coffee while it Boots (Score:4, Interesting)
Wow! (Score:4, Interesting)
2 Hewlett-Packard 7727.00 Los Alamos
The distance from the first to the second is pretty impressive. What on earth did NEC really do over there?
How many FLOPS (Score:2, Interesting)
Comment removed (Score:2, Interesting)
Beowulf clusters aren't so tricky... (Score:5, Interesting)
Step 1: Install the lam packages on all the nodes
Step 2: Create an account on all nodes, and use a passphrase-less ssh key to avoid prompting.
Step 3: Compile your code with mpicc (rather than gcc)
Step 4: Copy to all nodes.
Step 5: mpirun C
Admittedly it was only a 4 node cluster, but hey
Please, someone break it to me gently if this wasn't actually a Beowulf cluster
Re:FLOPs (Score:5, Interesting)
LinuxBIOS (Score:5, Interesting)
Re:Beowulf clusters aren't so tricky... (Score:3, Interesting)
The main comment that struck me was how easy it was to set up. The Engineer IT department is mostly Unix (they're all in retraining becuase they are dumping Sun Stations for Intel based systems running XP beleive it or not- becuase Intel chips are so much faster and machines running XP are much cheaper than Sun Sparcs (plus the software they want runs on XP)) so it was of course easy to set up for them.
Next they'll be setting up another LINUX cluster with maxed out dual or quad processor machines with more RAM. They're really excited.
Re:EARTH-SIMULATOR (Score:3, Interesting)
Re:Beowulf clusters aren't so tricky... (Score:5, Interesting)
I agree with you, in principle, it *is* easy to do but the problems increase with the number of nodes. IMHO, the main problems are:
-Administration effort per node has to be almost zero. Beyond a number of nodes you definitely need things like fully automatic instalation, automatic power control, automatic diagnostic tools, a batch system, etc. All these tools already exist but you need some know-how to put all them together.
-You need a large enough room with a cooling system that gives at least 100 W per node, 7kW in our case. Room temperature has to be about 20oC.
-Low cost PC hardware is not allways reliable enough for this application. If you have codes that run 24x7 for months in a large number of processors, the probability to have a hardware problem is very high.
We have found that our hardware suppliers do not carry out extensive tests on the systems they sell. This is because "normal" users run low quality OSs and they assume that it is normal that the computers just hang from time to time. Therefore, they do not allways detect failures in critical components such as RAM.
-Of course, your application has to be suitable for parallel computing, specially if your cluster uses a low cost 100Mb/s network. In this case, compared to a "conventional" parallel computer (eg Cray T3E), the processors are roughly equivalent but the network is about 10 times slower and is easily the bottleneck of the system.
Having said that, despite all the problems, I love Beowulfs. They have totally changed high performance computing, and they are definitely here to stay.
All this has been possible thanks to free software, so thanks Mr. Stallman/Torvalds and many others...
CNN and MS Bias (Score:2, Interesting)
As a side note, I find it rather funny that aside from technical issues, one can not legaly cluster Macrohard systems because EULAs gets in the way!