Linux Cluster attains 125.2 GFLOPS 66
akey writes "CPlant98, a Linux cluster composed of 400
Digital Personal Workstation 500as, achieved 125.2 GFLOPS, which would place it at #53 on the Top 500 list. And this was only a 350-node run... "
I'm hearing rumors of 1000+ Linux Clusters. I'm itchin' for
it to come out of the closet so we can see some real
benchmarks.
Stone SouperComputer (Score:1)
It's linked to from the Beowulf site somewhere.
Re:Slim chassis (Score:1)
Probably, there are so many different cases out there. Many of the 'slim' cases I have seen only have 120-175 watt power supplies, which may be adequate for most uses, but may not be the best for sustained heavy duty use. Typical mid tower cases usually have 200-250 watt supplies.
As for floppy drives - simple - don't put one in the box.
The problem there is that it may increase the time necessary to service a machine if it dies.
For bulk serving you don't need it.
Very true under normal circumstances, but my concern is when the proverbial excrement hits the rotary oscillator. I want to be able to fix any problem ASAP. Having to find and install a floppy drive in a machine to bring it back up is time I may not want to take.
Nor a CD ROM drive.
That is normally something that you can usually do without, provided one is available on a network reachable machine.
If the box is at a colocation, you're going to get to it via ssh, not standing in front of it.
Provided you aren't dealing with a crashed hard drive or some other issue that can't be solved remotely.
Another complaint I have about a lot of 'slim' cases I've seen is many of them have limited quantities of front-panel accessable drive bays. While that isn't a big deal for most things, one useful thing when you are dealing with a large number of servers is the inexpensive IDE 'lock and load' trays, which make swapping in and out hard drives much faster and easier. It can make large scale upgrades or dealing with crashed drives a lot faster since you can do the work on another machine and then only take down the server box to do the actual swap.
Re:Maybe you have an answer to my question (Score:1)
Re:Slim chassis (Score:1)
Re:Slim chassis (Score:1)
Of course, on this machine the compute nodes aren't equipped with hard drives... :) Seriously, in a huge cluster like this, if a node fails, they will take it out, and may not even bother trying to fix it, I imagine.
Re:Maybe you have an answer to my question (Score:1)
A cluster runs a separate operating system on each node. This generally (again, this is hearsay) makes it much harder to maintain a cluster than a supercomputer (meaing one with one operating system). We purchased a small 8 node IBM sp2 computer six months ago, and still haven't figured out how to make it act like a single computer. :( Oh well.
Re:Maybe you have an answer to my question (Score:1)
Re:Maybe you have an answer to my question (Score:1)
NERSC [nersc.gov], for example, has recently purchased an IBM SP system which has two processors (or was is 4?) per node, with plans to upgrade to 16 processors per node.
The problem with SMP and clusters is that the message passing software has to be smart emnough to take advantage of the shared memory situation, and needs to and this can also complicate things when you try to optimize your code.
yes! (Score:1)
How many Windows NT machines rank in the top 53 of the worlds fastest machines?
Slim chassis (Score:1)
Re:Slim chassis (Score:1)
Two things (Score:1)
Fermilab has plans to build a 2000-node cluster in the near future but is putting off purchasing all the nodes until the last second to maximize their value.
re: Rackmounts
They're more expensive, and typically the machine rooms at large Beowulf installations have enough space for whatever they choose to use. It's not like Los Alamos has to pay for space at the colo when they add a new pile of Alphas.
Re:yes! (Score:1)
2600 Node Cluster (Score:1)
Umm.. really fast Quake
Re:Maybe you have an answer to my question (Score:1)
Christopher A. Bohn
Re:Slim chassis (Score:1)
Christopher A. Bohn
Maybe you have an answer to my question (Score:1)
how much can this type of computer scale up.
At least 2000 it seems (if somebody try to do it then it must thoerically scale to that extent), but do we have a theorical limit or something like that???
And are these computers mono-processors or SMP?
If Linux was going to have great enhancement in SMP for 4+ CPU's then would it be worth to create a cluster of SMP boxes given the price difference between SMP and non=SMP boxes actually (I suppose if you do a 2000 SMP cluster then you must have special price).
Re:yes! (Score:1)
By that logic this cluster is no faster than my 386 with 2mb of RAM.
Re:User accounts (Score:1)
So it wouldn't be "distributed.net", but "joint research into highly parallelised, highly distributed encryption validation", and "SETI@Home" is actually "joint research into vastly parallelised radio inteferrometry, using test data from Aricebo".
Re:Slim chassis (Score:1)
Once vendors start selling 1U machines with one disk, one processor, and one slot, this kind of thing will be more accessible to those outside the USA. I know VA has some 2 or 3U machines, that's the right direction to be going.
-Rich
User accounts (Score:1)
1000? 2000? 2600? (Score:1)
Just how big of a room does it take to house 2600 nodes?
Re:Super HUGE cluster problems (Score:1)
I know that for what I do (pseudopotential plane wave calculations) one 100bT switch would be too slow to connect an 8 node cluster. As in, you'd be better off with all the memory in one computer, and forget parallelization.
Also, I think that typically you don't want more than two hops between nodes. Of course, it all depends what you're doing. If you're doing monte carlo stuff, you could probably get by with 9600 baud modems if it were cheaper than ethernet.
Re:How does price relate to "true" supercomputers? (Score:1)
machines. If you used dual processor systems, the cost/performance would be even better.
Re:yes! (Score:1)
The Linpack benchmark is one that's been around a long time, and is pretty much an agreed-upon benchmark in the computing industry.
In addition, the numbers are submitted by the USERS of the machines, not the vendors, and this makes a LOT of difference in the trustworthiness of the benchmark.
Re:Maybe you have an answer to my question (Score:1)
Re:Slim chassis - Compaq DS10 variant (Score:1)
USA isn't the only tech producer! (Score:1)
I wonder if this will make weenies go for more treaties. Ugh.
Re:Slim chassis (Score:1)
http://www.tesys.com/servers/rackmount.shtml
neal
Re:yes! (Score:1)
Re:1000? 2000? 2600? (Score:1)
Re:Slim chassis (Score:1)
(based on the DS10, I believe).
40 of these will fit in one rack.
Is that Compaq(t) enough for you? (sorry for the
very poor pun).
- Jim Gettys
Re:FIRST POST!!! (Score:1)
A brilliant epitaph for a Slashdotter would be:
HERE LIES JOE BLOW
LAST COMMENT!
--
Get your fresh, hot kernels right here [kernel.org]!
Re:Slim chassis (Score:1)
Personally I hate most 'slim' style cases for a few reasons:
Personally, I like the 'mid tower' type of case. In today's world with huge capacity hard drives, the big full tower cases are often overkill, but the mini-towers are too cramped to work in.
Super HUGE cluster problems (Score:1)
As far as the question about how large you can go with them, if you use an int to determine which machine you are addressing, that puts a theoretical limit of more than 60,000 nodes.
Cluster Scaling (Score:2)
Short answer: it depends.
Long answer: it depends on the applications and the usage patterns.
(I'm assuming we're talking about practical limits here, not theoretical ones -- the theoretical limit is probably the address space of a cluster's message-passing interface (i.e. 4 billion nodes).)
Some applications -- the so-called "embarassingly parellel" ones -- will scale with nearly no deviation from linear to any number of nodes, because they do loosely-coupled problems. (Which means the result of one part of the parellel computation does not depend on a result from some other parellel computation. The mandlebrot set is a good example of this.)
In general, the more tightly-coupled the problem is, the harder it is to scale, as the amount of data that has to be exchanged pushes the limit of the interconnects. A 32-node cluster constructed on a hub will be faster for loosely-coupled programs than a 24-node cluster on a switch, which could beat the 32-node cluster on a tightly-coupled problem because of communications overhead in the 32-node cluster.
Usage patterns also determine the maximum useful size. If you're at a large lab like Sandia, you can reasonably expect a large number of jobs to be running concurrently, which essentially parellelizes the cluster -- running 6 tightly-coupled programs, each on their own hypercube interconnect, will complete faster than running the six in series, each with the whole cluster.
-_Quinn
Re:Slim chassis (Score:1)
When we build our approximately 100 node cluster (hmm, where can I get the money for the other 500 nodes????) We're considering using rack-mount boxes just because they would be easier to handle and simpler to install.
I'ld tell you how many nodes we can stuff into a standard size 19" rack, but we're still building it!
How does price relate to "true" supercomputers? (Score:1)
Re:What is a GFLOP? (Score:1)
Re:Maybe you have an answer to my question (Score:1)
The reason is simple: The access to the resources are better with mono-processor systems, there is no need of competition among the procesors in the same machine.
Beowulf in a box (Score:1)
Forget about slim chassis; how about no chassis? Take a look at Beowulf on StrongARM boards for $2000 [dnaco.net]. These folks are looking at building 6 StrongARM processors with RAM and the necessary "glue" onto a single PCI card. Since easily obtainable PCs have 3 PCI slots in them, you should be able to set up an 18 node beowulf cluster inside one box (the PC itself acts as the controller). Can you usefully cluster a bunch of these (a cluster of clusters)? I don't know, but it's interesting to think about.
Doug Loss
Re:1000? 2000? 2600? (Score:1)
Unfortunately, I don't know of an ethernet solution that could scale beyond several hundred systems and still provide uniform bandwidth and latency among all of the nodes. Beyond that point, you will need to come up with a scheme to partition to the network or just accept some performance penalty for crossing subdivisions boundaries of the cluster.
Re:What is a GFLOP? (Score:1)
Re:Beowulf in a box (Score:1)
Re:Super HUGE cluster problems (Score:1)
if you use ip on a closed network, surely you get the basic math of 254*254*254*254 give or take a few addr's ?? [ thats approx 4162314256 nodes ].
am I wrong?
I run a 25 node mpi/pvm cluster of 486-66 DX2's over a NuSwitch fast switching ethernet hub, at just 10mb/s full duplex, and I have a load of fun with pvm-povray rendering my nut off!
I'm now working on a multi node crawler to feed my search engine ( www.websearch.com.au ).
++dez;
Re:What is a GFLOP? (Score:1)
Re:What is a GFLOP? (Score:1)
Re:Maybe you have an answer to my question (Score:1)
Too bad. this would have been cool if it was possible to add the performances of Clusters with the performances of SMP. But that's pretty rare to be able to have the better of two worlds
thanks
Re:How does price relate to "true" supercomputers? (Score:1)
Would such a cluster be close to such super computers? Would the State Department start stepping in?
Food for thought.
Re:Slim chassis (Score:1)
As for floppy drives - simple - don't put one in the box. For bulk serving you don't need it. Nor a CD ROM drive. If the box is at a colocation, you're going to get to it via ssh, not standing in front of it.
Re:yes! (Score:1)