Virginia Tech to Build Top 5 Supercomputer? 460
hype7 writes "ThinkSecret is running a story which might explain exactly why the Dual 2GHz G5 machines have been delayed to the customers that ordered them minutes after the keynote was delivered. Apparently, Virginia Tech has plans to build a G5 cluster of 1100 units. If it manages to complete the cluster before the cut-off date, it will score a Top 5 rank in the Linpack Top 500 Supercomputer List. Both Apple and the University are playing mum on the issue, but there's talk of it all over the campus."
Do they have a need for it? (Score:0, Insightful)
Parent is stupid (Score:1, Insightful)
Latency isn't the most paramount issue, otherwise render farms and clusters wouldn't be as popular as they were today.
and lets not forget about projects like distributed.net and Seti@Home. Latency is not at all the concern for them.
Re:Macs ? (Score:5, Insightful)
Getting a bit ahead of ourselves (Score:3, Insightful)
Err... I think somebody's getting a bit ahead of themselves here. =) Building parallel computing systems is complicated, and it may end up being quite a bit harder to realize the predicted performance than thought (not an uncommon occurrence). I'll believe it when they have the actual Linpack numbers.
Re:What? (Score:1, Insightful)
iWorks.
Hoax (Score:2, Insightful)
Re:What about latency? (Score:4, Insightful)
Now you are one optimistic AC. The day is still young. I am giving 30:1 odds that there are going to be way better morons than Thinkit3 before this thread is archived.
Re:What? (Score:1, Insightful)
Re:What about latency? (Score:5, Insightful)
Universities (and big business) often work together and exchange resources. Virginia Tech gets a large amount of bargaining power by having control over a large amount of processing power. They can easily trade CPU time on their cluster for CPU time on a low-latency supercomputer.
Re:AltiVec (Score:1, Insightful)
One example is in RC5-72 crunching. My plain old consumer eMac can crunch 10.47 million keys per second using altivec. It's not even an optimal architecture for getting all the data through to the altivec unit (133mhz bus etc). The fastest x86 based box doesn't come -near- that. (see http://n0cgi.distributed.net/speed/ for stats)
That's one extreme case, and possible the most extreme one. Only knowing the final purpose for the VT cluster would reveal just how sensible picking PPC machines is.
Re:Macs ? (Score:3, Insightful)
Without more details its hard to tell
Re:is that so? (Score:2, Insightful)
Re:AltiVec (Score:5, Insightful)
Uh, no. 2 years ago, my roommate and I were both running the distributed.net client. I have a 500 Mhz Powerbook G4 (100Mhz bus). He had a 1.4GHz P4 with rambus RAM. I got 4Million keys/sec. He got 2MKeys/sec.
So clock for clock, my machine was nearly 4 times faster.
Re:MOD PARENT DOWN (Score:4, Insightful)
http://n0cgi.distributed.net/speed/query.php?cput
Power PC 7450/7455 G4 1000 MacOS X 10.2 2.9005 RC5-72 10,594,666.00
Re:1100 G5s still can't... (Score:1, Insightful)
Re:Poor choice on Apple's part (Score:5, Insightful)
(seriously - I can take out a loan
1% of G5 orders (Score:5, Insightful)
And 8 hours@12.4GFlops...damn you Virginia Tech, you owe me a third of a quadrilion floating point multiplies!
Re:Poor choice on Apple's part (Score:3, Insightful)
Re:What about latency? (Score:2, Insightful)
Re:AltiVec (Score:2, Insightful)
In your "calculations" i think you forgot to add the added manhours for doing this...
and thats not even thinking about the order / failure rate...
buy mem for 1100 machines is bound to have 1, 2 or more (properbly the ladder) mem blocks not working...
And im sure there is alot more reasons to this that you ( and i) missed...
Re:Another brilliant idea, inspired by Apple (Score:5, Insightful)
But... a cluster should be redundant enough to withstand that sort of minor inconvenience and go on functioning without the errant node while it gets fixed, reboots or whatever.
I'll admit that building something smart enough to say "Node 206, you have a memory error. Bad G5, no donut!" is beyond the scope of my understanding.
Re:Think Secret's Record (Score:2, Insightful)
I believe we're honest and up-front about the reliability of what we report. If something has not been confirmed by multiple reliable sources, we say so. If something is mere rumor, we also say so, but most of the time, we would rather not report it if we can't confirm it.
As for mistakes, after Macworld Expo, for example, we typically take a look at the announcements and compare them to our pre-expo reporting to see what, if anything, we got wrong. We don't hide anything.
Nick dePlume
Publisher and Editor in Chief
Think Secret
But, what will they *do* with it? (Score:2, Insightful)
Chip H.
Re:Talk about a ton of desktops in a server room (Score:3, Insightful)
I said there wasn't capacity on that generator. I didn't say anything about the existence of other generators - and there's a distinction between pulling copper to get power grid capacity to the cluster and having emergency power backup for same. That diesel is for emergency backup, not local power generation - we've got a nice coal-fired beast on campus for that.
In addition, you have to remember that the cluster most likely has different backup power requirements than our production systems. Our production systems (the central mail server, the large database machines, etc) really need to stay up - we're talking about things that if the power substation near the airport goes out, the whole campus is screwed till it comes back online, if we don't have backup power.
Look at how many services were out when we did the cut-over - everything was affected, from E-mail to the library catalog. That's what emergency power is for - so you don't have an event like that without the multi-week notice that the install had.
On the other hand, does a compute cluster really need that level of power backup, or can we significantly trim the budget by merely having a really good power conditioner to make sure there's nice clean non-ripply power, and enough battery backup to allow the cluster to do a suspend-to-disk and poweroff cleanly? Yes, if we take a power hit in that scenario, some researchers have to wait for their results - but the business of the university as a whole isn't disrupted.
Or possibly the plan is to have the conditioner and batteries for now, and have someplace to cable in a backup diesel generator in the future - remember that 1,100 G5s, a bunch of Infiniband switches are a chunk of change - and then you get the expense of the power/cooling work. Suddenly you start thinking about how to put the non-essential costs into the next fiscal year.
(The above is not intended as an actual statement of the actual design or plans, just an illustrative discussion of the fact that there are a lot of tradeoffs that need to be made when planning and designing large-scale installations of any sort).