InfiniBand Drivers Released for Xserve G5 Clusters 134
A user writes, "A company called Small Tree just announced the release of InfiniBand drivers for the Mac, for more supercomputing speed. People have already been making supercomputer clusters for the Mac, including Virginia Tech's third-fastest supercomputer in the world, but InfiniBand is supposed to make the latency drop. A lot. Voltaire also makes some sort of Apple InfiniBand products, though it's not clear whether they make the drivers or hardware."
Proprietary Crap (Score:3, Informative)
Re:Proprietary Crap (Score:2)
> InfiniBand in a free/open source product without violating the
> licensing agreement of the spec, because of patent
> infringement.
Not even in the nvidia drivers kind of way, with proprietary kernel modules? Not the most optimal solution (probably nearing highly pessimal) but probably possible.
Re:Proprietary Crap (Score:5, Informative)
Re:Proprietary Crap (Score:2, Funny)
(Dress this up in a bunch stupid rhetoric, and you have the typical response around here.)
Re:Proprietary Crap (Score:3, Informative)
Re:Proprietary Crap (Score:4, Interesting)
Re:Proprietary Crap (Score:1, Interesting)
Re:Proprietary Crap (Score:1)
I was missing the point, I might read more carefully next time. Before I activate auto-fingers.
Storm
Re:Proprietary Crap (Score:2)
Don't waste your breath. The parent poster you respond to doesn't strike me as a Mechanical Engineer nor even deals with PDE's via FEM, let alone the vast aspect of accurately calculating Fluid Dynamic laws.
Re:Proprietary Crap (Score:3, Interesting)
In particular, Infiniband requires licensing under RAND terms, similar to that of IEEE 1394.
Re:Proprietary Crap (Score:3, Insightful)
so if you're there you're already pretty deep in "properiaty crap".
Re:Proprietary Crap (Score:5, Insightful)
#1 XServes runing (wait for it....) Mac OS X.
#2 Supercomputers
This is not your linux box you're using for a NAT server, or a Beowolf running SETI, so if you're building a super computer or just like drolling over them and thinking of using and expensive interconnect like InfiniBand, you're not looking to compare it to Beowolf over gigabit, and possibly not likely to care about if the drivers are binary only or not.
This article is in no way related to any LKML posting other than it's the same company. This is about OSX Infiniband drivers. RTFA sometime, and you might realize such things.
Welcome to the Apple section. If you're not interested in discussion of things related to Apple, please uncheck the appropriate box in your preferences, and we will all be happier. If you like to run Linux on Apple Hardware, please examine the OS discussed before trolling.
If you want to troll about Infinbands policies effecting Linux, then wait until the LWN article is public ("Alternatively, this item will become freely available on October 21, 2004"), and submit it to
Re:adjust your attitude, please (Score:1)
Calling me an Apple advocate because I see announcement of OS X drivers for a product as unrelated to a discussion about compatability with open source projects is assuming too much. In fact I'm about as cross platform as someone can get. I personally use Linux, OS X and Windows for both work and home. The count of servers at work, where I'm a sysadmin, is 12 Linux, 6 Windows,
Re:adjust your attitude, please (Score:2)
I didn't call you anything. Macintosh advocates in general regularly post all sorts of Apple trivia to Linux and X11 discussions. And Macintosh advocates in general make all sorts of claims about Linux and X11 in Apple forums on Slashdot. And that's OK: that's what Slashdot is about.
I don't inject Apple topics into Linux discussions, and I don't appreciate others injecting Linux into Apple related discussions.
Then you're on the wrong site. This is "News for Nerds", not
Re:adjust your attitude, please (Score:2, Insightful)
Re:Proprietary Crap (Score:2, Funny)
There was a dead project that I read about a few months ago that had 20microsecond latency over 100 ethernet. If anybody knows what I'm talking about, I would appreciate a refresher.
Re:Proprietary Crap (Score:2)
Performance differs by an **order of magnitude**
10GbE vs. Infiniband - maybe, but even so - Infiniband is cheaper and has lower latency.
Re:Proprietary Crap (Score:2, Insightful)
Re:Proprietary Crap (Score:5, Insightful)
[A 747] can do everything [the Joint Strike Fighter] can, although without the same [supersonic speed, air-to-air combat capability] of [the JSF]. [The JSF] just never caught on when it could, it was ahead of its time, but now [747s are] cheap, and soon [the Airbus A300] will strip it dry.
Smart. "Without the low latency of infiniband"? Idiot, what do you think it's for? We're not talking eDonkeying Halo 2 here... ultra-low latency is THE POINT.
Oh wait, you're just a stupid fucking troll. Why don't you go die?
Re:Proprietary Crap (Score:1, Offtopic)
Is that the one with Mario?
Proprietary, but definitely not crap (Score:5, Informative)
No offense, but you don't know what you're talking about. IB can sustain tranfer rates of 700 MB/s; the best I've ever seen from GigE was almost an order of magnitude lower, not to mention the two orders of magnitude drop in latency with IB. That might not mean much to you, but I guarantee you it's a big deal for folks with big parallel scientific codes.
Oh, and your pricing's wrong too. In the quantities you'd need it for a decent size cluster, IB gear is about the same cost as its direct competitors (Myrinet and Quadrics).
Re:Proprietary, but definitely not crap (Score:4, Insightful)
It would be nice to see IB actually come together, but it's an uphill battle. The Spec is a massive tangly mess. Vendor infighting and politics has nearly killed it dead two or three times now. The last thing it needs is for the specs to be priced like they're printed on gold leaf and patent battles to boot.
Meanwhile, I've seen lightweight reliable non-IP protocols over bog standard GigE hardware get 10 microsecond latency and as a result, 90MB/s ACTUAL transfer.
Given that, 10GigE could give IB a real run, especially if it's coupled with an onboard DMA engine (there's no reason it can't be). Consider that with the right protocol, GigE can get a little over 90% of theoretical, if 10GigE manages that, it'll beat IB.
There's a lot of good things about IB, but if the IBTA really wants it to catch on, they'd better start acting like they WANT people to buy it. Right now, IB's best chance looks to be the OpenIB project. However, if the IBTA decides to try locking it up tighter and tighter, OpenIB won't save them, the rest of the industry will do clever things with 10GigE and save itself a bunch of patent headaches.
Re:Proprietary, but definitely not crap (Score:2)
Re:Proprietary, but definitely not crap (Score:2)
The problem is, 90MB/s sustained isn't all that impressive. IB is dual-channel by nature, so a 4x HCA has 8Gb/10Gb in each direction. And 4x is just middle of the pack as far as IB is concerned.
No, but 900MB/s is better than 4x IB. I reported figures for 1GigE because that's what I had to benchmark. The bench shows GigE at a little over 90% of theoretical. If 10GigE can manage 70% it will be right in line with 4x IB.
12x IB sounds cool and all, but it's vapor at this point. 30GigE sounds great too, bu
Re:Proprietary, but definitely not crap (Score:2)
Also, 12x isn't really vapourware, it is available, it just costs a lot. There's also talk of setting a new mid-end spec out, 8x.
Re:Proprietary, but definitely not crap (Score:3, Insightful)
SUNET is having problems with 10GigE when they reach around 50-60%
If they're using IP, I'm not at all surprised. IP is designed to provide reasonable performance in a hetrogenous unreliable network. In a cluster environment, you would want the protocol to provide excellent performance in a homogenous and error-free environment and simple correctness (likely at a terrible penelty) in an unreliable environment.
The problem is that IP has to deal with fragmentation, out of order delivery and moderatly f
Re:Proprietary, but definitely not crap (Score:2)
You could actually look at it another way. It's been "killed dead" repeatedly and yet it keeps coming back. Does that mean that the underlying tech is *so good* that it cannot be resisted even with all these problems?
Absolutely! That has been the one and only reason it has stayed alive. However, all technologies have a shelf life. It can only keep doing that until something 'close enough' with fewer political problems comes along.
Clustered HPC is about the only opportunity for IB. Workstations can't
Re:Proprietary Crap (Score:3, Informative)
Where did you get this dollar amount and what exactly is it for, the HCA a switch, cable or all of them? HCA's are about 800-1000 dollars. Switches from Mellanox start at about $8,000 for a 480Gbit backplane 24 port switch. And up to $66,000 for a full 96 port modular switch. Cables though I will admit are costly at 100 bucks for a 4X 2m cable.
Also your statement that it is useless is complete FUD. It certainly will never gain the widespread
Re:Proprietary Crap (Score:3, Interesting)
4X infiniband is 10 Gb/s signal rate but actually 8 Gb/s data rate (8b/10b encoding). This is one of many facts that the IB marketing dept. keep forgetting (I keep telling them, but they won't listen for some reasons).
GigE and TCP are quite inefficient when compared to Infiniband
TCP over Infiniband is as inefficient, it has nothing to do with GigE. People us
Re:Proprietary Crap (Score:2)
Re:Proprietary Crap (Score:1)
The on-paper peak of 4x Infiniband is 1 GB/s, which is 8 Gb/s.
> On PCI-Express it's nearly double the bandwidth
Using 2 links or bidirectional ? The one-way maximum bandwidth of one 4x IB link is 1 GB/s. The only way to get over that is to use two links.
> The PCI-X bus is limiting the InfiniBand interconnect
The on-paper maximum bandwidth on PCI-X is 1 GB/s. It matches the one-way bandwidth of IB 4x. However, PCI-X is a bus, so bidirectional t
Re:Proprietary Crap (Score:5, Insightful)
One example of fake shared memory that I've seen is a cluster with an unusual design: Two IBM P5 570's with a total of 32 cores and 128GB RAM, linked together via IBM's NUMA interconnect. They also had a total of 36 Infiniband HCA's. The slave nodes are Xserve G5's with 2GB RAM and Infiniband, and a Xilinx FPGA-card that has its own memory banks. What the slave nodes do is essentially that they work straight against the RAM on the two 570's, with the local RAM only as a form of cache. The project runs as a multi-threaded app on the 570's, and are slaved out to the nodes. The project was originally meant to be used with some p690's.
Re:Proprietary Crap (Score:2)
Imagine (Score:4, Funny)
Shocking (Score:4, Insightful)
Re:Shocking (Score:2)
Not for PC girlymen (Score:1)
This is not stuff cobbling 2-4 PCs together. Its for people who want the ultimate Xserve solution. I have a 16-processor Xserve G5 with Gig-E and Myrinet. My next solution will be some 96 processors and all InfiniBand.
Infiniban into (Score:5, Informative)
This is a short into to infiband.
"InfiniBand breaks through the bandwidth and fanout limitations of the PCI bus by migrating from the traditional shared bus architecture into a switched fabric architecture."
"Each connection between nodes, switches, and routers is a point-to-point, serial connection. This basic difference brings about a number of benefits:
Because it is a serial connection, it only requires four as opposed to the wide parallel connection of the PCI bus.
The point-to-point nature of the connection provides the full capacity of the connection to the two endpoints because the link is dedicated to the two endpoints. This eliminates the contention for the bus as well as the resulting delays that emerge under heavy loading conditions in the shared bus architecture.
The InfiniBand channel is designed for connections between hosts and I/O devices within a Data Center. Due to the well defined, relatively short length of the connections, much higher bandwidth can be achieved than in cases where much longer lengths may be needed."
"The InfiniBand specification defines the raw bandwidth of the base 1x connection at 2.5Gb per second. It then specifies two additional bandwidths, referred to as 4x and 12x, as multipliers of the base link rate. At the time that I am writing this, there are already 1x and 4x adapters available in the market. So, the InfiniBand will be able to achieve must higher data transfer rates than is physically possible with the shared bus architecture without the fan-out limitations of the later."
That article is obsolete (Score:4, Insightful)
Re:That article is obsolete (Score:1)
Re:That article is obsolete (Score:4, Informative)
Re:That article is obsolete (Score:1, Informative)
speeeeed... (Score:2, Informative)
It's what I do for a living, actually... (Score:2)
Re:I'm curious (Score:3, Informative)
an one program them in Python or Perl, or only "Real Programmers(tm)" languages like Java and C++?
They can be programmed in any language. Fortran and C are by far the most common choices. It's common to see perl and shell scripts used as glue between standalone modular programs. It's about the only place where you'll still occasionally find hand assembly in the inner loops, though that's becoming less common as more compilers support MMX,SSE, etc instructions.
You won't find a lot of interpreted langua
Comparison with Myrinet (Score:5, Interesting)
And it has MacOSX Drivers:
http://www.myri.com/scs/macosx-gm2.html [myri.com]
Myrinet is used by 39% of the Top500 list published in November 2003
http://www.force10networks.com/applications/roe.a
Re:Comparison with Myrinet (Score:4, Informative)
1. Quadrics (EXPENSIVE! and closed standard) sub 4 microsec
2. InfiniBand (Realtively inexpensive, open standard) 4.5 microsec
3. Myrinet (Roughly the same price as IB, but closed standard) sub 10 microsec
4. GigE (cheap) 20+ microsec
All latency numbers are hardware not software latencies. Depending on how good your MPI stack is you can often triple those numbers.
There are so few companies making IB because there is only one chipset manufacturer right now. Mellanox. All the companies making IB products are startups and it will be a while before things get better.
Re:Comparison with Myrinet (Score:2, Informative)
Myrinet is not a closed standard. It's an ANSI-VITA standard (26-1998). The specs are available for free (http://www.myri.com/open-specs/ [myri.com]) and anybody can build and sell Myrinet switches, if they have the technology.
Furthermore, the latency is sub 4 microsec. Come to SuperComputing next month and you will see.
Re:Comparison with Myrinet (Score:5, Informative)
where did you get these numbers?
If you really want to compare the latency of actual interconnects you should use the official performance results achieved in real environments using the driver api:
(values from homepages)
1. SCI (dolphinIcs) : 1.4 us
2. Quadrics: 1.7 us
3. Infiniband 4.5 us
4. Myrinet 6.3 us
MPI latency and bandwidth highly depend on the mpi library. I suggest to compare the mpich results.
I rated these interconnects. But I'm sorry, I only have a german version.
http://stef.tvk.rwth-aachen.de/research/interconn
Re:Comparison with Myrinet (Score:2)
Re:Comparison with Myrinet (Score:4, Informative)
Of course, until IB over fiber media comes around, myrinet cabling is a hell of a lot easier to deal with, longer lengths, more bendable, and tighter bend radius.
Re:Comparison with Myrinet (Score:3, Informative)
The quality of 4x IB cable has gotten much better over the last two years. It will continue to improve as 10 GigE also uses the same style cable.
Re:Comparison with Myrinet (Score:1)
Not 3rd fastest, in fact not on list at all. (Score:4, Informative)
from http://www.top500.org/lists/2004/06/trends.php
* The 'SuperMac' at Virginia Tech, which made a very impressive debut 6 month ago is off the list. At least temporarily. VT is replacing hardware and the new hardware was not in place for this TOP500 list.
Re:Not 3rd fastest, in fact not on list at all. (Score:3, Informative)
(The link should be good until sometime this weekend, then it will be avaiable in re-runs)
Permanent link. (Score:4, Informative)
Re:Not 3rd fastest, in fact not on list at all. (Score:2)
BigMac already has I.B. (Score:5, Informative)
Note that V.T.'s cluster already uses InfiniBand, courtesy of Mellanox [mellanox.com].
It's mentioned at V.T.'s pages [vt.edu].
Re:But did it ever work? (Score:1, Offtopic)
That's true. (Score:2)
Mellanox made the HCAs (Score:2)
Well thank goodness.... (Score:3, Funny)
Third fastest what? (Score:1, Redundant)
Re:Third fastest what? (Score:4, Informative)
Re:Third fastest what? (Score:4, Informative)
Re:Third fastest what? (Score:1)
So are you saying upgraded Big Mac (currently 12.05 teraflops Rmax [appleturns.com], 20.2 Rpeak) and the apparently not-yet operational COLSA MACH5 [apple.com] (unknown Rmax, 25.1 Rpeak) will be 3rd and 4th on the Top 500 list? Aren't you forgetting Thunder [top500.org] (19.9 Rmax, 22.9 Rpeak) and ASCI Q [top500.org] (13.9 Rmax, 20.5 Rpeak) from the most recent list [top500.org] (June 2004)?
If we're assuming
Re:Third fastest what? (Score:1, Interesting)
You seem like the type that needs proof, so here' [top500.org]
the gigE card is more interesting (Score:3, Informative)
http://small-tree.com/mp_cards.htm
Gigabit has a latency of about 100 microseconds and realistic throughput of about 50MB/s. Infiniband has a latency of about 15 microseconds and a throughput of about 500MB/s.
I mostly sell small Apple workgroup clusters of 16 nodes, and these are almost always just a gigE backbone. There are certain classes of problems that can benefit from Infiniband at low node counts, but for the most common apps, like gene searching using BLAST, gigE is just fine.
Re:the gigE card is more interesting (Score:2)
Oh, and regarding Voltaire in the original poster's message, Voltaire does make Infiniband hardware, and they do support Mac OS X.
Re:the gigE card is more interesting (Score:2)
Re:the gigE card is more interesting (Score:2)
You could interconnect with optical fiber too, right? Although I don't really know if that'd be faster.
I'm expecting delivery of my 8-node Apple Cluster this week, which will almost exclusively do BLASTing, so I'm interested in picking your brain. And learning who your clients are: maybe they're in the market for on-site support? Reply to email above.
To keep this on topic, I'll plug the Apple Listserves that deal with this subject: Xgrid [apple.com], SciTech [apple.com], Cluster and HPC [apple.com].
Re:the gigE card is more interesting (Score:2)
Do you mean Fibre Channel? That's not a supported network transport, just a storage transport.
Re:the gigE card is more interesting (Score:1)
- Ethernet also works over Fiber
- Even at the same bitrate
- IP over SCSI or FC is not unheard of, and easily do-able.
Not 3rd fastest (Score:2, Informative)
http://www.netlib.org/benchmark/performance.pdf
http://appleturns.com/scene/?id=4980
"Calm down, Beavis; take a closer look at the third and fourth entries and you'll realize that they're the same exact cluster, before and after its owners added another 64 processors to it. In much the same way, System X is also listed in the seventh, ninth, and eleventh slots, with scores taken at various points along its journey to life as a complete 1,100-Xserve system. Factor out the doubles
Re:NASA and SGI could crash that party (Score:2)
semi-offtopic but deals with clusters and speed. (Score:1)
Re:semi-offtopic but deals with clusters and speed (Score:1, Interesting)
But, once the G5 goes dual-core, I would expect to see a dual dual-core G5 machine out there somewhere. Doe
Xsan (Score:2)
Re:Xsan (Score:2)
Re:Xsan (Score:2)
Re:Xsan (Score:1)
Phil
Obligitory "beleaguered company facts" (Score:2)
2) Macs are only for designers...so who cares?
3) Macs cost more than PCs...so who cares?
I'm surprised we haven't seen the usual, eight year old "facts" as to why this is a fruitless effort. Slowly but surely, Apple is making its way back into the limelight. After being the whipping boy for so long for a variety of reasons (no market share, higher outright cost, stability issues, etc), Apple is proving itself to be cheaper, more stable, and damn powerf
You might also want to look at Infinicon (Score:2)
As I understand it, the advantages of IB over gig-E are lower latency and scalability.
Third fastest my ass (Score:2)
Will you ever stop repeating that lie?
Second, it is under testing (not even in production).
(Third - not as relevant but still - why is a driver release still news? Topspin et al have been offering infinband drivers for Linux for a while; who wants
Re:Third fastest my ass (Score:1)
Re:Third fastest my ass (Score:2)
I will, can you please provide the URL?
> it was in third place
That's what I'm talking about - it *was* for the day or week they tested it, but it was probably crashing or something - in any case, they couldn't/wouldn't use it as it was, so they embarked on an upgrade (or "tuning" i.e. debugging) program.
A year (!) later the hardware has deprecated some 33% (3 year period, US$5.2m), they've _wasted_ US$1.716m and they're still not using it.
That's laughable. What
bandwidth (Score:2)
Re:So? (Score:2)
I bet your computer is using a MAC right now!