IBM Unveils Fastest Microprocessor Ever 292
adeelarshad82 writes "IBM revealed details of its 5.2-GHz chip, the fastest microprocessor ever announced. Costing hundreds of thousands of dollars, IBM described the z196, which will power its Z-series of mainframes. The z196 contains 1.4 billion transistors on a chip measuring 512 square millimeters fabricated on 45-nm PD SOI technology. It contains a 64KB L1 instruction cache, a 128KB L1 data cache, a 1.5MB private L2 cache per core, plus a pair of co-processors used for cryptographic operations. IBM is set to ship the chip in September."
Required (Score:4, Funny)
But will it run ... a Beowolf cluster of ...
[Comment terminated : memelock detected]
Re: (Score:3, Insightful)
[Comment terminated : memelock detected]
If Slashdot ever gets this working I'll instantly subscribe.
Speed times Quantity? (Score:3, Interesting)
So what is this beast supposed to be, a 64 core machine?
Didn't we retire the Ghz wars 5 years ago? I know, AMD style "more done per cycle", but isn't a quad core 3.1 Ghz per chip with 20% logistic overhead faster?
Re:Speed times Quantity? (Score:5, Informative)
The thing is that if you have 2 (say) 1.6 GHz processors, they aren't as 'powerful' as one 3.2 GHz processor.
For one - there are overheads, certain stuff common between them, pipelines - stuff which I forgot (computer engineering related problems).
But the main thing is that not all programs are multi-threaded, and a program with a single thread can only run on one processor. So yeah, GHz are still useful. Maybe for large single-thread batch processing - which is the kind of thing a mainframe would do.
Re: (Score:2)
But the main thing is that not all programs are multi-threaded, and a program with a single thread can only run on one processor. So yeah, GHz are still useful. Maybe for large single-thread batch processing - which is the kind of thing a mainframe would do.
I'm betting the code used on these z196 systems is multi-threaded. Shit, if you're paying hundreds of thousands of dollars per CPU you can afford some top notch programmers. With two co-processors used for cryptographic operations per chip I'd say they were after a bigger prize than, say, hardcore gamers ;-)
BTW, TFA mentions L1 cache per core but doesn't mention how many cores this chip scales up to. Could it be just one?
Re: (Score:3, Interesting)
It later mentions using 128Mbyte just for level 1 cache, so that would be around 1024 cores.
Re: (Score:3, Informative)
WP has the answer: http://en.wikipedia.org/wiki/IBM_z196_(microprocessor) [wikipedia.org]
Four cores, 128 KByte L1 data cache, 64 KByte instruction cache.
Re: (Score:3, Insightful)
But the main thing is that not all programs are multi-threaded, and a program with a single thread can only run on one processor. So yeah, GHz are still useful. Maybe for large single-thread batch processing - which is the kind of thing a mainframe would do.
I'm betting the code used on these z196 systems is multi-threaded. Shit, if you're paying hundreds of thousands of dollars per CPU you can afford some top notch programmers.
Actually I think this mainframe is for getting the last little bit of performance out of thirty year old cobol code. And the original top notch programmers are long dead.
Re: (Score:2)
Actually I think this mainframe is for getting the last little bit of performance out of thirty year old cobol code. And the original top notch programmers are long dead.
Considering that life expectancy in the developed world is in the region of 80 years, there is a reasonable chance that programmers who were under 50 when they wrote code thirty years age are still alive.
They may have little recollection of what they did 30 years ago, but to say they are all "long dead" is somewhat of an exaggeration.
Re: (Score:2)
I've never met a programmer over 50. I must therefore conclude that they all perish mysteriously upon their 50th birthday. Something like the planet of grim reapers from Futurama is how I prefer to envision it.
No programmers over 50? (Score:3, Informative)
I guess I'm a counterexample then:
I'm 53.
I believe (hope?) most people who know me would say that I'm still a pretty good programmer.
Terje
Re: (Score:2)
More processors = Share the Legacy.
Re: (Score:3, Insightful)
Mainframes are engineered fundamentally around two things: Reliability and IOPS.
When it comes to basic tasks, it isn't often that a large server ends up CPU bound (especially database servers). Instead what usually becomes the bottleneck is I/O and RAM.
Reliability is where mainframes take the cake. Some use multiple CPUs to execute the same instructions to make sure the output is correct. Mainframes have virtually redundant everything. Because they have been doing VM since the dawn of computing, it may
Re: (Score:2)
Shit, if you're paying hundreds of thousands of dollars per CPU
You aren't. FTA, the complete systems will cost hundreds of thousands of dollars, to a few million. Not the individual CPUs.
Re:Speed times Quantity? (Score:5, Informative)
When configured to run Linux, each core costs approx $125K. When configured for z/OS, each core costs approx $250K. A complete system (not including any storage or software) can cost up to around $30M.
Re: (Score:2)
And all the hardware will be there no matter what package you choose, and a "upgrade" will involve a IBM representative coming over to move a jumper.
Re: (Score:3, Informative)
Actually, IBM can upgrade mainframes over the internet. It can also downgrade it, if the lessee so chooses. The extra chips are used for failover.
Re: (Score:2)
If you're paying hundreds of thousands of dollars for a multi-GHz CPU then it's probably because you're trying to make up for the product of crap programmers, not the other way round.
Re: (Score:2)
Sorry to ask but who does IBM see using this?
At the price point and data sets that need sorting? - cheaper clusters or more expensive faster unique chips depending on math?
Re: (Score:2)
People with legacy mainframe programs that they don't want to port (translation: that they don't dare touch).
Re:Speed times Quantity? (Score:5, Informative)
Banks, Credit card companies, hospitals, Insurance companies...
Cheap clusters are great but they are not always the best tool for the job.
Very large traditional datasets involving lots of high value transactions, with 5 9s uptime requirements do not tend to scale well to COTS clusters.
IBM mainframes have uptimes measured in years if not decades.
They have hot swapable everything including CPUs. so you can do ugrades with zero downtime.
Also you need to take a look at the costs involved. The costs to throw out a working software system that has been used for decades and then the cost to redesign it to work on a Cluster of X86 boes will be huge.
Not to mention the investment in making it fault tolerant and if it is used in certain markets the cost of the auditing the software.
Not to mention that ZSystems tend to be really secure. There are just not a lot of exploits on Zsystems.
When downtime can cost millions of dollars hardware costs are just no that big of a deal.
Now if you are starting from scratch then you may save money by going with a cluster but then you may not depending on just how good your programmers are.
Re: (Score:3, Informative)
It has been a while but really?
I have never seen a mainframe that didn't use Zulu time. Also in the shop I worked all software was quality verified. One machine was at the five year uptime mark when I left but it was a none commercial system.
Re: (Score:2)
They say it's an old CISC architecture. This is probably the sort of system that runs horribly outdated and un-updatable code, like the tax system.
Re:Speed times Quantity? (Score:5, Informative)
"They say it's an old CISC architecture. This is probably the sort of system that runs horribly outdated and un-updatable code, like the tax system."
You mean like Windows?
The X86 is also an old CISC architecture.
Actually the Power line is RISC anyway. When it is used in a ZMachine the old style 360/370/390 CISC ISA is translated to RISC and then executed.
Before you go ew that is what modern X86 chips do as well as ARM when using the Thumb Instruction set. The ZSystem ISA is so high end it is almost a high level language so the translation doesn't really effect performance much at all. Also that old CISC architecture is much better than the mess that we have on the X86.
I am not sure about how IBM does the translation. On the System 38 AS/400 System-I the translation was done during the IPL aka Initial Program Load. On the Zs it may be done as a JIT but I am not sure.
Honestly I love the idea and wish that Linux would adopt it. You could then have one binary that would work on any Linux system on an CPU.
The AS400 way kept a native binary copy along with the TIMI copy. When the program was run the first time it would translate the TIMI copy into the native segment. Yes the first time you ran the program it might take a bit to start but after that it would run at full speed and start fast. Of course you could add a binary segment when you first released the code for the ISA of your choice.
All in all those old Mainframes and Minis had a lot of brilliant tech we still don't have today on our PCs.
Re: (Score:3, Informative)
The X86 is also an old CISC architecture.
Actually x86 is a new CISC architecture. The System/360 architecture predates it by over two decades. x86 was about the last CISC ISA to be developed outside of a few tiny niches.
Actually the Power line is RISC anyway. When it is used in a ZMachine the old style 360/370/390 CISC ISA is translated to RISC and then executed
Umm, no. POWER is RISC (well, RISC purists would say that's stretching the point), but POWER and System/z are completely unrelated. The POWER6 and z10, and POWER7 and this chip, were designed by cooperating teams, so they share some execution units, but they are very different architectures. This is not a POWER CPU running a S
Re: (Score:2, Insightful)
The thing is that if you have 2 (say) 1.6 GHz processors, they aren't as 'powerful' as one 3.2 GHz processor.
For one - there are overheads, certain stuff common between them, pipelines - stuff which I forgot (computer engineering related problems).
But the main thing is that not all programs are multi-threaded, and a program with a single thread can only run on one processor. So yeah, GHz are still useful. Maybe for large single-thread batch processing - which is the kind of thing a mainframe would do.
OK, firstly the OP should have said that this is the microprocessor with the highest clock speed. Calling it the fastest CPU is extremely misleading. In most modern CPUs, clockspeed is NOT related to throughput. The Intel Sandy Bridge or Nehalem CPU for example may be running its 4 cores at a clockspeed of 3.2GHz but overall, each core in the CPU is easily 4-5 times faster than a 3.2GHz Pentium4 core.
Secondly, many of the bottlenecks that you allude to are no longer major bottlenecks. CPU interconnect bandw
Re:Speed times Quantity? (Score:5, Insightful)
"clockspeed is NOT related to throughput"
Of course it is. It is not, however, the only factor, and other factors may indeed (and commonly do) outweigh it.
"IBM may have created a very highly clocked CPU and given it tons of transistors, but I seriously doubt if it will compete with a modern day server CPU from Intel or even AMD."
I think you underestimate IBM's technical ability. They do have some idea of what they're doing.
"pure performance maybe, but definitely not price-performance or performance-per-watt"
That's like saying a Ferrari is a poor performance car because it can't compete against a Ford Focus on cost-per-max-speed or miles-per-gallon.
Re: (Score:2)
That's like saying a Ferrari is a poor performance car because it can't compete against a Ford Focus on cost-per-max-speed or miles-per-gallon.
I doubt that IBM mainframes suffer from the equivalent of engine fires. [jalopnik.com]
Re: (Score:3, Interesting)
"clockspeed is NOT related to throughput"
Of course it is. It is not, however, the only factor, and other factors may indeed (and commonly do) outweigh it.
You took my comment out of context. I was responding to the original post that focused purely on clockspeed as a magic mantra. What you say is only true if you are talking about clock speed increase in the same microarchitecture, ceteris paribus. Making a blanket claim that we have the fastest CPU because we have clocked it at 5GHZ means nothing. I could overclock a P4 to 5GHZ using exotic cooling and my laptop would still probably beat it in terms of performance.
I think you underestimate IBM's technical ability. They do have some idea of what they're doing.
Of course they do. I wasn't talking trash ab
Re: (Score:2)
Re: (Score:3, Funny)
Not saying I'd recommend it, but if that's the measure that you want to use then I'd say a cheap clunker could probably beat the both of them*.
*Take the rest of the cash your would have spent buying either one and spend that on blowjobs.
Re: (Score:3)
According to the Passmark benchmark, a 3.20 GHz scores 524 [cpubenchmark.net], compared to 10221 [cpubenchmark.net] for a 3.20 GHz Core i7 970 six-core CPU. That works out to 3.14 times faster per core than the Pentium 4. While short of 4-5, the GP is not as far off the mark as your ridicule would suggest.
I actually think YOU (and the cretin who modded you insightful) fail.
Re: (Score:2)
Yup... there are so many dependencies on application and OS code that hardware capability matters very little.
I recently tried to tune a workload on a pSeries system. We gave it half a processor and 2 virtuals (with the Power version of hyperthreading so it saw 4 processors). Performance was a dog. Load was only 60% of capacity though. We doubled the number of virtual processors but kept the overall entitlement. Load dropped to 40%. Added another couple virtuals and load dropped to 25%. No increase in thro
Re: (Score:2)
Yeah, I'm uninformed.
Re:Speed times Quantity? (Score:5, Insightful)
More or less. They hit two walls - fabricating chips that could run faster while retaining an acceptable yield, and dealing with the heat such chips produced.
The fastest general-sale chips were the P4s - the end of their line marked the end of the gigahertz wars, as Intel switched from ramping up the clock to ramping up the per-cycle efficiency with the Core 2 and their complete architecture overhaul. As a result a 2GHz Core 2 duo will outperform a 4GHz P4 dual-core under most conditions. Better pipeline organisation, larger caches better managed.
Clock rate is no longer the key variable in comparing processors, unless they are of the same microarchitecture.
Re:Speed times Quantity? (Score:4, Interesting)
Yeah, it's actually kind of funny how today's Intel desktop processors actually trace their lineage to the Pentium M, which was a mobile chip. When the Pentium 4 came around, the Pentium Pro (Pentium II, Pentium III) architecture was pretty much relegated to the mobile market while Pentium 4 represented their desktop line. As you said, they ran into heat (and power) issues with the Pentium 4s and basically had no more room for expansion there. They went back to the Pentium M, which was doing pretty nicely in the notebook space, and since it was low-power and efficient it became the basis for their future desktop CPUs--the Core line, in particular. They just stopped playing up the clock speed because that architecture's clock speeds were substantially lower than the Pentium 4, despite being able to do more work. I read once that a Pentium M could do about 40% more work than a Pentium 4 of the same clock, so in essence a 2GHz Pentium M was about as powerful as a 3.2 GHz P4.
Switching everything over to the low-power and parallel-friendly Pentium M line is probably one of the smartest things Intel ever did. They would've dug their own grave had they stuck with building on Pentium 4 to the bitter end.
Re: (Score:3, Informative)
Clock rate has *never* been the key variable in comparing processors. Even back in the heady days of 1 MHz 6502/6510 vs 4 MHz Z80 the comparison was useless - the 6510 does way more per cycle than the Z80 and ends up being comparable speed-wise.
Re: (Score:3, Informative)
The word you are looking for is "sleight".
Sleight of hand.
Re: (Score:2)
Re: (Score:2)
You actually can go faster without x86 hogging you down
Re: (Score:2)
IBM BlueGene/L - runs at 700 MHz .... 596 TFLOPS
Cray XT5 - runs at 2.6 Ghz ... 2331.00 TFLOPS
Both of these are slower in Hz than the PC I am using to type this ....
Re: (Score:3, Insightful)
It's quad core. 24 MB of L3 Cache, and 96 MB of L4 Cache.
source [theregister.co.uk]
Price: RTFA (Score:5, Informative)
Re: (Score:2, Informative)
Re:Price: RTFA (Score:4, Informative)
You can buy or lease the hardware. The software is licensed under contract.
It seems like the GP is talking about software charges, not hardware. Software can be either monthly fee based or usage based. If it is usage based you must send a usage report to IBM so they can bill you. That is specified in the contract. In either case, the number of and performance of the CPs is calculated into the cost.
Hardware is a different story. With hardware, the number of cores you purchase is not the same as the number you get. For instance, you can buy a 1 core machine, but what you get is 16 cores. Only 1 core is enabled in the firmware though. IBM has offerings (again under contract) where you can buy the right to temporarily enable additional processors instantaneously (like if you lost one of your datacenters and need to transfer the workload to another one). With these offerings, you also need to send usage info to IBM so they can bill you for the time that the additional cores have been enabled.
Comment removed (Score:5, Funny)
Re:Great news for Mac OS X users! (Score:5, Funny)
"Please note: The revolutionary new MagsafePro 3-Phase/480 power connector is not backwards compatible with the Magsafe connectors of prior, non-containerized Mac Portables."
Re: (Score:3, Informative)
Re: (Score:2)
Re: (Score:2)
Uhm...
We're talking about Z-series mainframes. These are absolute beasts, with all the cooling, memory and processing speed that would leave a desktop in the dust without any problems whatsoever.
However putting this sort of hardware in a desktop is extremely prohibitive for a ton of reasons, one of the most important being cooling. You'd need a room just for that...
Re:Great news for Mac OS X users! (Score:5, Informative)
Wrong chip family. This is the Z-series mainframe chip, using an instruction set that is backwards compatible with the System/360 stuff from back in 1960 (the architecture of the future, as the marketing material trying to persuade my university to upgrade their IBM 1620 put it). The PowerMacs were using PowerPC chips, which use the same instruction set as the POWER CPUs from IBM (they used to be similar, with a common subset, now they are identical).
The chip that this is replacing, the z10, was designed concurrently with the POWER6. They share a number of common features, including a lot of the same execution engines (both have the same hardware BCD units, for example, as well as more common arithmetic units), but they are very different in a number of other aspects, including the instruction set, cache design, and inter-processor interconnect, because they are designed for different workloads.
I've not read much about this chip yet, but I think it shares some design elements with the POWER7, in the same way that the z10 did with the POWER6.
In short, while some of the R&D money spent on this CPU made it into chips that could, potentially, run OS X, this chip itself could not without a major rewrite.
Re: (Score:2)
It's not a PowerPC chip anyways. It's zSystem architecture, which is actually the modern-day descendant of what was originally the System/360.
True, true. For now. (Score:2)
CISC to save RAM? (Score:2)
IBM defines the z196 as one of the few remaining CISC chips, which allows for bulky, large programs that can require much more memory to execute in than RISC chips, including the PowerPC and ARM embeddded processors, among others.
For CISC you need more bytes per instruction, because there are more instructions. With RISC your executable has more instructions but they each use less storage.
I am not sure I believe their implication that CISC is better for humungus commercial applications. Sounds like marketing speak to management to me.
CISC seems to work well (Score:2)
Essentially all desktop and laptop computers use CISC chips and they are fast and cheap. RISC is a neat theory, but these days it seems that as the processors get decoupled from their ISAs anyhow, for various reasons, that it doesn't matter much. You choose the ISA for reasons of binary compatibility or features or the like, and it'll work just fine with the chip.
Also it is not true that CISC needs more bytes per instruction, at least not all implementations. With x86 you find instructions are variable leng
Re: (Score:2)
Actually, CISC uses less memory in general, but has traditionally been slower. CISC CPUs came out when memory was extremely expensive relative to CPU speed. cheaper memory is what made RISC (with its larger footprint but faster speed) possible. Nowadays, it really doesn't matter much, CISC is probably better nowadays that memory bandwidth is the big bottleneck. However, our CISC designs are not exactly modern, if you were to do a modern CISC design you would probably end up with something more akin to ARM's
Re: (Score:2)
CISC and RISC are marketing terms that incorporate a lot of loosely connected design elements. Most CISC architectures use variable-length instruction encodings. On x86, for example, a number of common instructions are a single byte, while the longest ones are 15 bytes. A RISC architecture typically has fixed-length instructions, typically either 4 or 8 bytes (although ARM chips tend to also support Thumb and Thumb-2 instruction sets which use a 2-byte encoding).
This is why x86 chips need smaller inst
Re: (Score:2)
This is a mainframe, it is designed to run many virtual machines, natively
Memory in this context is cheap, fast and readily available - all the things it is not on most RISC systems
If the chips are designed properly (which with IBM they will be) then for the tasks they are designed for they can be faster ...for other tasks they may well be slower
This chip snickers at my 6502... (Score:4, Insightful)
Can't even imagine writing in assembly code for this monster. I miss dinking around with a nice 6502 system.
Re: (Score:2)
Can't even imagine writing in assembly code for this monster. I miss dinking around with a nice 6502 system.
Yeah the 6502 is nice and friendly. I taught myself how to hand assemble on the 6502 when I was 12 or 13.
Re: (Score:2)
I miss dinking around with a nice 6502 system.
Start playing with ARM then, its design was somewhat inspired by the 65xx series and there are plenty of affordable ARM-based systems available.
Re: (Score:2)
Or something "lower" among many popular microcontroller families. AVR is quite pleasant, for example.
You really don't anymore (Score:3, Interesting)
These days, compilers take care of almost everything. It has gotten complex to the extent that a programmer trying to do things all in assembly will probably do a worse job than a good compiler. Chips have many, many tools to solve their problems.
That isn't to say it is never done, in some programs there may be some hand optimized assembly for various super speed critical functions. However even then it is most likely written in a high level language, compiled to assembly (you can order most compilers to do
art (Score:2)
Re: (Score:2)
If you're going for code perfection every time. In the real world you have deadlines and have to maintain your code. Writing in assembly is going to make your code harder to port across platforms, should that happen say from PowerPC to x86.
Not saying that its never justified to use assembly. Within reason, of course.
Re: (Score:3, Informative)
An appropriate conference for the announcement (Score:2)
Announcing a 5.2 GHz, 1.4 billion-transistor processor at "Hot Chips 2010" just makes sense. Strangely, no power numbers were given...
Re: (Score:2)
From Wikipedia each multi chip module takes as much as 1800W (six processors)
No figures for 1 chip though
http://en.wikipedia.org/wiki/IBM_z196_(microprocessor) [wikipedia.org]
highest clock speed? not really (Score:2, Interesting)
Intel's netburst architecture (of pentium 4 fame) featured the 'Rapid Execution Engine', which consisted of two ALU's running double the clock speed, on 3.8 GHz Pentium 4's, that would be 7.6 GHz
Granted, that is not the entire cpu, but still..
About time! (Score:2)
*sigh* FINE, begin the switch back....
-Steve
Sent from my iPad 2
With CoProcessors.... (Score:2)
My 386DX has an external Maths Coprocessor, => can only do floating point functions :( ... your one does how 5.2 Ghz -> Sure my M series superseeds your G series..right?.... ....
However mine's now a bit faster overclocked it from 33Mhz to 52Mhz
right?
The S/360 architecture lives on! (Score:2)
It's crazy that an architecture developed in the '60s lives on in the System Z today. IBM bet the company on the S/360 product line. I think the investment has paid off-- and still does!
Wait....what? (Score:3, Insightful)
I'm assuming the cache referred to in the second paragraph is off-chip cache, otherwise it would sort of negate the first sentence.... Would be nice if the article would have actually said that though.
Re: (Score:3, Insightful)
Considering the ratio between the two sets of figures is ~96, it seems that the "four-node system" contains 96 cores with their own L1 and L2 caches, but shared L3 and L4 caches.
Re: (Score:2)
That is correct.
Re: (Score:2)
Indeed. "Off-chip L1 cache" seems like a self-contradiction.
64KB L1 (Score:2, Funny)
That ought to be enough instruction cache for anybody.
Costing hundreds of thousands of dollars... (Score:2)
The codename for this processor, was "Ming Mecca".
-Laxitive
So much for the 3.3GHz speed of light limit. (Score:2)
Re:So much for the 3.3GHz speed of light limit. (Score:5, Informative)
The comments were about the fact that at 3GHz light travels 10cm per clock speed, which limits how far you can have 2 items on a bus if you want them to communicate within 1 clock cycle. There is no "light speed barrier" or anything of the sort, however at these frequencies you design knowing that it will take measurable time for an electric signal to propagate. For example, for this particular system whose core is at 5.2GHz, if you try to send a signal to an external memory that is say 11-12cm away, then it will take about two clock cycles just for the signal to travel the distance.
Re: (Score:2)
A lot of nonsense was spoken in that thread, but the issue is real. The time taken for light to travel is not yet a problem, but the skew is. Most communication between parts of a chip is parallel. If the connections are not precisely parallel then signals arrive at slightly different times. The clock speed is limited to the amount of time that is the maximum where signals will arrive in the same time slice. A similar limit also affects fibre optics, due to total internal reflection causing paths taken
A Little more detail here (Score:2, Informative)
Is this chip a bargin? (Score:2)
How much would it cost for me to put together a system with the same computing power, using off-the-shelf products, like a Xeon chip, or something? How long would it take for me to save $1 million in electricity, or whatever?
And to think that back in 1980... (Score:2)
...my trusty VIC-20 was clocked at a mere 1.02 MHz.
Its truly amazing how far we've come in so short a time.
(Well, maybe not so short for you whippersnappers...)
Bad Golf? (Score:3, Funny)
Had a golf game ended differently, would we be seeing these in power macs?
Put that in your iPad (Score:3, Funny)
IBM also previously claimed the title of fastest microprocessor with the POWER6 chip, which ran at speeds of up to 4.6 to 4.7 GHz, and its own z10, a 2008 chip which ran at speeds of up to 4.4 GHz.
I seem to recall that one of the official reasons Apple gave for the switch from Power to Intel was that IBM couldn't/wouldn't deliver a fast enough processor.
Re:Yeah, I read about this (Score:5, Insightful)
Yes, but their article comments are much closer to Youtube than Slashdot.
Re: (Score:2, Redundant)
If I could mod you insightful, I would.
Re: (Score:2)
Re: (Score:2)
Re: (Score:2)
Can we still enable "Coffee Crisp"?
Re: (Score:3, Funny)
Wait, do I have it turned on or off right now?
Re: (Score:2)
Given that the Z architecture doesn't even have PCI, that would be a no.
Re: (Score:2)
Actually, I believe it does now. 3D driver support is even worse than Linux, though. :)
Re: (Score:2)
Re: (Score:2)
Re: (Score:2)
Re: (Score:3, Interesting)
iChip?
Re: (Score:2)
Re:I doubt it's the fastest ever... (Score:4, Informative)
ummmm.......
It's a quad-core chip. Each core has two integer, two load and store, one binary floating point, and one decimal floating point unit. Up to 24 CPUs can be placed in the frame. It can connect to another whole rack of POWER7 blades running AIX as an application accelerator platform.
The z196 is for the stuff a mainframe is good at: big batches and fast I/O. The application accelerator is for stuff the clusters of supermicro servers are good at. As a hybrid system connected across the GX bus, it should pump data in and out of applications out pretty well.