NVIDIA Predicts 570x GPU Performance Boost 295
Gianna Borgnine writes "NVIDIA is predicting that GPU performance is going to increase a whopping 570-fold in the next six years. According to TG Daily, NVIDIA CEO Jen-Hsun Huang made the prediction at this year's Hot Chips symposium. Huang claimed that while the performance of GPU silicon is heading for a monumental increase in the next six years — making it 570 times faster than the products available today — CPU technology will find itself lagging behind, increasing to a mere 3 times current performance levels. 'Huang also discussed a number of "real-world" GPU applications, including energy exploration, interactive ray tracing and CGI simulations.'"
Goody! (Score:5, Funny)
Nope! (Score:5, Funny)
Then we can use our GPUs as our CPUs!
No No. GPU's only become CPU's when they are 570.34567 times faster. You will note that he precisely said only 570 times faster. That is he did not say an even 600 or 1000 or 500, but precisely 570, so we can assume he knew it was not 570.34567.
Re:Goody! (Score:4, Funny)
Yo dawg, I herd you like CPUs, so I put a CPU in your GPU so you can still get work done in vi while emacs is loading!
Predictions of the future (Score:2)
What troubles me is that this sort of cell GPU is not more widely used in everyday applications. We who program for a living are feeling like we have been engaging in 'self stimulation' for years and wish there were some new target platform/market that we could so some interesting work in.
Re:Predictions of the future (Score:4, Insightful)
Both seem highly unlikely.
Re:Predictions of the future (Score:5, Insightful)
Well, it comes down to simple math. For the performance to get to 570-fold more than what it is now, in the same style package, either:
Both seem highly unlikely.
You don't feel it could be a combination of both? Kind of like they did with multi-core CPUs? Make a single unit more powerful, then use more units ... wow!
There is more than one way to skin a cat.
Re: (Score:2)
Or 3d-erize the chip?
Re: (Score:2)
Perhaps have them running at lower clock speeds. Slightly slower clock speeds means much cooler chips afaik.
Re: (Score:2)
From reading slashdot, I know there are several technologies which have been "a couple of years away" for a while now which could (if people bothered with the expense) turn the most common problem in computer heat dissipation into "how do I prevent it from getting too cold and forming condensation on everything?"
Re: (Score:2)
One wonders what night-miraculous improvement in process, packaging, logic design, etc. will improve GPUs by hundreds of times, while somehow being virtually useless for CPUs...
Re: (Score:2, Funny)
Re: (Score:3, Informative)
But only to the degree that your problem maps to that level of parallelization. There are many problems that do not perform well on the GPU.
Re:Predictions of the future (Score:5, Interesting)
Or... not.
Currently CPUs and GPUs are stamped together. Basically, they take a bunch of pre-made blocks of transistors(millions of blocks, billions of transistors in a GPU), and etch those into the silicon, and out comes a working GPU.
It's easy - relatively speaking - and doesn't require a huge amount of redesign between generations. When you get a certain combination working, you improve (shrink) your nanometre process and add more blocks.
However, compiler technology has advanced a lot recently, and with the vast amounts of processing power now available, it should be simpler getting more complex blocks fully utilized. A vastly more complex block, with interconnects to many other blocks, could perform better at a swath of different tasks. This is evident when comparing the performance hit from Anti-Aliasing. Previously even 2xAA had a huge performance hit, but nVidia altered their designs, and now Multisampling AA is basically free.
I recall seeing an article about a new kind of shadowing that was going to be used in DX11 games. The card used for the review got almost 200fps at high settings - with AA enabled that dropped to about 60fps, and with the new shadowing enabled, it dropped to about 20fps. It appears the hardware needs a redesign to be more optimized for whatever algorithm it uses!
Two other factors you're forgetting...
1) 3D CPU/GPU designs are coming slowly, where the transistors aren't just on a 2D plane... that would allow vastly denser CPUs and GPUs. If a processor had minimal leakage, and low power consumption, 500x more transistors wouldn't be a stretch.
2) Performance claims are merely claims. Intel claims a quad-core gives 4x more performance, but in many cases it's slower than a faster dual-core.
570x faster for every game? Doubtful. 570x faster at the most advanced rendering techniques being designed today, with AA and other memory-bandwidth hammering features ramped to the max? Might be accurate. A high end GPU from 6 years ago probably won't get 1fps on a modern game, so this estimate might even be low.
A claim of 250x the framerate in Crysis, with everything ramped to the absolute maximum, might be even accurate.
But general performance claims are almost never true.
Re:Predictions of the future (Score:5, Insightful)
1. The GPU has to become 570-fold more efficient
2. The GPU has to become ~570-fold smaller so they can fit 570 of the things onto a card
Both seem highly unlikely.
If graphics card development in the last 10 years is anything to go by, nVidia's plan is that the GPU will become 570 times larger, draw 570 times more power and the fan will spin 570 times faster
Re:Predictions of the future (Score:5, Funny)
If graphics card development in the last 10 years is anything to go by, nVidia's plan is that the GPU will become 570 times larger, draw 570 times more power and the fan will spin 570 times faster
At that point, it would effectively become a helicopter, no?
Re: (Score:3, Informative)
Well, it comes down to simple math. For the performance to get to 570-fold more than what it is now, in the same style package, either:
Both seem highly unlikely.
It's not a linear relationship.
Re: (Score:3, Funny)
Either
Neither seem highly unlikely.
Re:Predictions of the future (Score:5, Informative)
The GeForce 9 series was a rebrand/die shrink of GeForce 8, but the GTX 200 series has some major improvements under the hood:
* Vastly smarter memory controller including better batching of reads, and the ability to map host memory into the GPU memory space
* Double the number of registers
* Hardware double precision support (not as fast as single, but way faster than emulating it)
These sorts of things probably don't matter to people playing games, but they are huge wins for people doing GPU computing. The GTX 200 series has also seen a minor die shrink during the generation, so I don't know if the next generation will be more of a die shrink or actually include improved performance. (Hopefully the latter to keep up with Larrabee.)
Re:Predictions of the future (Score:4, Insightful)
Re: (Score:2)
I don't doubt the prediction at all, I just have concerns about the vat of liquid nitrogen I'm going to have to immerse my computer in to keep that thing from overheating, and the power substation I'm going to need to build in my backyard to power it.
But GPUs today are somewhat more than 570x more powerful than they were several years ago and we haven't had to submerge them in a vat of liquid nitrogen yet, so what makes you think that's going to be the case in the next 570x power increase? (whenever that happens ...)
Re: (Score:2)
Re: (Score:2)
Re: (Score:2, Insightful)
Jen-Hsun Huang is full of shit (Score:3, Insightful)
He constantly runs his mouth without any real thought to what he's saying. It's just attention whoring.
Re: (Score:2, Flamebait)
Mod parent up.
Jen-Hsun Huang is a certified clown who just a short while back was running around saying things like 'we will open a can of whoop-ass on Intel'.
What a dumbass ...
Re: (Score:2)
It's even more bullshity than normal since he's also evidently predicting the end of Moore's law. CPUs only improving by 3x in 6 years?!
6 years/1.5 years = 4 Cycles of Moore's law.
2^4 = 16x performance increase.
So I guess Moore's law in the next year is going to go from a doubling every 18 months to a doubling every 4-5 years? When did that happen?
Re: (Score:3, Interesting)
Its easy to get a 570x increase with parallel cores. You will just have a GPU that is 570 times bigger, costs 570 times more and consumes 570 times more energy. As far as any kind of real break through though, I'm not seeing it from the information at hand.
There is something worthy of note in all this though, which is that the new way of doing business is through massive parallelism. We've all known this was coming for a long time, but its officially here.
Re:Predictions of the future (Score:5, Informative)
The prediction is complete nonsense. It assumes that CPU processors only get 20% faster per year (compounded). That would only be true if they did not add more cores to the CPU. And finally GPUs are hitting the same thermal/power leakage wall that CPUs hit several years ago - they will at best get faster in lock step with CPUs.
A GPU is not a general purpose processor, as is a CPU. It is only good at performing a large number of repetitive single precision (32 bit) floating point calculations without branching. Double precision (64 bit) calculations - double in C speak - is 4 times slower than single precision on a GPU. And the second you have an "if" in GPU code, everything grinds to a halt. Conditions effectively break the GPU SIMD (single instruction multiple data) model and bring the pipeline to a halt.
Re: (Score:3, Insightful)
"It assumes that CPU processors only get 20% faster per year (compounded). That would only be true if they did not add more cores to the CPU."
"It is only good at performing a large number of repetitive single precision (32 bit) floating point calculations without branching."
If we wanted a 64-bit GPU it would be easy enough to make. GPUs used to do weird mixes of integer and floating point math until the manufacturers made an effort to guarantee 32-bit precision throughout. That leaves the branching part o
Re: (Score:2)
Moore's law is ending. The fab issues at the scale can be costly. Moore's law is about cost, not speed.
In other news... (Score:5, Informative)
In other news, ATI is selling their 4870 series cards for $130 on newegg, which are twice as fast as an Nvidia 9800GTS which is the same price (at least on Left 4 Dead, Call of Duty, and any other game that matters). ATI is blowing Nvidia out of the water in terms of performance per dollar and will continue to do so through at least the middle of next year. See here:
http://www.tomshardware.com/charts/gaming-graphics-cards-charts-2009-high-quality/benchmarks,62.html [tomshardware.com]
Yeah, I'd be making outrageous statements too if I were Nvidia.
Re: (Score:2, Informative)
Here's the L4D comparo, sorry for the wrong link:
http://www.tomshardware.com/charts/gaming-graphics-cards-charts-2009-high-quality/Left4Dead,1455.html [tomshardware.com]
The 9800GT and 8800GT are both in the 40-60fps while the 4870 (single processor) is in the 106fps range. It's a pretty staggering difference.
Re: (Score:2)
Yeah and the 280 is in the 120fps range. Whats your point?
Re: (Score:2)
Re:In other news... (Score:4, Interesting)
>> provides features you can only appreciate on a 120hz display
well thats a new one. There's not even slight technical merit to that statement but its certainly demonstrates the amusing creativity of ATi fanbois.
>> The 9800GT and 8800GT are the same price and the ATI card blows it out of the water
I have no argument that you should go with Ati if you're windows only and looking at cheaper-end cards.
Its totally irrelevant to me though as I go for best overall performance, decent drivers, and only consider cards that have drivers that work well with Linux. ATI suck on all counts in my areas of interest.
Re: (Score:3, Informative)
I enjoy the following features of my GTX280 (used for calcs not games):
CUDA (I compile C code, throw in a couple of lines of stuff for the GPU and it runs on the GPU, easy)
Hardware optimizes my memory accesses and at times branchy code so the GPU is doing as much work as possible (makes it easy to get good results on the GPU)
Re: (Score:3, Insightful)
Re: (Score:3, Informative)
While his example is wrong (Nvidia's competitor to the HD4870 is the GTX 260 c216), AMD do have better value for money on their side. The HD4870 is evenly matched but a good bit cheaper.
The situation is similar in the CPU domain. The Phenom IIs are slightly slower per-clock than the Core 2s they compete with, but are considerably cheaper.
Re:In other news... (Score:4, Informative)
Depending on vendor it is now possible to get a 275 less than a 4890 and a 260 for only slightly more than a 4870; at lower prices its very competitive too. My point is that both NV and ATI are on pretty level ground again and the ONLY reason I now choose NV over ATI is because of the superior NV drivers (both Linux and Windows side)...oh and the fact that ATI pulled a fast one on me with their AVIVO performance claims. Shame on you ATI!
Re:In other news... (Score:4, Interesting)
In other news, ATI is selling their 4870 series cards for $130 on newegg, which are twice as fast as an Nvidia 9800GTS which is the same price (at least on Left 4 Dead, Call of Duty, and any other game that matters). ATI is blowing Nvidia out of the water in terms of performance per dollar and will continue to do so through at least the middle of next year. See here:
http://www.tomshardware.com/charts/gaming-graphics-cards-charts-2009-high-quality/benchmarks,62.html [tomshardware.com]
Yeah, I'd be making outrageous statements too if I were Nvidia.
Even when it comes to GPGPU (General Purpose computing on the GPU), ATI's hardware is much better than NVIDIA's. However, the programming interfaces for ATI suck big times, whereas NVIDIA's CUDA is much more comfortable to code for, and it has an extensive range of documentation and examples that provide developers with all they need to improve their NVIDIA GPGPU programming. It also has much more aggressive marketing.
As a sad result, NVIDIA is often the platform of choice for GPU usage for HPC, despite it having inferior hardware. And I doubt OpenCL is going to fix this, since it basically standardizes the low-level API, keeping NVIDIA with its superior high-level API.
Re: (Score:2)
In addition to VDPAU enabled mplayer, I can actually FIND CUDA enabled apps. There's CUDA enabled md5 crackers, cuda enabled BOINC, Matlab has a CUDA plugin. I'm considering buying CUDA compatible card so I can install it at work just to play with it in Matlab.
Re: (Score:3, Insightful)
Re: (Score:3, Insightful)
I agree. I recently bought a laptop with an ATI card and the biggest reason why I did that is because I heard they went open source. I was disappointed by the fact that their latest Catalyst driver doesn't work well on Ubuntu 9.04. The one recommended by Ubuntu works but it's VERY slow when restoring a window in Compiz. All in all it feels like a downgrade compared to my Intel integrated graphics card. Sigh. :(
Re: (Score:2)
Agreed. My primary use for the nvidia gpu is watching HD. Let's do some math.
1080 * sqrt570 = 25784
I like. Considering even the most basic of today's gpus, the ion and tegra, for example, are capable of 1080p, Mr Nvidia is predicting that my handheld 6 years hence will be able to smoothly decode mkvs and output them real-time to my new UltraMegaFullHD(TM) 25784p tv? Bring on the future!
But how? (Score:3, Insightful)
I read the article, but I don't see any explanation of how exactly that performance increase will come about. Nor is there any explanation of why GPUs will see the increase but CPUs will not. Anyone have a better article on the matter?
Re:But how? (Score:4, Insightful)
It'll come about because BUY NVIDIA GPUS THEY ARE THE FUTURE, CPU SUX
Re:But how? (Score:5, Funny)
Are you crazy? Why would I buy a GPU now if they're going to be 570 times faster in just a few years? I'd just be wasting my money!
Re: (Score:2)
I expect the real intention is more like BUY NVIDIA STOCKS THEY ARE THE FUTURE, CPU STOCKS SUCK.
Who needs profits from actual sales when you can manipulate the market?
Good to know! (Score:5, Insightful)
Thanks for the heads up, Nvidia! I'll be sure to hold off for 6 years on buying anything with a GPU.
Re: (Score:2)
That was my immediate thought as well.
We're about to drop $250K on a GPU cluster and if performance increases to that amount in 6 years, why on earth would we buy now?
Dammit, there's just no win when you fork out for clusters (of any kind).
Should spend 50K now, stick the rest into stocks and buy 50K every year. Of course, the dudes up the tree don't like that kind of thinking.
So... (Score:3, Funny)
I have to wait six years to play Crysis?
Re: (Score:2)
Now Arma 2...
Re: (Score:3)
Thanks a lot (Score:2)
My netbook tried to load that image and broke down in tears.
Re: (Score:2)
Re: (Score:3, Insightful)
All that resolution, and it still looks rendered. Instead of merely pushing more pixels, it would be nice if they did more to them, so it doesn't look so artificial.
OK (Score:2)
But they'd better hurry up with "Mr. Fusion" which will be needed to power that thing, and finally buy the license for that demon of Mr. Maxwell's to cool it.
Re: (Score:2)
Re: (Score:2)
why so serious? (Score:2)
And one more? (Score:2)
> 'Huang also discussed a number of "real-world" GPU applications, including energy exploration, interactive ray tracing and CGI simulations.'"
Add to that 'MD5 collisions etc"
GPU coding really is going to separate the men from the boys. I sense a return to the old days, where people had to think about coding, and where brilliant discoveries were made.
( like this: http://en.wikipedia.org/wiki/HAKMEM [wikipedia.org] )
Darn, pity I'm too old now. I'll have a play though...
The math (Score:2)
6 years = 72 months
Moore's Law states a doubling in transistors (but we'll call it performance) at every 18 month interval, so:
72/18 = 4 Moore cycles
2^4 = 32
So in six years, Gordon Moore says we should have 32x the performance we have now.
But it's indeed interesting... Silicon was a much easier-to-predict medium in the 20th Century. And yet here we have these two mature, opposing approaches to silicon-based computing, represented by the CPU and the GPU, with some predicting unprecedented growth for one and
Re: (Score:2)
Or what he is actually saying is that they (nVidia) will have more than 9 generations (~9.15) within 6 years... 1.5 generations/year... which I believe is fairly doable and actually slightly slower than the 6 month release cycle we have been accustomed to since 1998.
In other words "business as usual"
Re:The math (Score:5, Insightful)
So in six years, Gordon Moore says we should have 32x the performance we have now.
No - 32x the transistors.
You fail to predict how using those transistors in a more optimized way(more suitable to modern rendering algorithms) will affect performance.
Just think about it - a plain old FPU and SSE4 might use the same number of transistors, but when the code needs to do a lot of fancy stuff at once, one is definitely faster.
(inaccurate example, but you get the idea)
Re: (Score:2)
No - 32x the transistors. at the same cost, in the same space.
GPUs need more RAM for us (Score:5, Insightful)
I do high-performance lattice QCD calculations as a grad student. At the moment I'm running code on 2048 Opteron cores, which is about typical for us -- I think the big jobs use 4096 sometimes. We soak up a *lot* of CPU time on some large machines -- hundreds of millions of core-hours -- so making this stuff run faster is something People Care About.
This sort of problem is very well suited to being put on GPU's, since the simulations are done on a four-dimensional lattice (say 40x40x40x96 -- for technical reasons the time direction is elongated) and since "do this to the whole lattice" is something that can be parallelized easily. The trouble is that the GPU's don't have enough RAM to fit everything into memory (which is understandable, they're huge) and communications between multiple GPU's are slow (since we have to go GPU -> PCI Express -> Infiniband).
If Nvidia were to make GPU's with extra RAM (could you stuff 16GB on a card?) or a way to connect them to each other by some faster method, they'd make a lot of scientists happy.
Re: (Score:2)
Perhaps you know my brother? He has been doing a lot of this stuff.
http://www.google.com/search?client=firefox-a&rls=org.mozilla%3Aen-US%3Aofficial&channel=s&hl=en&source=hp&q=vandesande+qcd&btnG=Google+Search
Re: (Score:2)
Re: (Score:2)
Re: (Score:2)
Could you grab some motherboards with multiple expansion slots and load them up with dual-gpu boards?
Re:GPUs need more RAM for us (Score:5, Interesting)
You can -- that's what people are trying now. The issue is that in order for the GPU's to communicate, they've got to go over the PCI Express bus to the motherboard, and then via whatever interconnect you use from one motherboard to another.
I don't know all the details, but the people who have studied this say that PCI Express (or, more specifically, the PCI Express to Infiniband connection) is a serious bottleneck.
Re: (Score:2)
If Nvidia were to make GPU's with extra RAM (could you stuff 16GB on a card?) or a way to connect them to each other by some faster method, they'd make a lot of scientists happy.
Do you really need to ask them to do this for you? I'd think if you are a grad student you might be able to get together with some Electrical Engineering students and rig up something and turn a profit! The only thing you really need to know is how much memory the GPU can address, if you can get a hold of the source for the drivers, etc..
A video card isn't much more than a GPU with memory soldered on to it...
Re: (Score:2)
That product was actually specifically mentioned in the plenary talk at the 2009 Lattice Gauge Theory conference as the most likely contender for doing QCD on GPU's. It's still got the problem I mentioned, though -- not enough RAM to store everything, and not enough bandwidth to talk to the other units that are storing it.
Re:GPUs need more RAM for us (Score:4, Informative)
You mean like the Tesla [nvidia.com]?
No, that won't do. The NVIDIA architecture (which is shared between Tesla and graphic cards) is 32-bit, meaning that it can only flat-address 4GB of RAM tops. The more sophisticated Tesla solutions are essentially built from clusters of Tesla cards, each with its own 4GB of RAM tops. Separate memory spaces means expensive memory transfers to share data between the cards, which is not an issue if you can get good domain decomposition, but is a BIG issue if you cannot.
The revolution for HPC on GPUs would be a 64-bit GPU architecture.
Proper support for doubles and possibly even long doubles would be a plus, for applications that need it.
Re: (Score:2)
That too.
It turns out single precision is enough for lattice QCD. The step that requires the most CPU time doesn't need to generate an exact result; it only needs to get close. If the result is too far off then you wind up wasting time, but the result will still be valid.
(This is the Metropolis procedure, if you're familiar with it: the accept/reject step takes care of any computational errors that occur)
The big question is. (Score:4, Funny)
Brilliant sales pitch (Score:4, Insightful)
"Did I mention that our next model is going to be SO amazing that you'll think that our current product is crap? The new model will make EVERYTHING obsolete and the entire world will need to upgrade to it when it comes out. People won't even be able to give away any older products. Sooooo... how many of this year's model will you be buying today?
"Hello? Are you still there?
"Hello?"
% VS Times (Score:5, Interesting)
For the rest of us of course 570% increase is 5.7X faster.
So, CPUs increasing 3X in the next 6 years and GPUs increasing 5.7X I can maybe believe.
Re: (Score:2, Funny)
So a 100% increase would be 1.0X faster?
Re:% VS Times (Score:5, Informative)
It seems the rest of us don't understand what a "percent increase" means, either.
(hint: 570% increase == 6.7X)
Re: (Score:3, Informative)
Where is my math wrong?
This isn't about math (well maybe a little) as much as it is about wording. Basically, the difference between "as fast/increasing/the speed" and "faster/increase". The first is a multiplicative action, while the second is an additive. So if you say, 100% as fast, you are basically saying a * 100% = a. While if you are saying 100% faster you are saying a + a*100% = 2*a. Now looking at the posts. You started by this assertion.
seeing 570% increase and going
OK. A 570% Increase. That would be a+a*570% = 6.7*a.
570% increase is 5.7X faster.
increase/faster. Good.
GPUs increasing 5.7X
Ouch. Her
about 2.88x better each year for 6 years (Score:2)
Bitchin' fast! 3D (Score:5, Funny)
http://www.russdraper.com/images/fullsize/bitchin_fast_3d.jpg [russdraper.com]
285 Trillion floating point operations per second? (Score:2)
Comparison (Score:2)
Re:haha yeah right (Score:4, Interesting)
Intel said 4 nm for 2022, that's in 13 years. What precisely allows you to doubt that claims, except maybe the fact that deadlines are often missed? Let me rephrase that, what allows you to think that it'll be reached much later than anything else?
Also, queue a dozen+ posts explaining to the armchair pundits how 560x is possible.
Re: (Score:3, Informative)
The IEEE figures that semiconductor tech will be at the 11nm level around 2022. Intel and Nvidia both claim that they'll be significantly further along the path than the IEEE's roadmap. Maybe they're right, and I hope they are, but there are some very significant problems that appear as the process shrinks to that level.
Re: (Score:2)
570x is not that far (Score:3, Funny)
Re: (Score:2)
Re: (Score:3, Insightful)
I'm not shorting Intel's capabilities, but the IEEE has some solid people in it, too -- many of whom work at Intel -- and they're very capable of recognizing the potential problems with process shrinks. The issues that come about at the sizes they're discussing involve quantum tunneling effects that would (as I understand it) interfere in accurate computing. There is also doubt that transistors can be made to work at all at sizes below 16nm because the mechanisms that might deal with quantum tunneling may
Re: (Score:3, Insightful)
Stupid I know, but I would have had more confidence in a 500x increase, just because there's less significant digits and a wider error margin.
Re: (Score:2)
Intel said 4 nm for 2022, that's in 13 years. What precisely allows you to doubt that claims, except maybe the fact that deadlines are often missed? Let me rephrase that, what allows you to think that it'll be reached much later than anything else?
I'm dunno. Most CEOs don't make claims unless their business plan includes said claims else they look like a fool at the next shareholder meeting. That doesn't stop them from making claims that don't come through.
Remember Steve Jobs saying they would break the 3.0
Re: (Score:2)
Anyways... Intel seriously uses Moore's Law as their road map so its a self predicting prophecy.
No, they don't. It's descriptive, not something that ties your hands or, conversely, guarantees anything.
Re: (Score:2)
Re:Arbitrary number? (Score:5, Funny)
Re: (Score:2)
I'm wondering how you came up with 565 and 593x instead of say 564.8 or 593.82745109200174822x
Re: (Score:2, Funny)
Re: (Score:2)
It will at least have the same useful functionality over that Server 2008 has - non-admin users can schedule tasks, and Powershell.
Re: (Score:2)