Nvidia Discloses Details On Next-Gen Fermi GPU 175
EconolineCrush writes "The Tech Report has published the first details describing the architecture behind Nvidia's upcoming Fermi GPU. More than just a graphics processor, Fermi incorporates many enhancements targeted specifically at general-purpose computing, such as better support for double-precision math, improved internal scheduling and switching, and more robust tools for developers. Plus, you know, more cores. Some questions about the chip remain unanswered, but it's not expected to arrive until later this year or early next."
Another article here (Score:4, Informative)
http://www.pcper.com/article.php?aid=789 [pcper.com]
Just for a second glance.
More than just graphics (Score:5, Informative)
But this technology isn't straightforward. Someone asked why not replace your CPU with it? Well for one, GPUs didn't use to be able to do ANY floating or double-precision calculations. You couldn't even program calculations directly -- you had to figure out how to represent your problem as texel- and polygon-operations so that you could trick your GPU into doing non-GPU calculations for you. With each new card released, NVIDIA is making strides to accommodate those who want GPGPU, and for everyone I know those advances couldn't come fast enough.
Re:More than just graphics (Score:2, Informative)
Start looking at OpenCL as soon as possible if you want to learn gpgpu, cuda is nice but opencl is portable between vendors and hardware types :)
Re:Embedded x86? (Score:2, Informative)
What I'd like to see is nVidia embed a decent x86 CPU,
They did, its called Tegra. Except its not using the x86 hog, but way more efficent ARM architecture
Re:How did this get modded "insightful"? (Score:1, Informative)
GP was foolish to assume people would know what hir was talking about but:
GPUs are SIMD machines (Single Instruction Multiple Data), they process large quantities of numbers in parallel this is what makes them "fast" despite their low clock speed (compared to a CPU). They also have massive pipelines to decode as many instructions as possible simultaneously. All this makes them very powerful except for one major problem: branches. GPUs stall with major latency on branches.
If you can write general purpose software and operating system code that rarely uses Goto, Ternary, If, For, Do, While or Switch statements in C then you could pull it off, however, such a subset of C minus those constructs would not be Turing Complete so it'd be damn hard.
And then, even if you did succeed it would still be slower than the CPU code since not all workloads are compatible with SIMD, it only works on parallel streams; workloads that consist of multiple unrelated units (eg. pixels) that need to have the exact same operation performed on them whilst not depending on the results from the other pixels as part of those operations.
This is the major benefit of a CPU vs the GPU, the CPU can handle branch dense code with lots of interdependencies without too much stalling, the GPU cannot.
---
Of course, Intel's Larrabee GPU may change all this but that remains to be seen until it hits the market
Re:But does it... (Score:3, Informative)
Re:AWESOME (Score:5, Informative)
So no, I do mean the GT200. The GT200 processor supports double-precision, the G8x and G9x processors do not.
Re:But does it... (Score:3, Informative)