AMD Unveils Vega GPU Architecture With 512 Terabytes of Memory Address Space (hothardware.com) 125
MojoKid writes: AMD lifted the veil on its next generation GPU architecture, codenamed Vega, this morning. One of the underlying forces behind Vega's design is that conventional GPU architectures have not been scaling well for diverse data types. Gaming and graphics workloads have shown steady progress, but today's GPUs are used for much more than just graphics. In addition, the compute capability of GPUs may have been increasing at a good pace, but memory capacity has not kept up. Vega aims to improve both compute performance and addressable memory capacity, however, through some new technologies not available on any previous-gen architecture. First, is that Vega has the most scalable GPU memory architecture built to date with 512TB of address space. It also has a new geometry pipeline tuned for more performance and better efficiency with over 2X peak throughput per clock, a new Compute Unit design, and a revamped pixel engine. The pixel engine features a new draw stream binning rasterizer (DSBR), which reportedly improves performance and saves power. All told, Vega should offer significant improvements in terms of performance and efficiency when products based on the architecture begin shipping in a few months.
Re: (Score:1)
512 TB should be enough for any *CLUSTER*... (Score:2)
Fixed the title for you.
And I think that's genuinely the point of this:
it would be possible to run a whole cluster of compute nodes with VEGA GPUs,
and have all the data within a single unified address space across the whole cluster.
Just throw in a few IOMMU to handle access rights, and a fabric like Infiniband, or some PCI-E based one.
For bonus point have the storage it self being memory mapped non-volatile RAM (X-Point, etc.)
(But then you DO run out of address space - clusters tend to have data in the peta
512TB of address space means nothing (Score:1, Insightful)
Most high end GPU cards available have 8Gb, a large number of budget versions settle for 4Gb, and only a few offer 16Gb. Marketing this as a stand out point is iffy.
Re: (Score:1)
Oh, I know. That is still only 64Tb...
Re: (Score:2)
Anyway, it doesn't mean the cards themselves would need to have such huge amount of memory, it means they are now able to handle a memory address space that vastly exceeds their physical memory. Graphics cards don't need to have everything in local memory.
I've replied this below a good number of times below, but for completeness sake: this is not new. Not even for AMD, which i'm unable to find the exact number of addressable bits for each GPU family but they all support unified virtual memory with 64 bits CPUs since the days of the HD7700. Hell, Linux has support for this feature since 3.20.
nVidia? CUDA 5 has you covered as well.
My point is: current GPU offerings can already address way, way more memory than they usually physically carry. No idea why someone
Re: (Score:2)
512TB, sorry.
Re: (Score:3)
Most high end GPU cards available have 8Gb, a large number of budget versions settle for 4Gb, and only a few offer 16Gb. Marketing this as a stand out point is iffy.
What you will find is that most cards have only a fraction of their RAM as addressable, so a 16GB card either 4 or 8 gigs addressable. The increase to 512GB is a godsend to AI researchers and other fields with large datasets.
Re: (Score:1)
512Gb GPUs? Doubt you'll see that anytime soon. The very largest memory on any commercial GPU is AMD's own FirePro W9100, at 32gb. It is more of a cost issue than a limitation on addresable space; 64Gb is right around the corner though.
Re: (Score:1)
yea thats 32 gig's in banks of 4 bankswitching
even if you do not use it all just the wasted time and overhead saved is quite a bit in serious applications (and maybe even a few fps in gamez!)
Re: (Score:3, Informative)
Re: 512TB of address space means nothing (Score:3, Informative)
There is already technology available to feed this monster. Things like the EMC DSSD can have 1/2 PB of NVMe flash connected via a PCIe bridge, and presented as a single shared memory mapped space to an entire rack if servers. I assume that is the use case for these cards, mostly in the supercomputing space.
Re: (Score:2)
You really need to re-read my parent post.
Re: (Score:1)
"The increase to 512GB is a godsend to AI researchers and other fields with large datasets."
References?
While there's no doubt that there is SOME application that could use that amount of physical addressability, it would seem extraordinarily unlikely that a single GPU would be sufficient for such an application and, even so, it's absurd to refer to such a niche as a "godsend". Meaningless hyperbole, most likely without any supporting insight.
Re: (Score:3)
What you will find is that most cards have only a fraction of their RAM as addressable, so a 16GB card either 4 or 8 gigs addressable. The increase to 512GB is a godsend to AI researchers and other fields with large datasets.
Nope.
1: The GPU addresses the whole damn pool.
2: We're talking about 512 TB, not GB.
3: They're not planning to release a card with 512 TB of RAM, but they are releasing professional cards with lots of RAM (8 GB, 16 GB, or more) AND onboard connections for flash storage (SSDs). Vega will likely continue and extend this. By having a huge address space, you simply have the ability to keep the entire dataset in your cache on the card. The memory controller then decides what needs to live in the fast HBM2
Re: (Score:1)
You just spend two hours loading everything into the GPU's memory. Then you start managing it, updating what parts of it change, etc.
That PCIe bandwidth you mentioned? It's pretty scanty when you're shuttling 512 TB of data through it.
Re: (Score:3)
You don't understand.
The architecture can address that much, but the actual product will only address what's available.
There will be on-package HBM2 and the ability to connect to on-board (but off-package) storage in the form of fast flash.
512 TB of addressable space is just future proofing to allow for seamless work with a dataset regardless of whether it's on the 16 GB of ball-smackingly fast HBM2, on the SSD on your RadeonPro card, or in your system memory (or potential even abstracted out to disk storag
Re: (Score:2)
Re:512TB of address space means nothing (Score:5, Informative)
Lisandro you COULD RTFA, you know? It's even an effing meme around here.
The HBCC gives the GPU access to 512TB (half a petabyte) of virtual address space and gives the GPU fine-grained control, for adaptable and programmable data movement. Often, more memory is allocated for a particular workload than is necessary; the HBCC will allow the GPU to better manage disparities like this for more efficient use of memory. The huge address space will also allow the GPU to better handle datasets that exceed the size of the GPU’s local cache. AMD showed a dataset being rendered in real-time on Vega using its ProRender technology, consisting of hundreds of gigabytes of data. Each frame with this dataset takes hours to render on a CPU, but Vega handled it in real-time.
Re: (Score:1)
Wow, marketing lingo is eerily efficient on you.
Re: (Score:2)
Re: (Score:2)
AMD has supported this for years [wikipedia.org].
Re: (Score:2)
Re: (Score:3)
Pretty much all of it. I'm having a hard time finding out the number of addressing bits supported by, say, an Arctic Lake (4xx) GPU, but considering that AMD's GCN offered unified memory on the entire 64 bit space since 2011 [anandtech.com] and nVidia offers 49 bits of unified address space since CUDA 5 it surprises me that someone tried to make a selling point out of this feature.
They went as far as comparing a CPU render against their new GPU. WTF.
The only reason i can imagine someone would try to push this feature is th
Messing up the meme. (Score:1)
Re: (Score:2)
Re:trash posting (Score:1)
Re: (Score:2)
Re: (Score:2)
Re:512TB of address space means nothing (Score:4, Insightful)
But this is not new at all. IIRC Nvidia's CUDA 5 already gives you 49 bits of unified address space. Don't really know the addressing limitations on previous AMD architectures, but I doubt it was substantially lower.
Realistically, large address spaces when you can only practically fill 0.05-0.1% means little for performance. I don't want to attack AMD with this, who usually manufacture really good GPU hardware, but this sounds like a marketing gimmick and nothing more. I particularly enjoyed the "hours to real-time" comparison... against a CPU.
Re: (Score:2)
Re: (Score:2)
Oh, it is nice, don't get me wrong :) I'm just saying that promoting this as a dealbreaker is insane.
Re: (Score:2)
Re: (Score:1)
Re: (Score:2)
2^64 is 1.8x10^19. 1TiB is 2^40. So any 64 bit addressing scheme should at least cover 2^63 or 1,048,576TiB, assuming 1 address bit is sacrificed for the firmware or whatever else is needed.
So 512TiB is nothing, given what a flat 63 bit address space is capable of achieving. Also, supporting it on the address ain't difficult, given that there have been both data-address multiplexed lines as well as address-multiplexed lines.
Re: (Score:2)
Don't be a moron. You already got virtual memory mapping over huge address spaces on previous-gen GPU products, both from AMD and nVidia. Vega looks like a nice architecture but all these hyperbolic performance claims based solely on having 512TB of addressable space are utter bullshit. I'm actually surprised that most people here can't tell the difference.
Ah, and i'm pretty sure that "future-proofing" a GPU architecture that will be obsolete 5 years from now was certainly not a consideration for AMD engine
Re: (Score:1, Informative)
Framerate on the outside doesn't matter to the lifeforms inside the simulation.
Re: (Score:1)
And we are completely safe from said lifeforms, since nothing they do will ever emerge.
Person A: "Is it bricked, or just really really busy?"
Person B: "Hmmm, maybe if we wait a little longer..."
Re: (Score:2)
Re: You know, for Gamers! (Score:1)
Re: (Score:1)
Re: (Score:1)
How would that even work?
Re: (Score:2)
I had a computer instructor in the early 1990's who said that 4GB on a 32-bit processor was enough RAM for anyone to use. For the most part, he was correct. I had a 4GB system for nine years before I upgraded to a new motherboard with 8GB. The only time I ever use more than 4GB is when I'm playing a videogame or encoding video.
2017 might be AMD's year (Score:4, Interesting)
Re: (Score:3, Informative)
But 2017 is already the year of the Linux desktop...
Re: (Score:1)
Re: (Score:1)
1982 [wikimedia.org] was the year of AMD hardware.
Re:2017 might be AMD's year (Score:5, Funny)
Linux will never be on the desktop.
10 years from now, the linux powered terminators being commanded by the linux powered skynet, riding their flying linux powered bikes will still use Windows on their desktop computers, while performing experiments on installing linux on living people.
Re: (Score:3)
Sure as long as you only played their one bland boring dx12 game, ashes of the singularity. Anything else and dual r480's still usually get smoked by a standard 1080, let alone whatever the ti version might be. Hell they just about match a 1070
http://arstechnica.com/gadgets... [arstechnica.com]
Re: (Score:2)
I have only built AMD systems for nearly 20 years. My current AMD system is almost 4 years old now but still drives 5760x1080 pixels in most modern games with reasonable quality and frame rate (30fps or more). It also boots in less than 10 seconds and is an all-around snappy computer. I have not yet found a reason to upgrade.
News for nerds? (Score:5, Informative)
The "news for nerds" version of this story's headline is "AMD Unveils Vega GPU Architecture With 49 bits of Memory Address Space"
Re: (Score:3)
Re: (Score:1)
Re: (Score:2)
Did they blend a Beowulf cluster of them?
Re: (Score:3, Insightful)
The "news for cynics" version of this story's headline is "AMD unveils yet another set of powerpoints". Where is (Ry)Zen? Where is Vega? Every month is another month Intel and nVidia rule unchallenged on the high end. We need actual product on the shelf, not more tech demos. And I bet so does AMDs financials, you have to actually hard launch it before you get any revenue. I'm a bit hyped out, now it's more like hoping for a miracle.
Re: (Score:1)
I think he meant SKUs in retail outlets and in eShops that can be ordered, not vapor-hardware at a trade show.
Re: (Score:2)
Strange vapor-ware[sic]. Did you even read the article? The CPU exists, the motherboards exists, the coolers exists.
Well one of them does, at least...
Re: (Score:2)
Well, don't bother looking at any Microsoft Benchmarks. They killed RemoteFX performance by half in Server 2012. I still run my stuff on 2008R2 for this reason.
512 terabyte addr space should B enough 4 anybody (Score:2)
Is that how the saying goes? I am not sure :p
Re: (Score:2)
I really want to see AMD releasing a CPU competitive with Intel's latest offerings. I love my 8-core FX, but the real reason I bought it is that it costs almost 1/3rd compared to the competition.
Re: (Score:1)
Re: (Score:2)
Re: (Score:2)
Not really, sadly enough. Intel's single core performance is so above AMDs that for most computational tasks it makes little impact.
Re: (Score:1)
We been here before... (Score:2)
Re: (Score:2)
Well, that shouldn't have happened under normal conditions.
I was a professional videogame tester at the time.
You probably bought a card way out the league of your PC.
Back then I was probably upgrading my video cards every year. After I rebuilt my gaming PC for Windows Vista in 2007, I upgraded my video cards every three years.
Re: (Score:1)
Are they really allocating 49 pins on the CPU package to the address bus?
Will there be Vega-SX parts that are in much cheaper packages and only put out 40 pins?
512TB? Why? (Score:2)
What's the point? By the time we hit that amount of memory on a GPU, we're looking at this architecture being entirely obsolete.
Should've just said "We're slapping 1TB on this bitch!" and been done with it. No point in fussing about the scalability of the architecture when we're likely never going to see it hit full potential until long after its deprecated (AGP slot, anyone? When PCI-E cards came out, we'd barely even thought of saturating a 4X AGP slot.)
Re: (Score:2)
If you have a small address space then you need to write code that manually pages / caches the working set for an algorithm from storage. If you have a large address space then you use an interface similar to mmap and address the large dataset directly. It makes the code easier to write, and means that the paging / caching can be handled in hardware, where there are opportunities to speed it up.
Re: (Score:2)
Keep in mind large address spaces were here long before Vega. Hell, AMDs own "Graphics Core Next" architecture already supports flat 64-bit addressing, an that's been out since 2011.
Re: (Score:3)
Re: (Score:2)
Let's get real here (Score:3)
Someone check me on my logic here. The way I read this article is that AMD has created a new architecture with a memory controller that can address 512TB of memory address space. That's great and all but are we going to see cards any time in the near future with 512TB of GDDR on them? Not likely. How many years away are we? Who knows. It seems to me this is highly theoretical and possibly to put pressure on the memory industry to innovate on even more dense memory to push graphics even farther to the limit. It could also be to get some investor interest in the next "big thing".
Side question: How did AMD validate that their architecture works without actually being able to fabricate an actual board in practice, simulation?
Re: (Score:2)
Side question: How did AMD validate that their architecture works without actually being able to fabricate an actual board in practice, simulation?
You don't need to actually hook up memory to see if a memory bus works correctly. I used to test addressing on 8-bit CPUs using a Tektronix logic analyzer back in college.
Re: (Score:2)
AMD can actually connect SSDs directly to a GPU. So you can have your 32GB of HBM2 memory with ridiculous bandwidth, and a multi-TB SSD on the other side of the board
That's a neat idea. It's like you turn a PCI-E SSD into a modern day Voodoo 2 card
First step towards integrated CPU (Score:2)
Given the relative sizes of CPU's and GPU's, it makes sense that an 'APU' will be a GPU with a bundled CPU, rather than the other way round. Having a large address space is one requirement for doing virtual memory on a card.
AMD ruined ATI with shit drivers (Score:2)
512TB of ADDRESSABLE memory!!! Seriously, RTFA. (Score:1)
It's funny how not RTFA in Slashdot is even a meme but yet I can see that 95% of the people here didn't read it. The 512TB number is the amount of ADDRESSABLE memory, which means that you can reserve for example 300Gb of that memory to read a texture file that big. Then, as you start reading it, a secondary controller will transfer data there from main memory, directly from disk or from wherever. To you it will be as if you were reading a 300GB block from Video memory and thanks to that external controller