AMD's Fusion Processor Combines CPU and GPU 240
ElectricSteve writes "At Computex 2010 AMD gave the first public demonstration of its Fusion processor, which combines the Central Processing Unit (CPU) and Graphics Processing Unit (GPU) on a single chip. The AMD Fusion family of Accelerated Processing Units not only adds another acronym to the computer lexicon, but ushers is what AMD says is a significant shift in processor architecture and capabilities. Many of the improvements stem from eliminating the chip-to-chip linkage that adds latency to memory operations and consumes power — moving electrons across a chip takes less energy than moving these same electrons between two chips. The co-location of all key elements on one chip also allows a holistic approach to power management of the APU. Various parts of the chip can be powered up or down depending on workloads."
The simpsons (Score:2, Funny)
no "chipset" anymore; pr0n cache sniffers? (Score:3, Interesting)
But in this architecture, there is no physical barrier - they're on the same silicon.
Look for the bad guys to try to force the graphics drivers to sneak over and sniff the memory of the CPUs - I can imagine how they might be able to load some code in a pr0n movie that could tell some pointer in a GPU driver to point to addresses of cache which [at least ostensibly] belong to a CPU, at which point they should be able to r
vs Larrabee (Score:2)
Re: (Score:2)
It's not so interesting. I rarely wait on my CPU. It's my I/O and my GPU that hit the limits. When will NVIDIA make a GPU with a CPU core? That could be a real game-changer.
Re:vs Larrabee (Score:5, Insightful)
AMDs product is just a desperate attempt at trying to be relevant. They need to show they have a product competing with the big boys in all the right channels.
AMD is plenty relevant. It is Intel that scrambled to put out a 6 core desktop processor, which was so poorly planned that the cheap version is $1000. Meanwhile nVidia is desperately trying to get people locked into their CUDA API because their video cards just dont bang the performance drum like they used to.
AMD and Intel have different visions. AMD is clearly focusing on getting more cores on chip for more raw parallel performance (12 core CPU's in 4 chip configs are owning the top end server market.. brought to you by AMD), while Intel is clearly trying to maximize memory bandwidth to peak out raw single threaded performance (triple channel ram and larger cache is owning the software rendering and gaming markets)
Normal people are within the $50 to $200 CPU range, and at those price points, solutions from both camps perform about the same. On the video card front, you just can't beat AMD right now. Best price/performance ratio on top of best performance period.
Re:vs Larrabee (Score:5, Interesting)
The 6-core Intel processor is the Extreme Edition (always was introduced at $1000), and frankly smokes every other desktop processor out there.
AMD is the value-choice - they're cheaper at the same performance point, but they don't really compete in the over $250 desktop arena.
On the server front, Intel's introduction of Core2 based Xeons allowed it to compete again, and right now AMD is leader only in some cases in server performance (some are draws, but most I think go to Intel). Too bad, as server processors were producing a lot of money for AMD.
Intel is also leader in performance/watt, due to a complex power delivery architecture and better processor production facilities.
Meanwhile, AMD competes where it can on the processor front (but ruled the previous 6 months on the performance graphic front).
Re:vs Larrabee (Score:4, Insightful)
Intel is also leader in performance/watt, due to a complex power delivery architecture and better processor production facilities.
As long as you look only at raw CPU performance and power usage. Add GFX perf into consideration and...
(plus that would be quite recent development for Intel; their power consumption numbers weren't that great by themselves, when adding also chipsets of previous gen)
Re: (Score:2)
We go more and more towards integrated solutions. Accidentally, those are also the areas where perf/price ratio can really make a huge difference.
Re: (Score:3, Insightful)
The 6-core Intel processor is the Extreme Edition (always was introduced at $1000)
If ((not realistic for server marker) && (cant sell for less than $1000 without undercutting our other offerings))
{
setlabel("Extreme Edition");
}
Where is Intel's budget 6-core design? Is it because they refuse to make budget 6-core CPU's, or is it because they can't make budget 6-core CPU's?
Either way, the proof is in the pudding. They are not targeting the highly parallel market either by choice ("ignoring that market" scenario) or by mistake ("caught with pants down" scenario)
Re: (Score:2)
If ((not realistic for server marker)
Afaict other than not having dual socket support the i7 980x is pretty much the same design as the xeon 5600 series. So it seems they are selling this design into both the very high end desktop market and the server market.
Where is Intel's budget 6-core design?
At least according to wikipedia a cheaper model (though still pretty expensive) is coming Q3 of this year. I don't know how reliable that info is though.
BTW unlike most other extreme editions the i7-980x isn't tha
Re: (Score:2)
they don't really compete in the over $250 desktop arena.
Maybe they don't really want to.
Under $250 is by far the biggest market. Competing with the high end Intel chips would probably lose money.
Re: (Score:2)
Meanwhile, AMD competes where it can on the processor front (but ruled the previous 6 months on the performance graphic front).
Ahem -- still rules. http://www.amd.com/us/products/desktop/graphics/ati-radeon-hd-5000/hd-5970/pages/ati-radeon-hd-5970-overview.aspx [amd.com]
(And don't say it cheats because it has two processors, as Nvidia has been doing the same thing for their last two model lines, as well.
Re: (Score:2)
Well right now the fastest Supercomputer is an AMD. Next year maybe it will be Intel.
I think you are a correct in many ways one AMD is the choice in desktops that use CPUs in the under $250 mark. Which is probably close to 90% by volume.
AMD is or had better performance in visualization which is a huge market in servers.
Frankly the need for big fast CPUs is dropping. I work at a software company. Our flagship product was a monster. It really could use just about all the CPU, memory, and IO you could throw a
Re: (Score:2)
Most servers these days are running at well under 10%. You might as well run six servers on one box and save power, money, and space.
This is why AMD's 12 core chip is so attractive in the server market. On a 4-socket board thats up to 48 cores. Intels strategy just doesnt map well to this many-servers-in-one-box space at all. Sure there are some Big Boys that absolutely need plenty of memory bandwidth and map better to Xeon's, but thats just not a very big market. It doesnt keep the foundry running 24/7 printing money, the ultimate goal of any production facility.
Re: (Score:2)
AMD leads in availability of features. Intel has always locked features, unless you were willing to PAY for them. All AMD processors are sold with the same virtualization and other capabilities. No matter which price range I am looking at, I'll go with AMD. I never liked Intel very much.
Re:vs Larrabee (Score:5, Informative)
Re: (Score:2, Insightful)
Re: (Score:2)
"cleverly heating everything along the whole length of pipe."
Not to mention roasting any yarbles in the vicinity.
Re: (Score:3, Informative)
It should be plenty good for space and power consumption. Just look at Intel's US15W chipset which includes the GMA500 IGP.
It's tiny and consumes 2W compared to previous gen chipset + GPU setups (GMA950) that consume 15W, lengthening the battery life by a huge margin.
The chip itself has good performance, hindered only by terrible outsourced drivers (Tungsten, I'm looking at you), currently only optimized for video decoding (who said two smooth 1080p streams at less than 100% CPU usage using EVR in MPC?)
Comb
Re: (Score:2)
Now that ARM is trying to scale up and and the implementers are shipping SoC's with graphics optimizations, Intel and possible AMD on the other hand whats to do more in the mobile/small devices space. I wonder how this will play out.
Re:Relevant? (Score:5, Informative)
AMD designed/implemented the 64bit instruction that will be running our desktop PCs for decades to come.
Intel was the one scrambling to catch up [wikipedia.org] on that.
Re: (Score:2, Insightful)
Back in the days of Athlon64 vs Pentium 4 and Itanium AMD were ahead. Still since Core2 I'd say Intel are doing better. That being said Larrabee seems to be dead and I still think the idea has legs. Hopefully AMD will be to Larrabee what AMD64 was to IA64 - i.e. a more pragmatic version of the idea that ends up working better.
Re: (Score:2)
I'll enjoy my TWO systems with cash-to-spare that together trivially outperform your one system.
Re:vs Larrabee (Score:4, Informative)
How so?
AMD's offer is real, it uses a real performant GPU, not a GMA joke. Larrabee is stil vapourware, and it will be for a long time.
Re: (Score:2)
Intel's on-processor graphic in the i5-661 (the fastest on-board graphic from Intel) is trading blows with AMD's last generation (and current generation, as AMD didn't improve performance of its graphics core over its last generation).
If you're refering to an add-on card from AMD/ATI, then by all means the Intel IGP is crushed (just like the AMD's IGP or NVidia's IGP)
Re: (Score:2)
But when you actually look at the state of drivers...
Re: (Score:2)
AMD's current (and latest) IGP isn't better than the IGP they launched two years ago. As for promises, I've had plenty from NVidia, I'll wait until Fusion is in stores.
Re: (Score:2, Informative)
Intel's on chip GPU is just that - a GPU, and a primitive one at that. It can't even do OpenCL. It's certainly not a competitor to anything that AMD will release. Never mind Intel's appalling graphics drivers (and consistent history of poor driver releases), and benchmark cheating (so that they look competitive in reviews).
Re:vs Larrabee (Score:5, Insightful)
AMD and Intel need to have a contest on the shittiest driver category. I have one of each. Each revision of xserver-xorg-video-intel bricks my laptop in a new and exciting way. And AMD's fglrx is a steaming pile of rendering errors, inconsistent performance, and crashes.
On the other hand, both Intel [intellinuxgraphics.org] and AMD [x.org] have released specs and participate in open source development. So in the long run, either one is a better choice than NVidia. So I'll continue to complain about them and submit bug reports. It's the open source way.
Re: (Score:2)
Actually - AMD bought ATI with just this scenario in mind. Make the chips work together, and complement each other. How much more complementary can you get, than putting them on the same chip? Integrated circuits are MUCH faster than trying to pump everything through a pipeline!
Enough with hyping eye candy (Score:5, Insightful)
“Hundreds of millions of us now create, interact with, and share intensely visual digital content,” said Rick Bergman, senior vice president and general manager, AMD Product Group. “This explosion in multimedia requires new applications and new ways to manage and manipulate data."
So people watch video and play video games, and it's still kinda pokey at times. We're way past diminishing marginal returns on improving graphical interfaces.
I bring it up, because if you're trying to promote a technology that actually uses a computer to compute, you know, work with actual data, you are perpetually sidetracked by trying to make it look pretty to get any attention.
Case in point: working on a project to track trends over financial data, there were several contractors competing. One had this software that tried to glom everything into a node and vector graph, which looked really pretty, but didn't actually do anything to analyze the data.
But to managers, all they see is that those guys have pretty graphs in their demos and all we had was our research into the actual data... all those boring details.
Re:Enough with hyping eye candy (Score:5, Insightful)
|Bergman, senior vice president and general manager, AMD Product Group. "This explosion in multimedia requires
|new applications and new ways to manage and manipulate data."
So people watch video and play video games, and it's still kinda pokey at times. We're way past diminishing marginal returns on improving graphical interfaces.
Well sure YOU DO, but your Gran still has a 5200 with "Turbo memory" (actually that's only 3 years old, she probably has worse). This will be the equivalent of putting audio on the motherboard, a low baseline quality but done with no cost.
I bring it up, because if you're trying to promote a technology that actually uses a computer to compute, you know, work with actual data, you are perpetually sidetracked by trying to make it look pretty to get any attention.
Bloat is indeed a big problem, programs are exploding into GIGABYTE sizes, which is insane. OTOH linux reusing libraries seems not to have worked. There is too little abstraction of the data so each coder writes their own linked list, red-black tree, or whatever algorithm instead of just using the methods from the OS.
Case in point: working on a project to track trends over financial data, there were several contractors competing. One had this software that tried to glom everything into a node and vector graph, which looked really pretty, but didn't actually do anything to analyze the data.
Sounds like a case of "not wanting to throw the baby out with the bathwater." If they have someone of moderate intelligence on staff, that person can find a way to pull useful information out of junk data. He/she will resist removing seemingly useless data, because they occasionally use it and routinely ignore it. A pretty presentation can also be very important in terms of usability, remember you have to look at the underlying code but the user has to look at the GUI, often for hours a day.
But to managers, all they see is that those guys have pretty graphs in their demos and all we had was our research into the actual data... all those boring details. I can't comment on the quality of your management, but once again don't underestimate ease of use or even perceived ease of use (consider how long you will remain trying to learn a new tool if frustrated, the perception that something is as easy as possible is a huge boon... think iCrap).
Anyway back to Fusion, this is EXACTLY what Dell wants, bit lower power, less heat, significantly lower price and a baseline for their users to be able to run Vista/7 (7 review: better than Vista, don't switch from XP). So while it's true that this chip won't be dominant under ANY metric, and would therefore seem to have no customer base it's attractiveness to retail is such, that they will shove it down consumer throats and AMD will reap the rewards.
I'm curious about these things in small form factor, now that SD cards/MicroSD cards have given us nano-size storage we can get back to Finger sized computers that attach to a TV.
SFF Fusion for me!
Re:Enough with hyping eye candy (Score:5, Interesting)
| This will be the equivalent of putting audio on the motherboard, a low baseline quality but done with no cost.
I don't think you are viewing this correctly. I wish they didn't call it a GPU because your thought on the matter is what people are going to think of first. Instead think of it as the Fusion between a normal threaded CPU and a massively parallel processing unit. This thing is going to smoke current CPUs in things like physic operations without the need of anything like CUDA and without the performance limit of the PCIe bus. The biggest problem with discrete cards is pulling data off the cards because the PCIe bus is only fast in one direction (data into the card). This thing is going to be clocked much higher then discrete cards in addition to having direct access to the memory controller.
I don't think many have even scratched the surface of what a PPU (Parallel Processing Unit) can do or how it can improve the quality of just about any application ... I think this is going to be Hott.
Re:Enough with hyping eye candy (Score:5, Interesting)
What about memory? (Score:2, Insightful)
This thing is going to smoke current CPUs in things like physic operations without the need of anything like CUDA and without the performance limit of the PCIe bus.
Ummm, but videocard has its own super-fast memory (and a lot of it), and it uses direct access to system RAM, while this little thing will have to share the memory access and caches with CPU.
without the need of anything like CUDA
I dare to say, that this is totally false.
Re: (Score:2)
I agree with you. Something like CUDA will be required.
C//
Re:What about memory? (Score:5, Insightful)
while this little thing will have to share the memory access and caches with CPU.
Sharing the cache is not necessarily a bad thing. Its nice when the data that the CPU now needs is already sitting in L1 because the GPU just computed it, or vise-versa. That was, in fact, the point of the poster.
Re: (Score:3, Informative)
Well sure YOU DO, but your Gran still has a 5200 with "Turbo memory" (actually that's only 3 years old, she probably has worse).
What year are you living in?
1: Turbocache didn't exist until the 6100.
2: The 5200 is seven years old
3: You can apparently still buy them: eBuyer Link [ebuyer.com]
Re: (Score:3, Funny)
Re: (Score:2)
They specifically said about "creating, interacting, sharing ... managing, manipulating" - and you just dismissed that part and criticised "consuming"?
There is a quite popular usage scenario which is nowhere near diminishing marginal returns - video editing. Architecture of Fusion seems perfect for that. Will help also in image editing; even if its not so desparatelly needed in this case, it will come handy with what's enabling video boom - reasonably cheap digicams shooting fabulous 720p, even at this poin
Re: (Score:2)
Yeah, sure, go ahead and call it "crap"...but that sea of crap will give us many great videographers. Especially if large portion of them will be able to finally even afford quite sensible camera and editing rig (remember, world encompanesses not only developed countries)
I'm not saying it's crap. Other comments pointed out how this is far more than simply glomming a GPU onto a CPU and I don't doubt that. I'm complaining about the eye-candy oriented hype, and I'm stupefied as to how, even in third-world countries, there's this desperate shortage of video. Do you really think that their problems would be solved if only they could set up their own cable news networks?
Re: (Score:2)
...
If not "developed" then it's suddenly "third world"? For that matter, why does first world (numbering designation is a bit obsolete btw) accept the existance of indy videographers? Aren't they useless?
Re: (Score:2)
How well does it handle virtualization? (Score:2)
Re: (Score:2, Interesting)
It doesn't bring anything to the table yet. Firstly, IOMMUs need to be more prevalent in hardware, then secondly there needs to be support for using them in your favourite flavour (Xen will be there first) of virtualisation.
That said, we'll get ugly vendor-dependent software wrapping of GPU resources. Under the guise of better sharing of GPUs between VMs, but really so you're locked in.
Re:This has nothing to do with virtualisztie. (Score:5, Insightful)
They specifically were pointing out the benefit to having the GPU and CPU on the same chip which is quite a bit different than a mobo integrated solution. It probably isn't as powerful as a Xeon quad-core processor and a $500 video card but the question is how well it is setup to handle many different GPU tasks. I'd at least assume it's quite a bit faster for these types of tasks than a standard CPU and I wonder how well they can scale the technology for a better CPU and GPU.
I'm not sure I agree it's a niche market. I'd say more of a market poised to explode when the right products make it attainable. For virtualization it's more important that it can handle several unrelated tasks at a reasonable speed than that it can handle a single task at a high speed. If each CPU core also had a paired GPU it'd open up possibilities. Bulk, power consumption, and heat are often as big of issues for server farms as for laptops which is another reason why an interpreted GPU might be of interest.
Grid computing uses goes hand in hand with virtualization. Again coming down to how well these can work in parallel. Being able to fit a number of CPU and GPU cores on a single physical chip could be very beneficial I think.
Moving electrons (Score:4, Informative)
Re: (Score:2)
Last I checked nobody was quite sure if all electrons moved or some electrons moved. Has this actually been ironed out? Do any of these chips actually switch fast enough that your statement is correct regardless?
Re: (Score:2)
Re: (Score:2)
Adding buffers will increase the power of the chip.
Just to be clear, you mean power consumption, right? And to be still more clear, adding buffers increases latency because you have to wait for more transistors to switch? And of course more power means more opportunity for noise in other systems...
Re: (Score:3, Informative)
Re: (Score:2)
Electrons will drift( look up 'electron mobility' on Wikipedia), but the GP is right in that it is the wave motion of the potentials that are primarily the means in which the information will travel. At the same time, he is just being a bit nit picky.
Agreed, and it's a pretty minor nit. Electrons do move (there would be no current if they did not), they just don't have to move all the way for the electric potential to reach the destination.
The nit I would have picked would have been with the phrase "the sa
Re: (Score:2)
Yeah! (Score:5, Interesting)
I'm hoping moving things into the CPU will make it easier to take advantage of the huge parallel architecture of modern GPUs.
For what, you ask?
I'm personally interested in sound synthesis. I play the piano, and while you can get huge sample libraries (> 10 GB), they're not realistic enough when it comes to the dynamics.
Instead people have been researching physical models of the piano. So you simulate a piano in software, or the main components of it, and extract the sound from that. Nowadays there are even commercial offerings, like Pianoteq (www.pianoteq.com) and Roland's V-Piano. Problem is that while this improves dynamics dramatically, they're not accurate enough yet to produce a fully convincing tone.
I think that's partly because nobody understands how to model the piano fully yet, at least judging from the research literature I've read, but also very much because even a modern CPU simply can't deliver enough FLOPS.
Re: (Score:2)
I thought the whole instrument modeling thing died because disk space is cheaper than processing power. That 1000GB of samples fits on a $100 drive.
just like my Core i3, then (Score:2, Insightful)
Just like my Core i3 sitting about 20 inches to the left, then. Yes, I know they're incorporating a better GPU, but they're touting too much as new.
Re:just like my Core i3, then (Score:5, Informative)
The technical difference is that while your Core i3 has its GPU as a separate die in the same packaging, AMD Fusion has the GPU(s) on the same die as the CPU(s). The Intel approach makes for shorter and faster interconnects, the AMD approach completely removes the interconnects. The main advantage is probably (as is alluded to in the summary) related to power consumption.
Re:just like my Core i3, then (Score:5, Interesting)
I also heard that they *share* FP math units between CPU and GPU.
Re: (Score:2)
You're entirely correct, sorry! Inferring wrongly from a high-level flow diagram, I thought the second die was used for PCIe/memory, with graphics on the CPU die.
Re:just like my Core i3, then (Score:4, Interesting)
Well, "incorporating a better GPU" makes quite a bit of difference, considering i3/i5 solution isn't much of an improvement almost anywhere (speed - not really, cost - yeah, I can see Intel willingly passing the savings...anyway, cpu + mobo combo hasn't got cheaper at all, power consumption is one but mostly due to how Intel chipsets were not great at this); and seemed to be almost a fast "first" solution, announced quite a bit after the Fusion.
Re:just like my Core i3, then (Score:5, Interesting)
This is where Intel's monopolistic behaviour rears its ugly head. In the past, the GPU needed to be integrated on the motherboard. Now it's on the CPU but Intel motherboard chipsets cost the same as previous generations. Seems like a terrific opportunity a market for 3rd party chipset vendors to make an offering (like the good old days when you could choose from VIA, Nvidia, SiS, Intel,
But wait, Intel will no longer allows 3rd parties to produce chipsets for their CPUs and keeps the profits from the artificially inflated chipset market to itself. Intel may have the performance crown, but its reasons like this (and the OEM slush funds to lock out AMD from Dell and other vendors) that keep me from supporting "Chipzilla"
Whoa, graphics on the CPU? (Score:4, Insightful)
Let's party like it's 1995! Again! [wikipedia.org]
Slightly less cynically, isn't this (in like-for-like terms) trading a general purpose CPU core for a specialised GPU one? It's not like we'll get more bang for our buck, we'll just get more floating point bangs, and fewer integer ones.
Re: (Score:2)
Seems a bit better example... [wikipedia.org]
Re: (Score:2)
Slightly less cynically, isn't this (in like-for-like terms) trading a general purpose CPU core for a specialised GPU one? It's not like we'll get more bang for our buck, we'll just get more floating point bangs, and fewer integer ones.
Until it takes less than four intel CPU cores (This is a psuedorandom number, it's what I recall from some intel demo) to do the job of a halfway decent GPU, this approach will be rational for any users who care about 3d graphics. Intel would like us to have "thousands" of CPU cores (I assume that means dozens in the near term) and to ditch our GPUs and the change cannot come fast enough for me... but it's not here yet.
It's that time again... (Score:2)
Very few people need more than a dual-core - the cores just sit their twiddling their bits. Sacrifice a core or two for a good GPU, and you have massively simplified the design of the system, saved power and saved space.
Sure, it's not a new idea - in IT we seem to progress in spirals. It's time this idea came around again...
Re: (Score:3, Interesting)
It's not like we'll get more bang for our buck, we'll just get more floating point bangs, and fewer integer ones.
You can accelerate integer operations as well on "new" GPUs. This means that for highly parallel, data independent operations you will get a ton of bang for your buck and without having to send data to the graphics memory first and then pulling the results back.
holistic is (Score:2)
Open Source drivers? (Score:5, Interesting)
Will the drivers for the graphics be open source or will we be crawling out of this proprietary driver hole we have been trying to climb out of for over a decade?
Re: (Score:2)
The more things change, the more things stay the same.
Don't expect to have high quality, high performance open source drivers as people high in the company will think the information revealed in those drivers will help the competition
Re:Open Source drivers? (Score:4, Informative)
The documentation needed to write 3D graphics drivers has been consistently released by ATI/AMD since R5xx. In fact, yesterday I was setting up a new system with a RV730 graphics card which was both correctly detected and correctly used by the open source drivers. Ever since AMD started supporting the open source DRI project with both money, specifications and access to hardware developers things have improved vastly. I know some of the developers personally; they are smart and I believe that given this support, they will produce an excellent driver.
It's sad to see that with Poulsbo Intel did quite an about-face, and stopped supporting open source drivers altogether. The less said about nVidia the better.
In conclusion, seeing who is making this Fusion chip, I would have high hopes for open source on it.
APU (Score:3, Funny)
Looks like the Quickie Mart has a lawsuit on their hands.
Meh. (Score:2, Insightful)
Sounds like a non-advancement to me.
"Look, we can build a VCR *into* the TV, so they're in one unit!"
Yeah, so when either breaks, neither is usable.
Putting more points of failure into a device just doesn't sound like a great idea.
In the last 4 computers I've built/had, they've gone through at least 6-7 graphics cards and 5 processors. I can't remember a single one where they both failed simultaneously.
Now, if this tech will reduce the likelihood of CPU/GPU failures (which, IMO, are generally due to heat or
Re:Meh. (Score:4, Insightful)
Sounds like you need a new power supply, or a surge suppressor, or a power conditioner, or an air conditioner.
You shouldn't see that many failures. Are you overclocking like mad? Silicon should last essentially forever compared to other components in the system, as long as you keep it properly cooled and don't spike the voltage. Removing mechanical connectors by putting things on one die should mean fewer failure modes. A fanless system on a chip using a RAM disk should last essentially forever.
A single chip with N transistors does not have N failure modes. It's essentially tested and will not develop a failure by the time you receive it. A system with N mechanically connected components has a failure rate of N*(probability of failure of one component), and it's always the connectors or the cheap components like power supplies that fail.
Re: (Score:2)
I'm willing to bet that the GP buys cheap crap like ASRock and generic PSUs that couldn't perfor
Re: (Score:2)
Then you'd be wrong.
Firstly, I've been building PC's since perhaps 1984. (I wouldn't include the early computers that I built from kits in 80-81.) So we're talking over a long, long span of time.
I learned early on that you get what you pay for - shit components=shit performance.
Thus PRECISELY the point I was making: when sinking a lot into individual components because you're not buying cheap crap, it's useful to be able to purchase incrementally.
Now, I'll answer all the other commenters: first, recognize
Re: (Score:2)
Further, ambient temperature shouldn't be any issue to the CPU if your HSF is good enough. I don't let my C
Re: (Score:2)
Well you should have mentioned the rural challenges earlier!
Those of us in more urban settings with more controlled climate and power don't experience these problems and yours should not be taken to be normal experience. Particularly, I've never seen a CPU failure and motherboard and video card failures are nearly always the result of bad capacitors.
Re: (Score:3, Interesting)
32C? Is that considered high?
I've played Call of Duty 4 in my cheap P4 in the summer. I don't know the temperatures inside, but outside there were 40C and I have no A/C. All with stock cooler too. The CPU is now seven years old and still works perfectly.
In fact, the only thing that died in that cheap system was the power supply due to some construction workers in another floor which connected their machines directly to the building's power without protection and caused a power surge.
Re: (Score:3, Interesting)
Re:Meh. (Score:5, Informative)
People who don't know better seem to skimp on the power supplies more than anything else.
I can understand cheap boards; they'll (usually) last the useful life of the system provided they're not really crappy. But the power supply is existential: it's the heart of the system.
If it doesn't pump your electricity properly (at the correct rates and the like), your brain and various peripherals will die a slow death. Sometimes it is not so slow.
Invest in a decent power supply: it's worth it. It's probably the only part of a typical user computer I'd consider an investment, too, because it is an insurance policy (of sorts) on the parts. Buying a cheap power supply so you can get a UPS is backwards. Your components are still going to be getting crap power if the PSU is crap.
I've had a total of one power supply failure, 2 disk failures, and 0 peripheral/RAM/CPU/motherboard failures in the 12 years I've been buying my own parts to build systems.
The current PSU I've got in my main home computer is a Seasonic something or other (they, and Antec, I've found are very good). I'm amazed at how good this converter is: yes, it's got PFC and all those bells, which certainly help, but it delivers amazingly consistent power, evening out the voltage nicely. Hell, we had the power go out for long enough to stop the motor in the washing machine, make my wife's laptop go to battery, and kill the lights, and make my LCD lose power: the computer didn't turn off (and no, I'm not currently using a UPS). This little power supply caches enough power for a full second or so of operation while playing a CPU and graphics intensive game.
So yeah, paying $70 or more for a PSU does not seem unreasonable in the least. With PSUs, you're paying more so for quality than you are for advertised performance or anything like that, so throw down the cash.
Re: (Score:2)
Re: (Score:2)
Wow, obviously you are a consumer. I cannot imagine anyone worth their pay in the business world replacing individual components in a computer. Usually, it is just tossed.. or handed back to the oem to get fixed..
Second, when is the last time you had a processor fail?
Re: (Score:2)
Re: (Score:2)
1 year is definitely unusual for the life on a PC. My parents still have a nice Dell from 2001 (with the ram tripled from the original and an extra hard drive) which is working fine. The floppy drive on it died but nothing else has had a fault. But ignoring the anecdotes, the data shows that the average life is far more than one year. Here in the EU it is not legal to offer less than 2 years warranty so they need a decent survival rate just to stay in business.
Re: (Score:2)
You may have missed the memo - Intel is the largest supplier of graphics units for PCs and Mac. And no - none of their graphics units are discrete. They're all mounted on the motherboard. Just like the audio controller. And the USB controller. And SATA controller. And NIC.
And for some strange reason, there's still a market for discrete controllers.
What this is doing isn't taking away your choices. It's giving you more choices. Though, to be realistic, it's probably aimed more at OEMs and business who have n
Re: (Score:2)
Discrete controllers (video, sound, RAID, NIC) offer better performance than motherboard-integrated options.
The integrated parts are fine for the vast majority of users, but many users, especially professionals in various endeavours, need better.
Re: (Score:2)
You are having way to high of a failure rate.
Fusion is going into notebooks first which don't tend to get a lot of cpu or gpu swaps.
Re: (Score:2)
These days, the mechanical components of computers are the only parts that fail. Fans and non-SSD hard drives will need replacement. Chips and circuits won't. In fact, your computer should be shutting itself down if critical fans fail (though your power supply may cook itself).
Always use a UPS and be sure to use dust filters on your intakes if you find your fans choking to death. Every few months, blast it all out with an air can.
Your computers won't fail if you do this.
Re: (Score:2, Informative)
Taken as a whole, GPU+CPU is simpler and more robust than two separate components connected via an external bus. It does away with connectors, bus drivers (need something to drive those signals across connectors and inches of trace) , level shifters (external busses don't operate at the same voltage as core silicon), bridges (external busses are often shared by multiple devices) and all of the complexity, signal integrity issues and points of failure that these things introduce. GPU+CPU on one die means t
Future Multi CPU + GPU Combos (Score:2)
Packing the GPU into the CPU makes a lot of sense but also raises some questions.
Does this mean that in the future we can have chips that contain not only a multi-core CPU but also a multi-core GPU? For example could AMD pump out a frag-tastic 6 CPU + 4 GPU chip for hardcore gamers and scientists?
How is this going to effect the cooling for the chip? If I fire up Crisis will my computer melt? (Assuming a GPU is packed in with enough power to play crisis.)
Also how is this going to effect memory bandwidth? Mos
Re: (Score:2)
GPU's are already highly "multi core". They have been for years. I am not sure if they call them cores, they are slightly different from CPU cores but it is effectively the same idea. his si why GPU's are so much better than CPU's at graphics and some other parallel tasks with things like CUDA.
Re: (Score:2)
Simple answer no.
They are not taking everything that makes up a video card and putting it all on the chip, they are just integrating some GPU like functions into a CPU.
Re:heat (Score:5, Funny)
Re: (Score:2, Funny)
"retarded cunt" is completely offensive and a terrible distraction which hinders your message.
In the interest of sensitivity, please use the term to "mentally challenged cunt" in future.
Thank you.
Re: (Score:2)
This is not going to make "faster gaming systems". It's just another, cheaper and lower power form of onboard graphics. Gamers are definitely still going to be using discrete graphics for now.
Re:heat (Score:5, Insightful)
AMD chipsets with integrated GFX were quite good at power consumption already; using a dozen or so watts. Considering AMD puts out quadcores with sub 100W TDP, Fusion shouldn't be that big a problem.
Re: (Score:2)
Yes, heat is going to destroy this thing if they want to make it a Fermi plus an overclocked Extreme Edition processor.
However, if the graphic performance is limited based on available (main) memory bandwidth (from which the main processor also takes a chunk), they don't need more than a quarter of a 5850, and starting with a 65W TDP processor, they're in the 125W TDP (where they released plenty of processors).
If they drop graphic performance even further, they co
Re: (Score:2, Informative)
This demo was of Ontario - AMD's low power solution for netbooks and low-end notebooks. This will be using the low power Bobcat cores and probably something similar to an HD5450 graphics-wise.
I seriously doubt heat is going to be an issue.
Re:heat (Score:4, Interesting)
This GPU-on-the-CPU is targeting the mobile/lightweight market.
Think about how the other solutions work. That GPU chip sits next to the CPU chip and they both must be connected to the system bus in order to access ram. With AMD's solution here, you remove that GPU chip and therefor also remove the external BUS connection that it required. This is a very big win for manufacturers, who would even pay a premium for the chip because of the lower production costs. But knowing AMD, they wont be charging a premium for it. Instead they will try to push Atom's out of the market.
Re: (Score:2)