DDR4 RAM To Hit Devices Next Year 233
angry tapir writes "Micron has said that DDR4 memory — the successor to DDR3 DRAM — will reach computers next year, and that the company has started shipping samples of the upcoming DDR memory type. DDR4 is more power-efficient and faster than DDR3. New forms of DDR memory first make it into servers and desktops, and then into laptops. Micron said it hopes that DDR4 memory will also reach portable devices like tablets, which currently use forms of low-power DDR3 and DDR2 memory."
Would have gotten a FP except (Score:5, Funny)
... I'm still stuck on good ole DDR2
Realistically, while there are benefits for "faster", it's no substitute for reducing inefficient bloatware.
Re:Would have gotten a FP except (Score:5, Interesting)
Re: (Score:3, Insightful)
Well, one way we can reduce power consumption is to go to operating systems that aren't as bloated. If you've tried the Windows 8 Consumer Preview, you already know that Windows 8 isn't just the worst product Microsoft has ever made - it's also bloatware. Microsoft would be better off making an XP 2014 release and selling it.
The same with LXDE as opposed to bloatware like KDE.
Another thing is screen savers - not only not needed, but a total waste of energy. Just have the OS turn the stupid screens off
Re:Would have gotten a FP except (Score:4, Insightful)
LXDE == Lubuntu Linux?
Good release.
>>>capable of executing 1,000 times more instructions per second than the original pc
Heh. More than that. The IBM PC was 4 megahertz? And now we have double-clocking where CPUs execute instruction on both rising & falling edges. And dual-core CPUs are now standard, so 3000*2*2/4 == 3000 times faster. And yet as you pointed-out we still have to deal with annoying "wait" states while the PC thinks or redraws a screen. Bloat.
Re:Would have gotten a FP except (Score:4, Insightful)
Instead of competing on features, why not have a 6-month moratorium where people just fix current bugs? It would make everyone more conscious of bad practices that lead to bugs in the first place, hopefully reducing future breakage (and slow/fugly code to work around buggy cruft).
Re: (Score:2)
Because, to paraphrase many WONTFIX bugs on the openoffice project (under Sun's watch): It's less fun to fix bugs than to focus on new features.
Re: (Score:2)
This is *so* true of so many projects. It's a pervasive problem - nobody likes to do a bughunt - you're basically seen as the maid cleaning up after other people's messes.
Maybe there should be a "you can only add 1 feature for every x number of bugs you've removed - and your 'x' reverts to zero every time someone else finds a bug you created."
Similar to "you can only add n bytes of code for new features if you first remove n+1 bytes of code w/o losing any existing features or introducing any bugs". Even
Re: (Score:2)
Re: (Score:3)
Maybe you meant XP?
Windows 7 isn't brain-dead over swapfiles like XP is. It doesn't swap unless you're really out of RAM (an amazing breakthrough!)
Re: (Score:3)
Unfortunately, in Windows 7 it is possible for the hard drive check to consume over 11 GB(!!!!) of RAM!! Believe it or not that is the ONLY process I have ever had use that much RAM (12GB in my laptop, and I run Photoshop, Gimp, Illustrator, Lightroom, DPP, and embroidery design apps - oh, and virtualbox)
Re: (Score:2)
My Core Duo laptop runs Windows 7 very smoothly with 2.5GB. My girlfriend has an identical laptop, except for a a slower HDD and 1.5GB RAM and it also runs better than XP or Ubuntu for her. Modern OSes take advantage of all the memory you have... if you have 16GB RAM, it will use as much otherwise-unused memory as possible for optimization. If you only have 2GB, it will allocate as much as needed to apps and still use the rest for optimization.
Re: (Score:3)
I understand the purpose of RAM. However, I also understand that the reason my applications were terminating without warning was that Windows was killing the apps in the foreground to give the background scan more RAM. With 12GB of RAM I've seen little to no need to run a swap file. To complete that scan while getting work do
Re: (Score:2)
So I can turn-off the virtual memory in Win7 without any problems? I've always been told to avoid that with previous versions of windows.
Re: (Score:2)
If a program requires more memory than you have, and you turn off the swap, that program (and maybe your whole computer) will crash. However, that is unlikely to happen, but with smaller RAM sizes (
Re: (Score:2)
You can run XP on a half-gig of ram and no swapfile.
Just like a lightweight desktop such as LXDE runs fine in 1 gig with no swap, even when running something like eclipse or libreoffice.
Of course, all of these are still bloated compared to the days when a multitasking OS (microware OS9 Level 2) would run multiple copies of Flight Simulator in separate full-screen windows, a bunch of text terminals, and a copy or two of Rogue, all in a half-meg of ram.
Re: (Score:3)
My XP install doesn't run on only half a gigabyte and no swapfile. I've tried it, and it refuses to open any programs (except tiny crap like notepad). Forget trying to browse the web or run Word.
>>>microware OS9 Level 2
Heh. Commodore Amiga OS multitasked in just 1/4 meg of RAM.
Re: (Score:2)
>>> It doesn't swap unless you're really out of RAM
That's not what I've heard. I've read that Seven, like Vista, precaches everything into RAM that it believes you will need and makes it full. This of course means if you access something new (or unexpected), it has to first swap out something from RAM to the HDD in order to make room for the new item. Slooooow.
PLUS you didn't really answer my question. Can Windows 7 run with the virtual memory/swapfile set to 0? How much Real RAM would I need
Re: (Score:2)
No, it just has to use that memory for whatever you want to load. When the cpu generates a page fault, the OS will just have to load the code that used to be cached off the hard disk, same as any other cache miss.
Re: (Score:2)
Windows Vista/7 will run without swap on 4 GiB of ram if you are a light user. You do want 16 if you do more than web browsing and editing text files though. And apparently some programs (such as Photoshop) will bitch if you don't have any swap.
Re:Would have gotten a FP except (Score:4, Informative)
You say that like a bad thing? Low-end machines today have 8GB. My desktop at work has 32GB. I want Windows to "waste" as much memory in an effort to minimize physical I/O as it possibly can! Needing to go to even the lowest-latency-on-the-market SSDs means potentially "wasting" 250 thousand CPU cycles. An HDD, more like 25 million. And if you actually need to wait for an idle HDD to spin up or a network request... Ouch!
In any case, keep in mind that flushing the FS cache doesn't mean hitting the pagefile. You may need to actually hit the disk if you then go to access something that got dumped, but you would have needed to anyway if Windows hadn't cached it in the first place.
Re: (Score:2)
Seriously. I don't mind the OS caching itself in the RAM so long as I don't need that RAM at the moment. This was a more serious issue several years ago when the OS was twice the size of the RAM installed on most machines; the amount of memory available today is large enough to store Windows itself, with all the whistles, completely, with enough left over for Visual Studio & friends. And the fact that the OS is taking a throw-away approach (not swapping, but throw-away) to caching itself in the RAM is f
Re: (Score:2)
That's not what I've heard. I've read that Seven, like Vista, precaches everything into RAM that it believes you will need and makes it full. This of course means if you access something new (or unexpected), it has to first swap out something from RAM to the HDD in order to make room for the new item. Slooooow.
Nope, how 7 works is that it allocates unused RAM to precache stuff, yes, but if it needs that RAM it doesn't push it to swap (which would be stupid, since that basically is even slower than just reloading it in the first place), it just dumps the precached stuff from memory, which is (practically) instantaneous. You can even see in the Performance Manager how much it is pre-caching (usually fills as much as it can: again, because it gives no performance penalty with potentially large performance returns).
Re: (Score:2)
Re: (Score:2)
Re: (Score:2)
Which gives me an idea. Hats, on Steam, from MS, when you send them the scores from the Windows Experience program.
Re: (Score:2)
Re:Would have gotten a FP except (Score:5, Informative)
How does crap like this get modded insightful? Oh wait.. it's because it plays up to the bigoted prejudices that prevail on this site.
1. I've actually used the Windows 8 preview on a 4 year old PC and it is more responsive than Linux for desktop use. I don't like Metro, but everything under the hood in Windows 8 is in very good shape and some changes to the UI could make it a good successor to Windows 7.
People on this website who brag about being Linux "experts" because they got Ubuntu to boot one time should know the difference between the UI presentation layer and the underlying OS services. Unfortunately a bunch of self-proclaimed "experts" who troll this site are anything but.
2. I also use KDE on the desktop and I've used LXDE. Guess what? KDE is faster for my use because of the ability to reconfigure its setup. I don't want or need a taskbar to switch between apps, and because of KDE's flexibility I have a very efficient keyboard shortcut system in place to handle window management. Additinally, yakuake gives KDE a big edge for handling the konsole in a smart way and guake (which cloned yakuake) is still not as good.
Firefox under KDE starts up in the same amount of time as on LXDE.. and so does every other application I try. Windows don't move faster across the screen on LXDE either and they resize at the same speed on both desktops!
Re: (Score:2)
Guess what - you don't need a taskbar to either launch or switch between apps in LXDE. It makes me wonder if you even tried it, or are just repeating someone else's BS.
Re:Would have gotten a FP except (Score:5, Interesting)
Actually, I am running the Windows 8 Consumer Preview on the same hardware that I was previously running a clean XP installation, and Windows 8 is definitely snappier, plus has better search/launch functionality. I can't say that I am particularly fond of the Metro UI (I mostly use the Explorer-style interface), and I preferred the search UI in Windows 7 to the one in Windows 8. But saying that Windows 8 is a worse OS than such champions as Vista, 98, and ME is quite a stretch.
Re: (Score:2)
There was a time when XP could run on 64MB of RAM. I normaly use a 1GB computer that can barely run it nowadays.
Re: (Score:2)
Windows 8 isn't just the worst product Microsoft has ever made - it's also bloatware
Funny how less memory, CPU, while booting faster turns into "bloatware". I would love to see your definition for the word.
Re:Would have gotten a FP except (Score:4)
Bloatware: software I dislike and wish to deride but for which I am unwilling or unable to give reasons why.
Re: (Score:2)
Re: (Score:2)
You're probably comparing Windows ME on hardware at that time with Windows 8 on today's hardware. Put ME on the same hardware and it will FLY.
Just for fun, I once made a boot cd that loaded dos 7.0 (just 'cuz). I'd then launch Windows for Workgroups from the cd. Remembered how slow it was back in its' day - but launch time was under a second on a stupid 900mhz machine with a quarter gig of ram - and all the ram was available to WfW.
Both ME and Vista were real bugfest when they came out - and both impro
Re: (Score:2)
Re:Would have gotten a FP except (Score:5, Insightful)
Luckily we're getting both. I just purchased a video card that's twice as powerful as my current one, and only uses 2/3 the power. I'm upgrading from a CPU using up to 130W to just 77W, but still gaining 20-25% performance.
Those are some good jumps in performance, but great leaps in efficiency. Total power consumption is a big factor moving forward in trying to reduce what we need from the grid.
Re: (Score:2)
Re: (Score:2)
Luckily we're getting both. I just purchased a video card that's twice as powerful as my current one, and only uses 2/3 the power. I'm upgrading from a CPU using up to 130W to just 77W, but still gaining 20-25% performance.
Those are some good jumps in performance, but great leaps in efficiency. Total power consumption is a big factor moving forward in trying to reduce what we need from the grid.
7950 and 2500k?
Comment removed (Score:5, Insightful)
Re: (Score:3)
A fashionable opinion on Slashdot, no doubt, but go back and actually try out an older piece of hardware. I bet it will seem absolutely bog-slow. I remember the days not so long ago when I would shut down everything to fire up a browser (Netscape), and really think hard before opening a new window (no tabs, of course). Now I sit here with two browsers, each with dozens of tabs, mp3s playing in the background, bit-ticket software like Photoshop and Illustrator running, and a disk-scan going, without the slig
Re: (Score:2)
A fashionable opinion on Slashdot, no doubt, but go back and actually try out an older piece of hardware. I bet it will seem absolutely bog-slow. I remember the days not so long ago when I would shut down everything to fire up a browser (Netscape), and really think hard before opening a new window (no tabs, of course).
I remember those days. It was 1996, when I had 4MB of RAM on my laptop and had to run both Apache and Netscape for web development. I was really glad when I managed to get another 4MB and eliminated the perpetual swapping.
Otherwise, unless you had an insanely low amount of RAM or were running Vista, I can't see why you'd have had that problem 'not so long ago'.
Re: (Score:2)
What's a long time ago to you? I'm talking 2001ish I guess. I had a G3 250mhz PowerBook with 64mb of RAM. Classic Mac OS sucked at swapping--that's my whole point: software improved along with hardware.
Re: (Score:2)
Re: (Score:2)
Me too, my 2008 machine has a relatively cheap SSD in it, and I'm not going to need to upgrade until it breaks (despite doing computationally expensive stuff like motion graphics).
Re: (Score:2)
This. While software vendors certainly deserve some part of the blame for eating more cycles, much of that is not bloat, and any realistic analysis of the problem must also take into account usage patterns. Does Photoshop use more cycles and more RAM than it used to? Yes, for certain. It's also able to do many more things than it used to, and is regularly run on huge images by relative standards. I also think nothing of having a browser with 20-30 tabs open, while listening to MP3s, editing a photo, a
Re: (Score:2)
Well, that's the funny thing. If you optimize too much, the code will only run on that one machine. With increasing levels of abstraction, we run into increasing costs.
You have the engineer's dilemma -> write one program that eats several cores & 40 GBs of RAM, but runs on every machine, or you write one program that uses 5 processor cycles & 1 KB of RAM, but runs only on one machine.
Re: (Score:2)
My primary computer is a 7 year old laptop with 512M of RAM and it works great.
After you turn off all the crap, XP takes up like 50M of RAM plus 100M of "System Cache", whatever that is.I haven't fully tweaked it out of sheer laziness: some guy built a "Micro XP" distro that can get it to run in 64M.
Thing is, any stupid browser takes up more memory than the entire operating system, and leaks *heavily* due to the insanity that is JavaScript. The browser alone will easily eat all the RAM available. Don't get
Re: (Score:2)
I just made the switch from DDR2 to DDR3 in March and only did it because it's cheaper and easier to get a DDR3 motherboard and 16GB DDR3. Current mobo supports up to 32GB RAM, so I'll probably be good until DDR5 comes out. I still have a number of PCs and Servers on DDR and DDR2 and foresee it staying that way for a while.
Re: (Score:2)
What do you think it would take to convince motherboard manufacturers to put a few more DIMM slots on those boards?
I want 256GB on my main machine, but I don't want to use a server motherboard (not enough expansion slots).
Re: (Score:2)
More importantly, what does "faster" mean?
Higher bandwidth or lower latency? It's supposed to be both, but my guess is that latency is mostly not affected.
Great (Score:5, Funny)
I predict a 33% performance increase going from DDR3 to DDR4 based on my own super-secret analysis of the press release.
DrrDrrArr (Score:2)
Re:DrrDrrArr (Score:5, Informative)
QDR's already around. In fact, a popular console already uses it. It's still heavily patented though, so it's not very appealing.
The Playstation 3 has 256MB of XDR-DRAM by RAMBUS (yes, that RAMBUS). It does QDR - two bits on falling edge, two bits on rising edge (using multi-level signalling).
It's tricky for memory because the bus speed is high, signalling ovltages low, and motherboard traces bad enough that the eye window is very small, so a lot of (patented) tricks are needed to "open up" the eye and recover the bits from it. Impedance mismatches are a killer (and they happen at connectors especially).
Re: (Score:3)
I know that the PS3's RAM is soldered directly onto the mainboard; but that is normal for consoles. Does RAMBUS' secret sauce allow them to handle less controlled environments(in servers, say, if you can't do at least 8 DIMMs per socket you might as well go home) or are there technical reasons, as well as legal togetherness issues, that drove them to pursue specialty embe
Re:DrrDrrArr (Score:5, Informative)
DDR2 effectively *is* QDR –it transfers 4 words per clock cycle... It just doesn't do it in quite the same way that true QDR RAM would. DDR3 effectively is ODR (octa-data-rate) RAM. DDR4 will effectively be HDDR (hexa-deca-data-rate) RAM.
Latency? (Score:5, Interesting)
Re:Latency? (Score:5, Informative)
13 clock cycles according to the all-knowing Wikipedia [wikipedia.org], so similar to the latency increas going from DDR2->DDR3; theoretically it will be made up for by increasing clock frequency, I guess, with DDR4 starting at 2133 MT/s (unfortunately I'm not clear on how transfers/s translates to MHz for DDR4 - is it the same two transfers per quad-pumped cycle?).
Re:Latency? (Score:5, Informative)
Latency has significantly decreased, thanks to higher clock frequencies. See the chart on this page: http://en.wikipedia.org/wiki/CAS_latency [wikipedia.org]
But RAM will always be slower than L1 and L2, simply because of the size of the memory.
Re: (Score:2)
Actually, it is the proximity to the CPU core that is the primary mitigating factor here. A 512MB on die Cache will be faster than one off chip (assuming competent designers) because you can clock the RAM much faster when the CLK (clock) signal has to travel microns rather than inches.
Re: (Score:2)
Not so much size (Score:2)
But type and location. DRAM has worse access times than SRAM for various reasons. Also there is simply the distance from the processor. When you start wanting super low access time, distance matters. That's why L2 and L3 are on CPU dies these days. For L1, even that isn't enough, it has to be near the core to get the kid of speeds you want there.
The good news is with judicious use of caching, you can have your cake and eat it too for the most part. You can use cheap DRAM for most of your memory, but get ove
Re: (Score:2)
Incorrect, CAS on DDR3 1333 parts is ~7 transfers. Meaning about 4.5-5ns.
Re: (Score:2)
has favored pure throughput to latency
Hey, sounds [apcmag.com] like [skytopia.com] the Linux OS.
Sadly, I think you believe it ... (Score:2)
My Linux box turns on in under 10 seconds (from sleep mode - didn't have that in the IBM PC/XT days) and I get right to work. All of my apps are already open and ready to go, and Internet Connectivity is up and running (You remember the Internet and WiFi from the 80's right?). Try booting an IBM
Re: (Score:2)
If you try say,
Re: (Score:2)
GCC: Warning - Conditional not reached.
Re: (Score:2)
Re: (Score:2)
Yes. Most of what he said is ridiculous. ... and I blockquote:
Re: (Score:2)
Quote1:
The main problem was that there simply was not a convincing way to prove that staircase was better on the desktop. User reports were not enough. There was no benchmark. There was no way to prove it was better, and the user reports if anything just angered the kernel maintainers further for their lack of objectivity.
Quote2:
And there are all the obvious bug reports. They're afraid to mention these. How scary do you think it is to say 'my Firefox tabs open slowly since the last CPU scheduler upgrade'? To top it all off, the enterprise users are the opposite. Just watch each kernel release and see how quickly some $bullshit_benchmark degraded by .1% with patch $Y gets reported. See also how quickly it gets attended to.
Quote 3:
Then I hit an impasse. One very vocal user found that the unfair behaviour in the mainline scheduler was something he came to expect. A flamewar of sorts erupted at the time, because to fix 100% of the problems with the CPU scheduler we had to sacrifice interactivity on some workloads. It wasn't a dramatic loss of interactivity, but it was definitely there. Rather than use 'nice' to proportion CPU according to where the user told the operating system it should be, the user believed it was the kernel's responsibility to guess. As it turns out, it is the fact that guessing means that no matter how hard and how smart you make the CPU scheduler, it will get it wrong some of the time. The more it tries to guess, the worse will be the corner cases of misbehaving. The option is to throttle the guessing, or not guess at all. The former option means you have a CPU scheduler which is difficult to model, and the behaviour is right 95% of the time and ebbs and flows in its metering out of CPU and latency. The latter option means there is no guessing and the behaviour is correct 100% of the time... it only gives what you tell it to give. It seemed so absurdly clear to me, given that interactivity mostly was better anyway with the fair approach, yet the maintainers demanded I address this as a problem with the new design. I refused. I insisted that we had to compromise a small amount to gain a heck of a great deal more. A scheduler that was deterministic and predictable and still interactive is a much better option long term than the hack after hack approach we were maintaining.
Disclaimer: I'm not sure ho
Re: (Score:2)
I can't take you seriously anymore.
Re: (Score:2)
Re: (Score:2)
So you want links to the part of the brain in the developers that shows that they are not afraid to mention these? Is Kolivas claiming that the developers are somehow removing reports of performance issues from lkml?
How do you propose I refute a rhetorical question?
Re: (Score:2)
That doesn't take anything away from the fact that as far as average desktop users are concerned, latency is given (to put it politely) second priority. In any case, a 0.1% increase in bandwidth performance at a cost of 2-4x latency drop in GUI responsiveness is pretty short-sighted in my opinion, no matter which way you look at it, even if he was exaggerating somewhat. Especially if the desktop is a goal for Linux (which
Re: (Score:3)
There is no problem. Any problem encountered amounts to improperly configured kernels. If I select Voluntary Preemption rather than Preemption, like Ubuntu does, then I too will get a much slower GUI response. Kern
Re:Latency? (Score:5, Interesting)
Actually, RAM latencies have slightly improved over time, it's just not as fast as transfer rate, so the units (number of missed transfers) make it look like it's getting a lot worse. The main reason that RAM latencies haven't improved much is because they're not that important in the grand scheme of things.
In reality, it takes around 200 transfers to get from the CPU asking for something to getting it, of that, only about 7-9 are the RAM. An improvement of one transfer, makes that 199 transfers, instead of 200 – yay, we gained 0.5%. Except that in reality, the gain is not 0.5%, because in reality, most of the CPU's requests are in level 1 cache... Make that 0.005%. Except that in reality, the gain is not 0.005%, because in reality, most of the CPU's requests that are not in level 1 cache are in level 2 cache... Make that 0.00005%... You get the idea.
The real way to sort out the latency issue is via tighter integration of things onto the CPU (hence why we've seen memory controllers move on board, and more levels of faster cache), not in skimming one or two cycles off how quickly the RAM responds.
Re: (Score:2)
If your application does cache misses 99% of the time, then latency still matters.
Also, higher latency could allow removing some cache levels, which would make the CPU faster and free up some transistors.
Re: (Score:2)
From the benchmarks I've seen on DDR3, the increased clock speed does seem to increase performance up to around 1.6GHz. What I haven't seen is a comparison between max clock speed on DDR2 and DDR3.
Re: (Score:2)
There have been a few reviews involving modern DDR3 1066-1600, and the difference between 7-7-7(CAS-CasToRas-RAS) and 11-11-11 is less than 1% performance across nearly every benchmark. Multiple cores coupled with huge amounts of cache with advanced pre-fetch units has all but nul
Tell me again how this improves my life... (Score:3)
Slightly lower power consumption. Slightly faster memory. Sorry, but it's looking to me like just another way of obsoleting my portable faster, without significant performance improvement.
Re: (Score:2)
Like the car industry of the 1950s, the computer industry has now reached the point of incremental tiny improvements rather than revolutionary improvements (like jumping from 8 bit to 32 bit in one decade). I've had the same PC for 10 years and it still runs everything just fine (except the latest flash update). It would have been impossible to run a 1985 PC with Windows95 and the latest software.
Re: (Score:2)
Re: (Score:2)
As someone above pointed out, the biggest gains these days are in power efficiency, not performance. You can upgrade your CPU/motherboard and video card and probably get perhaps 25% more performance, but with less than half the power usage, which over a year will probably pay for itself in reduced electric bills (even more so in southern climates with A/C).
Re: (Score:2)
No, you just can't figure out your numbers... Remember, the jump from 32 bit to 33 bit is as big as the jump from 0 bit to 32 bit ;).
Re: (Score:2)
Re: (Score:2)
The price/GB sweet spot does seem to migrate to the 'current' flavor, after a period of new-hotness pricing; but the RAM industry doesn't seem to be pursuing its sinister forced upgrade strategy very aggressively...
Re: (Score:2)
Too late.
Re:Tell me again how this improves my life... (Score:5, Informative)
The initial DDR4 models will be only marginal increases over DDR3, true. But remember how the original DDR3 models were only marginally better than DDR2, or even how some initial DDR2 modules were *worse* than DDR?
DDR3 is hitting a wall, where increasing the frequency any further is causing exponentially higher power usage and heat. I can't find any air-cooled DDR3-1866 or DDR3-2133 - every module I can find is water-cooled, because that's the only way to dissipate the heat. DDR4 begins at DDR4-2133, apparently without even needing a heat sink. And it's expected to scale to double those speeds, over time. And *those* you *can* upgrade - if you buy a DDR4-2133 device now, you can upgrade to DDR4-3200 or DDR4-4266 whenever you wish, if your memory controller supports it.
DDR4 is also making a rather significant shift in architecture, going from a dual/triple/quad-channel-memory paradigm to a point-to-point system. So better scalability with multiple modules.
Oh, and one quote cited a 40% decrease in power usage compared to an equivalent DDR3 module. That's hardly "slightly" lower.
Re: (Score:2)
Secondly there are no ddr4 devices now, because as the article summary said, ddr4 won't be out till next year. It will be pin incompatible with ddr3 (to protect it from wrong voltages and different signal methods). Also, ddr3-1866 and higher ram is available (I just bought some), they come with air cooled heatsinks, just t
You can all thank me for this (Score:5, Funny)
No pricing (Score:2)
Screw It. I'm going straight to DDR5. (Score:2)
Which devices? (Score:2)
Intel has already confirmed that the 2013 "tock", Haswell, will still use DDR3.
Not sure about AMD's position, but this sounds like DDR4 will wait on desktops and laptops for 2014 or 2015.
Re: (Score:2)
Doubtful. You can bet you'll see Asus and Gigabyte having boards out late this year for testing with full releases probably early or mid-next year in time for the new CPU cycle.
Re:Isn't it time to drop the "D"? (Score:4, Informative)
Re: (Score:3)
Re: (Score:2)
Right, and DDR2 transferred 4 times per cycle, and DDR3 8 times per cycle. Neither sounds very "double" data rate to me, except when referring to the previous generation ;D.
Re: (Score:2)
Are we ignoring the fact that the major players in this very industry were at one time colluding to keep prices artificially high [arstechnica.com] until the DoJ stepped in?
Re: (Score:2, Offtopic)
ARPANET - one project out of millions that produced anything even remotely useful. TCP/IP? Big deal. Networking does not rely on just the protocol and there were plenty of protocols already and more would have been created, it's not like the gov't was needed to push phones or radio usage and development.
Re: (Score:2)
Sure! if you ignore the fact that it was government money that helped to pay to lay telephone lines across the country in the first place.
Re: (Score:2)
The facts are not what you think.
About 3000 competitors were destroyed by the gov't to give AT&T its monopoly [google.com].