Follow Slashdot stories on Twitter

 



Forgot your password?
typodupeerror
×
Data Storage Hardware Technology

Samsung '3D' Memory Coming, 50% Denser 87

CWmike writes "Samsung on Tuesday announced a new 8GB dual inline memory module (DIMM) that stacks memory chips on top of each other, which increases the density of the memory by 50% compared with conventional DIMM technology. Samsung's new registered or buffered (RDIMM) product is based on its current Green DDR3 DRAM and 40 nanometer (nm)-sized circuitry. The new memory module is aimed at the server and enterprise storage markets. The three-dimensional (3D) chip stacking process is referred to in the memory industry as Through Silicon Via (TSV). Samsung said the TSV process saves up to 40% of the power consumed by a conventional RDIMM. Using the TSV technology will greatly improve chip density in next-generation server systems, Samsung said, making it attractive for high-density, high-performance systems."
This discussion has been archived. No new comments can be posted.

Samsung '3D' Memory Coming, 50% Denser

Comments Filter:
  • Does ram really use that much power.
    Now 40% power savings on the latest 3D accelerator would be awesome. Probably help with heat issue.
    • by Anonymous Coward on Wednesday December 08, 2010 @02:13AM (#34484076)

      Googling a bit, one test showed 2x1 GB of memory consuming up to 7.28 watts.
      http://www.tomshardware.com/reviews/hardware-components,1685-13.html

      For PC, that's practically nothing. For mobile devices, every watt counts.

      • by 0100010001010011 ( 652467 ) on Wednesday December 08, 2010 @02:37AM (#34484184)

        Not just mobile. Newer generations of HTPCs, Plug like devices are using 20W. The AppleTV2 has a 6W power supply. Assume they overspec'ed it by 20%, that's 5W at full tilt.

        7W is a huge % of those numbers.

        • by 0123456 ( 636235 ) on Wednesday December 08, 2010 @03:17AM (#34484338)

          Not just mobile. Newer generations of HTPCs, Plug like devices are using 20W.

          Yeah, I measured my MythTV frontend at 26W from the wall; so if the 4GB of RAM is taking 14W, that would be more than half the total consumption of the entire system.

          • by MrNemesis ( 587188 ) on Wednesday December 08, 2010 @05:30AM (#34484782) Homepage Journal

            Out of curiosity, what hardware are you using? I've just picked up one of the new ASRock Vision 3D HTPC's (great little machine for Myth/XBMC; works OotB with Linux for everything except the IR receiver, although for some reason amazon won't publish my review) that pulls 23W from the wall on a bad day, and idles at about 17W at idle. My old C2D-based mATX box pulled more like 50-60W.

            But yeah, I've never been able to quantify those power usages of memory. I think they must take an absolute worst case scenario along the lines of "if every bit was flipped at once" or something like that. DIMMs even run cooler than they used to, making those ubiquitous heatspreaders all the more ephemeral.

            • by 0123456 ( 636235 )

              Out of curiosity, what hardware are you using?

              Probably a bit late now, but that's a Zotac Ion motherboard in a small ITX case with some 'silent'-ish fans, 4GB of RAM and an X25-V SSD. It doesn't really need 4GB, but since it's running off a cheap SSD I wanted to push all temporary storage into a RAM disk to reduce SSD writes.

              • Thought about buying an Ion myself, but I found the atom sucked for most non-video stuff; was crappy with XBMC and youtube bits (running XBMC on top of debian myself).

                Also using an SSD, a 30GB OCZ vertex, but from my experience with the Intel drives you can safely use them for temp storage. Also got 4GB in the ASRock (which is essentially just intel and nVidia laptop components in a Mac-mini-esque chassis) and it's laughable how little of it linux + XBMC ever use :)

                Surprised the atom eats so "much" power th

      • by Anonymous Coward

        For stationary devices that run on battery anything that peaks over 0.1W is unacceptable. When you are expected to run your stuff for at least 2 years on 3 AAA cells you'd better cut down the average consumption to the uW range.

      • by bigdaisy ( 30400 )

        Looking up some datasheets on Kingston's "valueram.com", 2x1GB DDR2 DIMMs use about 1.0-1.4W depending on clock speed. That drops to about 0.8-0.9W for DDR3 modules.

      • by spydum ( 828400 )

        For a server with say, 16x4GB dimms, that can add up real quick. Consider a farm of 40 such machines. Every watt counts.. when dealing in extreme scales (both small and large)
        Not to mention, heat generated is just as significant as power drawn.

        • Yes, cooling costs on a large server are substantial -- you have to run huge cooling towers to cool them down.
    • by Requia ( 1734466 )

      How much it is depends on what kind of limits you have. Server power draw can run up against building or power grid limits, at which point every watt counts.

      • by ihavnoid ( 749312 ) on Wednesday December 08, 2010 @03:45AM (#34484440)

        Additionally, an average server has 2x cpus, 8x memory, while having 0x graphics compared to an average desktop. Another problem is that we are running out of tricks for reducing dram power, which means that the portion of dram power may increase steadily in the near future.

        Even graphic cards have a sizable, high-bandwidth ram on-board.

        Trust me, DRAM power consumption is becoming a serious probpem.

        • by Ailure ( 853833 ) on Wednesday December 08, 2010 @08:49AM (#34485642) Homepage

          Trust me, DRAM power consumption is becoming a serious probpem.

          So is apparently cosmic rays. ;)

        • Perhaps more importantly, RAM power is close to constant. If your CPU load is low, you can underclock the CPU and lower the power usage. You can spin down disks when they're not in use. Pretty much any other component of a modern computer can be powered down when not in use, but RAM needs to constantly refresh its contents. This means that it is consuming power at a pretty constant rate. It takes slightly more power to read or write, but not very much.

          In theory, an OS could swap things out of most RAM

          • Suspend to disk defeats the purpose of suspend in many use cases. Point in fact: it's ass-slow. A suspended laptop is pretty close to instant-on. Also, a laptop that is writing its memory to disk can't be thrown in the backpack until the disk shuts off without risking a head crash.
        • So I'm guessing that much of the DRAM power budget is taken because of the requirement to Refresh. At 1 transistor per cell, DRAM has been 4x less expensive than SRAM, before taking consideration of economies of scale. So where is the SRAM market? Why do we still not see an alternative with better speed, and power, at the cost of price and storage density? Why can't I make that choice? And while we're on the subject of MIA memory technology, where are the FRAM devices? Those would be flash-based SSD killers

          • I heard that since sram cells are larger (6 transistors actually, 4 for the latch and 2 for select), on very large chips the bus capacitance overwhelms the advantage given by faster sensing.
    • Your average desktop with 2-4 sticks of 8 or 16 module memory isn't a concern. When you're talking servers with 16 sticks of 36 module memory and ECC, it really adds up.
    • by afidel ( 530433 )
      DDR3-1333 RDIMM's use about 5W each, multiply that by the 18 DIMM's available in a fairly typical 1U or 2U server and it adds up. For an even more extreme example look at the HP DL980, 128 slots so 640W just for the memory.
    • by Jeprey ( 1596319 )

      Yes. It really does.

      The reason is that to get the signals off the chip you have to amplify them and then take the losses of having to line-charge the bond pads, bond wires, package traces, PCB traces, etc. This charging is simply done due to physical parasitic losses induced by shipping the data off-chip. Keeping it all on-chip avoids this and allows nA-uA currents to be used throughout rather than kicking things up to mA currents and then back down again.

      This, combined with jitter limits, is part of the

  • by Anonymous Coward

    Great, does the CPU now need 3D glasses too ?

  • by Anonymous Coward on Wednesday December 08, 2010 @02:09AM (#34484058)

    Core memory is static in the true sense of the word. I've got core memory that hasn't changed a bit in 60 years. Punks !! You don't know memory.

  • by DWMorse ( 1816016 ) on Wednesday December 08, 2010 @02:14AM (#34484078) Homepage

    It'll fit right in with my ex's computer. Stupid P.O.S. Gateway.

    *takes a deep breath...* NOW WHEN SHE TYPES IN ALL CAPS and overuses LOL ON FOXNEWS.COM and adds a thousand!!!!!!!!!!!!!!! EXCLAMATION POINTS... her memory can be just as dense as she is.

  • by Anonymous Coward

    One way to look at this is "oh good, people have been talking about stacked chips for years, and they're finally rolling it out for mass production. Another tool to increase density. Yay!"

    The other point of view: "The geometries aren't going to be shrinking much longer, so chip makers are starting to turn to desperate measures to keep Moore's law going. This will work once or twice, but when the shrinks stop, and the chips are already stacked, we're going to run out of roadmap, probably soon".

    Not sure wh

    • by HiThere ( 15173 )

      Well, IBM seems to think the next step is liquid coolant. Then you can just keep stacking them higher. Not sure myself. I don't really like the idea of water inside the chips, and there doesn't seem to be a good replacement for freon. (Or maybe there is. What do modern refrigerators work on?)

      • tetrafluoroethane. [wikipedia.org] A fluorocarbon.

        Similar to what is in compressed air dusters. (usually difluoroethane. [wikipedia.org])

        With both compounds boiling at room temperature though, your ram chips will be internally pressurized, which means mechanical stresses during heating and cooling cycles.

    • We've got the 3rd dimension to fully use. Then we have memsistors (I won't call them memristors because it isn't a portmanteau and is sort of stupid sounding). These two things will be able to feed our high rate of growth for some time. But it will come to an end soon. Maybe within 30 years before we find something else to keep pushing us forwards.

      And if we ever DO have a period with no technological progress we have created ourselves a comfortable buffer zone... Software efficiency can be improved grea
  • From ECC, buffered DIMM's to RDIMM's in the Mac Pro's?
    • I'd be interested to see if they put it into a new generation of apple Xserve rack-mounted servers.

      • You mean... the ones [apple.com] they've already announced will be discontinued?
        From Wikipedia: "On November 5, 2010, Apple announced that it would not be developing a future version of Xserve."

    • Apple are already using DDR3 ECC (they don't say if it's registered or unregistered but I suspect it's registered) in the mac pro and xserve. It's not like apple had a lot of choice in the matter, memory controllers are now in the CPU so the CPU vendors call the shots as to what will be supported.

  • by nounderscores ( 246517 ) on Wednesday December 08, 2010 @02:59AM (#34484284)

    Anything that reduces the cooling load and the power bill will be welcome.

  • This is great for the big business side of things, but how soon will we see this on the consumer level? I mean, we keep seeing all these really high spec systems being used for the governments and large operations, but nothing for the little guys? TFA gives no hints.
    • Thing is the memory support on desktop boards is already ahead of what most people need even with todays "bloatware". LGA1156 supports 16GB and desktop LGA1366 suports 24GB yet even among "enthusiast" forums the consensus seems to be that 8GB is plenty.

  • Great. (Score:5, Insightful)

    by olsmeister ( 1488789 ) on Wednesday December 08, 2010 @03:36AM (#34484400)
    We've added another dimension, and got 50% denser. Sounds like we didn't do our jobs very well.
    • by ThreeGigs ( 239452 ) on Wednesday December 08, 2010 @05:06AM (#34484680)

      I've always wondered if there was a reason why manufacturers didn't use both sides of the silicon for lower powered chips, like memory. Seems like a win-win... twice the component count for the same silicon investment. Yeah, handling might be tricky, but not a showstopper.

      • by guruevi ( 827432 )

        They already do, buy a bit more dense memory than you're used to (or can afford) and you'll see it happen.

        This I believe is talking about stacking multiple chips on one of the sides, probably in the same packaging as a single chip.

        • by tlhIngan ( 30335 )

          This I believe is talking about stacking multiple chips on one of the sides, probably in the same packaging as a single chip.

          Not a new technique, either. It's just another stacked die - where you have multiple chips stacked on atop the other. Stacked dies have been commercially available for at least 5 years now (usually they're used in flash chips).

          Some form of packing together multiple dies has been around. We've had multi-chip packaging (like the Pentium Pro), package-on-package (where you put two ICs on

      • by Anonymous Coward

        Except for the plummeting yields, which could easily - or even likely - mean *more* wasted silicon.

      • Re: (Score:2, Informative)

        by Anonymous Coward

        They don't use both sides because the back side is where the robot handlers touch the wafer to move it. At several steps in the wafer process it is vacuumed down to chucks to hold the wafer and keep it flat. If you did print on the back the pattern would be damaged by all of the backside handling and ruin the chips back there. There is also the issue of front to back wafer alignment. While I am sure some college kids or some profs will come on and try and quote some things from some text books and sales

    • Is it really 'adding another dimension' or more like stacking up more thinner pancakes?
  • Will this perhaps give us a chance at having cubical memory stacks to plug into our motherboards like tiny processors? I could really enjoy 2GB RAM in a little 3/8"x3/8"x3/4" stack. Key it right and save costs on PCB. Might be able to be cooled just as easily.

    • Re:Cube memory? (Score:5, Interesting)

      by wierd_w ( 1375923 ) on Wednesday December 08, 2010 @04:31AM (#34484586)

      3D geometries have serious issues with line saturation and heat dissipation. This is because of thermal noise, and the increased voltage needed to overcome it. (which in turn, creates more heat.)

      We are already at the point where high performance RAM chips need heat spreaders, and that is with 2D chip geometries that can eliminate heat reasonably efficiently.

      When you start stacking multiple silicon fab layers together, heat builds up in the layers, requiring more voltage to overcome thermal noise, which produces more heat...... You get the idea.

      Without separating the layers with some kind of highly thermally conductive intermediate to pipe the heat out, the insides of the chips become little easy bake ovens, and estimated service life drops radically, as does performance metrics.

      I could see them going 2 levels deep in the geometry, with a special package with heat spreaders on both sides (of the package itself that is- not the DIMM) or something crazy like that-- but I really can't see a big "solid 3D block" of silicon getting plugged anywhere. IF such a technology were to come into being, it would need to be made from something that is damned near to being a room temperature superconductor to keep from being unreliable/a fire hazard from thermal noise.

      Alternatively, it could be done in a photonic computing approach, using optical transistors and optical interconnects... that would solve the heat problem too, but would make servicing the system substantially more difficult.

      • There are patents going back a decade pertaining to using microfluidic ducts as a heat transfer mechanism. Every few months now, there's another article on slashdot about one of the chip giants testing out such manufacturing techniques. Just a few links from a quick googling...

        http://www.xbitlabs.com/news/coolers/display/20031008155430.html [xbitlabs.com]
        http://www.electronics-cooling.com/2002/11/electroosmotic-microchannel-cooling-system-for-microprocessors/ [electronics-cooling.com]
        http://www.frostytech.com/articleview.cfm?articleid=2424 [frostytech.com]
      • by Khyber ( 864651 )

        For heat dissipation, just make the entire outside of the module the heat sink. It's what I do for ultra-power LED diodes, and lemme tell you, those get WAY hotter than any RAM chip could dream of, plus pull more power (some of these diodes are 100w a piece.) Drop a fan on it for when you overclock, just like normal. No big change in anything, really.

        Microfluidics got mentioned, but really that's pointless without a huge phase change section, and that addition renders my idea of 3D RAM useless, plus fluid+

      • The square cube law is always the elephant in the room when people start talking about 3D circuits. It is certainly a problem, but the field is still open to improvements. For example, the "through silicon via" process presumably means they etch a via entirely through a silicon wafer and plate it with a metal. These could also be used as heatsinking aids and not just ways to transfer signals through vertically stacked chips, and though some surface area is consumed it may be more than made up for by the

    • by dargaud ( 518470 )
      Why is DRAM so large compared to flash memory ? I mean, I have a 32Gb micro-SD card in my phone and it's smaller than a fingernail. But the 8Gb of DRAM in my desktop take 4 large slots and at the time (last year) there weren't any 4Gb modules available in this category. And also I had to add a fan and run them at 666MHz instead of their rated 800MHz or I get hard failures. Doesn't sound very exciting to me.
      • by vlm ( 69642 )

        Why is DRAM so large compared to flash memory ?

        flash is nice, but incredibly slow, especially in write, compared to dram. We are talking several orders of magnitude here, not just 10% or something. Also if you have 100 meg write rate (wishful thinking) and the drive burns out at 100K rewrites (wishful thinking) and its about 10 gigs, the numbers divide out to the drive will be dead in about 100 days. Different technologies have certain tradeoffs and flash is nice and small and low power and nonvolatile, but it is slower than molasses and short lived.

      • Most flash memories use a serial interface providing access to large amounts of memory through a small number of pins. The price is higher latency for memory accesses. DRAM uses a parallel bus to minimize bottlenecks at the cost of needing many more connections to a chip or module. Even RDRAM is parallel to some extent. Furthermore, about half the pins in a modern day memory module are grounds to minimize crosstalk at the high switching speeds. The I/O requirements for high speed memory all conspire to forc

  • I'm just waiting for the day when Intel and AMD will be competing on not the number of cores in the CPU, but the number of circuit layers on their 3D chip.
  • I find this funny... (Score:2, Interesting)

    by Lumpy ( 12016 )

    I have been doing "3d" ram stacking for decades... I did it first in 1983 on a TRS-80 Color computer. I had 2X the max supported ram the machine could handle. I simply used a toggle to switch ram banks, later I added logic to allow the computer to do that for me. Writing programs that consumed most of ram and stored data in the other bank were fun...

    What ele is samsung going to discover that hardware hackers have been doing for ever and a day?

    • they're talking about stacking the dice, not the devices. You know what dice are? They're the little chips of silicon that are then packaged to make the IC's that you typically see and use. Unless you can precisely align and drill little tiny microscopic holes in the dice and electrically connect the one on top to the one on bottom, then you haven't been doing what they're doing. Not even close.

      The closest anyone has ever got to this is stacking small dice on a larger die and wire bonding the pads of one to

    • by Fishbulb ( 32296 )

      Exactly. The Amiga 1000 I bought in 1988 had a hack like this done by the prior owner (in fact, it's still in my attic). Tripled the motherboards' memory (256 to 768k iirc...), and since the Amiga would detect any memory in the system and just tack it onto the address space, no configuration headaches. Damn, those were the days. :) (FWIW, it had that piggy-back chip hack, the front-loaded mem expansion, I added a 1.5 MB daughterboard that plugged into the CPU socket, and finally added some SIMMS to my X

    • Except for the fact that this development is absolutely nothing like what you describe. But hey, who let anything like logic stand in the way of a "I used to do X back in the day" post?

  • Radio Shack COCO 1 anyone?

    My BF did that to mine for me like back around 1980 or so..

  • Interesting that TSV is found to be useful after all. 29 years ago an AMD employee independently conceived of TSV and AMD refused to talk to the employee about this and other concepts, nearly all of which have subsequently been developed and patented by AMD's competitors.

  • by Junior J. Junior III ( 192702 ) on Wednesday December 08, 2010 @12:26PM (#34489226) Homepage

    As long as I don't have to wear those stupid glasses, I'm all for this 3D memory.

I owe the public nothing. -- J.P. Morgan

Working...