Follow Slashdot stories on Twitter

 



Forgot your password?
typodupeerror
×
Intel Hardware

Intel Moving Forward With 10nm, Will Switch Away From Silicon For 7nm 279

An anonymous reader writes: Intel has begun talking about its plans for future CPU architectures. The company is already working on a 10nm manufacturing process, and expects the first such chips to be ready by early 2017. Beyond that, things are getting difficult. Intel says it will need to move away from silicon when it develops a 7nm process. "The most likely replacement for silicon is a III-V semiconductor such as indium gallium arsenide (InGaAs), though Intel hasn't provided any specific details yet." Even the current 14nm chips they're making ran into unexpected difficulties. "While Intel didn't provide any specifics, we strongly suspect that we're looking at the arrival of transistors based on III-V semiconductors. III-V semiconductors have higher electron mobility than silicon, which means that they can be fashioned into smaller and faster (as in higher switching speed) transistors."
This discussion has been archived. No new comments can be posted.

Intel Moving Forward With 10nm, Will Switch Away From Silicon For 7nm

Comments Filter:
  • amazing (Score:5, Interesting)

    by schlachter ( 862210 ) on Tuesday February 24, 2015 @12:07AM (#49117069)

    Amazing that we're getting to 7nm, and rather than saying we can't do it, there's just casual talk about how they will have to switch away from silicone. Really incredible. Will they just keep marching forward to less than 7nm and into other exotic configs?

    • by ciaran2014 ( 3815793 ) on Tuesday February 24, 2015 @12:16AM (#49117087) Homepage

      Nope. They've decided to hit 7nm and then call it a day.

      • Re: (Score:2, Funny)

        by Anonymous Coward

        They've decided to hit 7nm and then call it a day.

        I asked Gordon Moore about this and he said it would be illegal.

      • One wonders whether if they reach 'the limits of silicon' whether implementing optimization techniques in hardware will be the next iteration.

        e.g. VLIW inspired designs such as Transmeta Crusoe, Elbrus 2000 or Mill CPU.

        • by itzly ( 3699663 )

          They've always done a lot of hardware optimization techniques. But advanced hardware techniques go hand in hand with extra transistors.

          • Re: (Score:2, Interesting)

            Those are the Tock's in Intel's Tick/Tock model. [wikipedia.org]

            Tick is smaller structures.
            Tock is new architecture.

            Each new architecture is optimized for the most common tasks at that time, together with a bazillion other changes. If they figure out a general optimization technique that still works with the x86 instruction set in the mean time, they'll go for it.

            The problem with some optimizations is that they do not work with the x86 instruction set. Abandoning that instruction set is expensive, although we are doing it

            • by itzly ( 3699663 )

              The problem with some optimizations is that they do not work with the x86 instruction set.

              I don't see why the x86 instruction set is a problem. Just translate them on the fly, as they've been doing for years.

              • by Anne Thwacks ( 531696 ) on Tuesday February 24, 2015 @05:12AM (#49117777)
                Just translate them on the fly, as they've been doing for years.

                You can, and people do. However, the issue is not translating one x86 instruction to one [insert ISA here] instruction. That has been done since x86 was invented, and was common with previous ISAs before that. The real requirement is to translate source code that maps to a bunch of x86 instructions into ONE [trendy ISA] instruction. This will obviously be easier if x86 is thrown out the window.

                Historical note: x86 is a bastadised rip-off of the PDP11 instruction set. The PDP11 was built as a "hardware Fortran machine" ie one instruction represents one Fortan instruction as far as was achievable in 1970. C is (just one) PDP11 assembly language! The VAX instruction set was an attempt to achieve a higher level machine code, which worked quite well - most VAX assembly instructions are actually function calls to application specific microcode.

                X86 was a poor ISA when the first 8086 chips were made (but good, given hardware capabilities at the time). That was about 40 years ago. MIPS and Sparc (and ARM) are all better than x86.

                The moral of this story is that it is "first past the post" in this game, cos people hate it when their favorite app stops working. (See Great Western Railway, Brunel and 8' gauge).

                • Couldn't Intel do a phased shift to a different ISA, and leave people to run older code in emulators? Different OS' would have to build in emulators, but it would probably work. Force compilers to start outputting new code withe the new isa. It wouldn't be different than the x64 move... When it would have made the most sense to replace x86...
                  • by itzly ( 3699663 )

                    It's not even clear that a new ISA would actually improve performance by a meaningful amount.

                  • by LWATCDR ( 28044 )

                    They tried that with the Itanium and it did not go well.
                    What Intel might do is faze out older parts of the ISA like the 16bit x86-x286 instructions to free up some space on the die. Even that might not be worth the effort since I am pretty sure those are already "emulated" on modern CPUs in the decoder.
                    I thought that it would a good idea for Intel to go only 64 bit on their mobile chips. They have no real installed code base in mobile to worry about.

                • by itzly ( 3699663 )

                  The real requirement is to translate source code that maps to a bunch of x86 instructions into ONE [trendy ISA] instruction.

                  No, the real requirement is to execute the program as quickly as possible. If that can be done by mapping N->1, that's great, and I'm sure Intel already does that where they can. But if you can get the same speed by using multiple instructions in parallel, that works too.

                  X86 was a poor ISA when the first 8086 chips were made (but good, given hardware capabilities at the time). That was about 40 years ago. MIPS and Sparc (and ARM) are all better than x86.

                  No, the x86 is a good ISA. You may not think it's pretty, but it gets the job done, as their market shares proves. It's also enlightening to look at the ARM ISA. From the original ARM1 to the latest ARM Cortex, there's been a clear trend

                • by AmiMoJo ( 196126 ) *

                  I think translation is becoming less important now because a lot of code is compiled from an intermediate form anyway, e.g. Java and .NET. If you look at x86 Android performance, which is a mix of Java byte-code compilation to x86 and binary translation of ARM to x86 performance just isn't an issue any more.

                  The other big issue used to be boot time with hardware that contained x86 code in ROM, executed by the BIOS. Apart from the security implications that meant that you needed special PCI cards for Macs whi

        • by tigersha ( 151319 ) on Tuesday February 24, 2015 @03:29AM (#49117551) Homepage

          Intel was heavily invested in VLIW, and developed Itanium. That did not go well, and AMD brought out x64 and ate their lunch. Intel adopted AMD's instruction set and Itanium is basically dead now.

    • The thing is, atoms are very, very small, but they still have a finite size. A hydrogen atom, for example, is about 0.1 nanometers, and a caesium atom is around 0.3nm. The atoms used in silicon chip fabrication are around 0.2nm.

      source: http://www.extremetech.com/com... [extremetech.com]

    • Re:amazing (Score:5, Insightful)

      by FlyHelicopters ( 1540845 ) on Tuesday February 24, 2015 @12:32AM (#49117119)

      There is some debate among people if 5nm will make sense or even be reasonable to do...

      Can a 5nm transistor be made? Sure.... Can 5 billion of them be packed onto a chip and sold for $200? That is a different question...

      Going to 5nm only helps if it is a functional product that is better than what we have.

      Anything further beyond that and it becomes really interesting... it might happen, but we're running out of room in the known universe.

      • Re:amazing (Score:5, Insightful)

        by itzly ( 3699663 ) on Tuesday February 24, 2015 @02:01AM (#49117313)

        Going to 5nm only helps if it is a functional product that is better than what we have.

        We still don't have the processing power of a human brain in a few pounds of silicon, running on 20 Watts. There's still a lot to do.

        • We don't? I don't know about you, but I sure can't do a billion math problems in a minute... but my Intel CPU sure can...

          I couldn't do a billion math problems in my whole life!

          Depends on how you measure processing power of course...

          • by itzly ( 3699663 )

            Depends on how you measure processing power of course...

            I was hoping this was obvious from my comment. I'm talking about the silicon chips doing the things that our brain can do, such as designing the next intel chip.

            • Re:amazing (Score:5, Interesting)

              by JanneM ( 7445 ) on Tuesday February 24, 2015 @02:54AM (#49117441) Homepage

              I'm talking about the silicon chips doing the things that our brain can do, such as designing the next intel chip.

              The major stumbling block isn't processor speed or capacity. It's that we don't know how to architect such a system in the first place.

              And if you think about it, a lot of the "smart" things we want to automate really don't need anything like human-level or human-like intelligence. A car with the smarts of a mouse would do great as an autonomous vehicle. Real mice manage to navigate around a much more difficult, unpredictable and dangerous environment, using a far more complex and tricky locomotion system, after all.

              • by itzly ( 3699663 )

                The major stumbling block isn't processor speed or capacity. It's that we don't know how to architect such a system in the first place.

                We have some ideas on how to architect such a system, but we can't try them out because of lack of good hardware. We already had ideas in the 80's to build neuronal nets, but the ideas failed because they weren't big enough. Now people have a lot more success with deep learning, mostly because they've been throwing a lot more hardware at it.

                A car with the smarts of a mouse would do great as an autonomous vehicle

                Well, we can't make an artificial mouse brain either, so that only enforces my point that there's still a lot to do.

                • Now people have a lot more success with deep learning, mostly because they've been throwing a lot more hardware at it.

                  Bullshit.

                  The success of deep learning coincided with the discovery of a novel training method, not improved hardware.

                  Why even open your mouth?

              • by msobkow ( 48369 )

                I think you underestimate how much of the design is actually done by computers and auto-routing/placement algorithms.

                • by itzly ( 3699663 )

                  Computers can help with the low-level design, yes. They can't come up with novel ideas to change the overall design.

          • by Kjella ( 173770 )

            No, but you're doing real-time 3D vision and context-sensitive pattern recognition with an amazing degree of parallelism any time you got your eyes open. Cue the "I'm blind, you insensitive clod" jokes. Do you know what the processing speed of a neuron is? Roughly 0.2 kHz, give or take a little depending on type. The Apple I from 1976 runs circles around a neuron with a 1 MHz processing speed. The difference? We have a *lot* of neurons with a *lot* of connections.

            The brain proves we can do a lot more of ext

            • We have a *lot* of neurons with a *lot* of connections.

              The second part is the important one. Neurones in the human brain have an average of 7,000 connections to other neurones. That's basically impossible to do on a silicon die, where you only have two dimensions to play with and paths can't cross - you end up needing to build very complex networks-on-chip to get anywhere close.

              • by itzly ( 3699663 )

                On the other hand, silicon is orders of magnitude faster, so you could use less hardware resources and do many things in sequence, rather than in parallel.

                • On the other other hand brains are orders of magnitude more energy efficient. I don't know if the efficiency is even related to the parallelism, asynchronicity, and ultra low "clock speed" of the brain, but it seems plausible that it is. The brain is optimized for efficiency above all else, where we have so far made the opposite trade-offs with computers.

                  We're doing that "real-time 3D vision and context-sensitive pattern recognition" with a few watts. Doing that with a bunch of GPUs would take thousands of

              • by Kjella ( 173770 )

                The second part is the important one. Neurones in the human brain have an average of 7,000 connections to other neurones. That's basically impossible to do on a silicon die, where you only have two dimensions to play with and paths can't cross - you end up needing to build very complex networks-on-chip to get anywhere close.

                We can implement that with a fairly simple grid with pass-through, say you have a grid (x,y) and (1,3) wants to pass it to (4,7), we can just pass it right to (2,3). It can do a simple compare(x=2, y=3) not for us, if x > 2 pass right else if if y > 3 pass down until we hit the right grid node. What's hairy is understanding how to program it into doing anything useful.

          • You are mistaken, every time you listen to music your brain is doing like 20k floating point ops per second, watching a ball and trying to hit is with your tennis racket involves millions of flops, on top of that balancing and moving your body requires a robot a few mega if not giga flops. Not so sure about your body, perhaps it needs less flops than a robit :)
            Your brain is by far the most powerfull computing device we right now have on the planet. Only beaten by my brain of course.
            I really doubt a peta flo

          • by xeoron ( 639412 )
            The human brain does "20 million billion calculations per second" (The Age of Spiritual Machines: When Computers Exceed Human Intelligence By Ray Kurzweil).
      • Re:amazing (Score:5, Interesting)

        by DrTJ ( 4014489 ) on Tuesday February 24, 2015 @02:51AM (#49117427)

        From Metal-Pages:

        In: $600/kg
        Ga: $220/kg

        vs

        Si: $3/kg

        The material part of the cost of the chip is likely to go up. I think however, that part today is minuscle,
        so that part of the price impact with be small. However, I do think the volume benefits to Si technology
        (50 years of development and industrial support, and with 13 gazillion Si units produced every year)
        will be very, very hard to beat with any III-V technology. There's so much new stuff to be done: defect
        density, passivation, via technology, lithography chemistry etc. The investment in III-V to reach current Si
        position will be huge and ultimately paid by the customers with higher unit prices.

      • by Aereus ( 1042228 )

        I think the bigger problem is, what happens when we reach the long-tail of process development, and demand tapers off to the point they can't fund further R&D?

        IE: Systems are "good enough" and people go from buying one every 3 years to "only when they break". That could be 10+ years.

        I suppose Intel would just follow the carrot to the next profitable market like they are pushing Atom CPUs lately?

      • by Theovon ( 109752 )

        I don't know if this'll apply to InGaAs, but for silicon, I did a projection based on ITRS numbers. As transistors shrink, they get faster. But at the same time, process variation gets worse, and that uncertainty requires wider safety margins. At what point does the increase in performance equal the increase in safety margin? 5nm.

        It's unlikely that InGaAs will suffer less in terms of random dopant fluctuation and lithographic abberations, unless it's less damaged by UV, in which case at least the lithog

      • by c ( 8461 )

        There is some debate among people if 5nm will make sense or even be reasonable to do...

        It's not a new discussion by any means. It was an old debate when people were asking whether a 100MHz bus was as fast as we could get, and 45nm was considered ridiculously small. The GHz barrier on clock speeds seemed insurmountable.

        Didn't stop anyone, did it?

        If it can be done, someone's going to try. If it can be done profitably, we'll see it on our desks or in our pockets in a few processor generations. That's just how

        • And to look around the GHz barrier *was* pretty damned insurmountable. Sure, it wasn't at exactly 1000MHz, but that particular number was always a "magical thinking" artifact of how the human brain regards numbers. We hit 1GHz back in 2000, and here we are, 15 years later and we haven't managed even managed a single order of magnitude increase in clock speed. Lets put that in proper context: 15 years earlier, in 1985, Intel had just released the 12MHz 386 with optional floating point module.

          So, from '85

    • For that, you can thank [ieee.org] IBM [extremetech.com].

      They have been at the leading edge of a number of computer technologies over the years. It's a shame that IBM has been so poor at capitalizing on them.

    • silicon - cone is for... cones.

    • Re:amazing (Score:5, Informative)

      by LWATCDR ( 28044 ) on Tuesday February 24, 2015 @08:46AM (#49118411) Homepage Journal

      Cray did it first.
      http://en.wikipedia.org/wiki/C... [wikipedia.org]
      Seymour Cray build a GaAs based computer almost 20 years ago. It actually worked but he ran out of money because of the end of the Cold War and the need for Super Computers decreased.

  • Always has been, always will be.

    • by Plammox ( 717738 )
      Yeah, just like terahertz technology.
  • InGaAs? (Score:5, Interesting)

    by serviscope_minor ( 664417 ) on Tuesday February 24, 2015 @12:38AM (#49117135) Journal

    GaAs was the future of super-fast transistors. The Cray 3 was made from GaAs.

    GaAs has a much higher electron mobility than silicon, 8,5000 versus about 1,500 for silicon. This allows for much faster switching. InGaAs has an electron mobility of 10,000 allowing even faster switching.

    But that's just electrons which are used in P channel MOSFETs. For CMOS, you also need N channel MOSFETS. The kicker is that GaAs and InGaAs have respectively lower and much lower hole mobility so the N channel FETs switch rather slower than silicon.

    CMOS is by far the only architecture. Historically it is the most power efficient since it only spends energy switching. On high speed, small scale CMOS, however, lots of power goes into the switching itself, the switching is fast enough that the devices don't really act very ideally and there's a lot of leakage.

    Perhaps at very extreme ends, other architectures can compete, power wise.

    • Is there enough minable indium on earth to allow mass production of InGaAs-based devices? The world production is just a few 100 tonnes per year.
      • by itzly ( 3699663 )

        According to Wikipedia, the natural occurrence of indium is 3 times that of silver, but current world production of indium is 40 times lower, so it is reasonable to assume that indium production can be scaled up if there's increasing demand.

        • And, apparently, it is three times as abundant as silver in the Earth's crust, so PARENT made no mistake here.

          The minerals in the mantle or core are not easily accessible, so the phrase "in the Earth's crust" needs to be observed.

          • It's a bit more complicated that that. Even if an element is somewhat abundant but evenly distributed in the earth's crust, then it's difficult to mine. It's only practical to mine something if it's concentrated in some areas. E.g. gold is rare but you can find it in macroscopic flecks or clumps that are concentrated in certain areas. If gold were not concentrated like that but was instead uniformly distributed in the crust, there'd be no economical way to mine it.

            That said, it looks like indium is concentr

    • Re:InGaAs? (Score:4, Interesting)

      by Beck_Neard ( 3612467 ) on Tuesday February 24, 2015 @03:11AM (#49117499)

      > CMOS is by far the only architecture

      No it's not. Complementarity is great, but there's no requirement for it to be MOS-based. MOS is just the best choice for silicon. There are transistors using Schottky barriers and other technologies that are far better suited to InGaAs. Five minutes of googling would have revealed this and nullified your "Score 5 Interesting" argument.

      No, the main issue with InGaAs is manufacturing difficulty and expense. You can buy InGaAs chips right now. It's just really expensive technology and not nearly as developed as silicon, both in terms of manufacturing steps and lithography tech.

    • Bandgap engineering baby. It's magical stuff, eg. strained Si to increase mobility.
  • Welcome InGaAs Valley
  • by Cafe Alpha ( 891670 ) on Tuesday February 24, 2015 @02:53AM (#49117439) Journal

    will involve making chips taller, ie various forms of 3D ICs. That would mean that we could continue to get the apparent effects of higher densities at least for a while, though we'd really just be making taller or chips or better interconnected layers, but it would also mean that the cost of transistors wouldn't go down, it would probably go up.

    • You can't just stack cpu chips on top of one another. They'd melt and vaporize. You either have to develop really good cooling tech or ways of reducing power consumption.

      One near-term solution is to stack memory (cache levels and main RAM) on the cpu chip. Memory doesn't produce that much heat so cooling would be straightforward. It would be a huge boost to speed to have memory right on top of the cpu. A few companies are working on this.

      • by afidel ( 530433 )

        GaAs chips have a very high thermal tolerance, temperatures of 250C have been shown to have no impact on MTTF, this is ~250% better than Si. The bigger issue is what do you attach them to, most commonly available PCBs can't handle that, though solutions do exist since I've read about very high temperature GaAs chips used in jet engine monitoring and control.

        • Chips that run hotter also have more thermal gradient, which can put mechanical stress on the various delicate layers of the chip. Being able to run hotter means you can support more of a thermal gradient to ambient, and thus support more heat flow and thus more computations/sec. However, at some point you're going to cause mechanical failure of the chip, especially if the stresses cycle.

          So not only termperature tolerance, but also coefficient of thermal expansion and strength of all the various materials

      • One near-term solution is to stack memory (cache levels and main RAM) on the cpu chip. Memory doesn't produce that much heat so cooling would be straightforward. It would be a huge boost to speed to have memory right on top of the cpu. A few companies are working on this.

        Another I've heard about is going vertical with the transistors. You still have increased worries about heat, but you can get a lot more density that way. Shorter average wire runs also result in less heat per transistor, on average, so increased density and efficiency might outweigh any need to throttle to manage heat.

      • by drinkypoo ( 153816 ) <drink@hyperlogos.org> on Tuesday February 24, 2015 @08:49AM (#49118431) Homepage Journal

        You can't just stack cpu chips on top of one another. They'd melt and vaporize. You either have to develop really good cooling tech or ways of reducing power consumption.

        On-chip heat pipes will become a thing to carry heat away from the center of stacks. We found out that water actually goes faster through channels so small that it has to pass one molecule at a time.

  • by Anonymous Coward on Tuesday February 24, 2015 @05:14AM (#49117785)

    > III-V semiconductor such as indium gallium arsenide (InGaAs

    I think the french will like it and possibly the swedes. They use Gallium and Indium based semiconductors in airborne electronic warfare systems, which allows for very high RF energy output in physically very small and high temperature tolerant packages. (For example used in the Dassault Rafale and SAAB Gripen fighter jets). The french SPECTRE jamming suite is especially famous: the Rafale plane is not stealthy, only has reduced radar reflection, but the french trusted their system enough so their pilots were already flying deep in lybian airspace by the time the US Navy started to launch Tomahawk cruise missiles at Gaddhafi. Supposedly there is something equal or better in the american F-35 JSF, but that airframe is so buggy one must wonder if it will ever enter service?

    On the other hand non-silicon semiconductors, like Ga and IN tend to cost twice the price of pure gold per weight or more. At the most extreme end, the soviet-russians even created diamond-based semiconductors, for use in space weapons and a planned Venus robotic rover. They invented a diamond crystal growing machine for the purpose, which after the Cold War was sold to a US company, which nowadays grows and sells multiple carat "cultured" yellow diamonds for ladyfolk decoration purposes. Beware, that femme fatale may wear a supercomputer on her finger! Now you know why multiple-finger gesture support was developed by Synaptics...

  • The prices in my condo development in Indium Gallium Arsenide Valley is going to explode!

We are Microsoft. Unix is irrelevant. Openness is futile. Prepare to be assimilated.

Working...