Stealth Startup Plans Fundamentally New Kind of Computer with Circuit-Rearranging Processor (zdnet.com) 107
VCs have given nearly half a billion dollars to a stealth startup called SambaNova Systems to build "a new kind of computer to replace the typical Von Neumann machines expressed in processors from Intel and AMD, and graphics chips from Nvidia."
ZDNet reports: The last thirty years in computing, said CEO Rodrigo Liang, have been "focused on instructions and operations, in terms of what you optimize for. The next five, to ten, to twenty years, large amounts of data and how it flows through a system is really what's going to drive performance." It's not just a novel computer chip, said Liang, rather, "we are focused on the complete system," he told ZDNet. "To really provide a fundamental shift in computing, you have to obviously provide a new piece of silicon at the core, but you have to build the entire system, to integrate across several layers of hardware and software...."
[One approach to training neural networks with very little labeled data] is part of the shift of computer programming from hard-coded to differentiable, in which code is learned on the fly, commonly referred to as "software 2.0." Liang's co-founders include Stanford professor Kunle Olukotun, who says a programmable logic device similar to a field-programmable gate array could change its shape over and over to align its circuitry [to] that differentiated program, with the help of a smart compiler such as Spatial. [Spatial is "a computing language that can take programs and de-compose them into operations that can be run in parallel, for the purpose of making chips that can be 'reconfigurable,' able to change their circuitry on the fly."]
In an interview in his office last spring, Olukotun laid out a sketch of how all that might come together. In what he refers to as a "data flow," the computing paradigm is turned inside-out. Rather than stuffing a program's instructions into a fixed set of logic gates permanently etched into the processor, the processor re-arranges its circuits, perhaps every clock cycle, to variably manipulate large amounts of data that "flows" through the chip.... Today's chips execute instructions in an instruction "pipeline" that is fixed, he observed, "whereas in this reconfigurable data-flow architecture, it's not instructions that are flowing down the pipeline, it's data that's flowing down the pipeline, and the instructions are the configuration of the hardware that exists in place.
ZDNet reports: The last thirty years in computing, said CEO Rodrigo Liang, have been "focused on instructions and operations, in terms of what you optimize for. The next five, to ten, to twenty years, large amounts of data and how it flows through a system is really what's going to drive performance." It's not just a novel computer chip, said Liang, rather, "we are focused on the complete system," he told ZDNet. "To really provide a fundamental shift in computing, you have to obviously provide a new piece of silicon at the core, but you have to build the entire system, to integrate across several layers of hardware and software...."
[One approach to training neural networks with very little labeled data] is part of the shift of computer programming from hard-coded to differentiable, in which code is learned on the fly, commonly referred to as "software 2.0." Liang's co-founders include Stanford professor Kunle Olukotun, who says a programmable logic device similar to a field-programmable gate array could change its shape over and over to align its circuitry [to] that differentiated program, with the help of a smart compiler such as Spatial. [Spatial is "a computing language that can take programs and de-compose them into operations that can be run in parallel, for the purpose of making chips that can be 'reconfigurable,' able to change their circuitry on the fly."]
In an interview in his office last spring, Olukotun laid out a sketch of how all that might come together. In what he refers to as a "data flow," the computing paradigm is turned inside-out. Rather than stuffing a program's instructions into a fixed set of logic gates permanently etched into the processor, the processor re-arranges its circuits, perhaps every clock cycle, to variably manipulate large amounts of data that "flows" through the chip.... Today's chips execute instructions in an instruction "pipeline" that is fixed, he observed, "whereas in this reconfigurable data-flow architecture, it's not instructions that are flowing down the pipeline, it's data that's flowing down the pipeline, and the instructions are the configuration of the hardware that exists in place.
Yawn (Score:5, Insightful)
Smoke and mirror and faked demos until they finally produce a prototype that's nothing more than an FPGA.
Sorry investors, you're not catching this unicorn.
any Bean Counters at /. ????? (Score:2)
PRO-TIP: It takes about 12 years to get to the point where the brains of the smartest human kids can be trained to perform simple calculations involving the Fundamental Theorem of Calculus [although only about 1% of those smartest kids will have any clue what the FTC actually says].
Machine learning in hardware is an awesome
Re: (Score:2)
for that matter of course any extant processor can do what they're talking about in software. Might be some slower than flexible hard wiring but then just throw more cheap parallel off the shelf hardware at the problem if you're really bottlenecked.
Turing-Complete-ish Hardware/Compilers (Score:2)
It's very very difficult to compete with off-the-shelf Turing-Complete-ish Hardware/Compiler combos which have had decades of debugging invested in them already.
As expensive as this thing would have to be [if they could even debug 99.9% of its functionality in the next decade], it would be damned near impossible to keep up with Intel & AMD & Samsung foundries pushing out generic chips at a tiny fraction of the cost of any specialized chips.
It seems lik
the blur describes fpga.. (Score:2)
well the blurb quite literally describes a system with a control loader cpu and fpga.
that the blurb does not mention fpga is what makes it suspect. if they could have super super cheap fpga's okay, they would have something going for them. but if they just have soc+fpga at existing soc+fpga pricing then who cares.
Re: Yawn (Score:2)
Such people are easily separated from their money by others with a good story and a bit of faked enthusiasm. "I don't really get how they're going to do it, but the inventor is excited so it must be good, and I can definitely see the applications" is a typical sentence often heard.
A
Re: (Score:1)
Not "investing". It's money laundering and/or tax evasion when the bullshit is this deep.
Re: (Score:2)
Re:Yawn (Score:4, Informative)
Try "spend 1 million of dirty money and get 300,000 of clean income" that you can move around. Or funnel a bunch of money you cannot explain into vendors that you actually control. Or "lose a few millions that you write off your taxes, because you just paid them to a shell company of yours". Or "drive your company's stockholders bankrupt, because you bought cheap services at exorbitant prices from a company which profits you reap".
But why should I give you an education in money-laundering? And why do you need one? Hmm...
Re: (Score:2)
Re: (Score:2)
I assume that by 'resident of the Bahamas', the OP meant a company registered and doing business in a tax shelter. The car would not have to be in the US, either.
Re: (Score:2)
Re: (Score:3)
Re: (Score:1)
Re:Yawn (Score:4, Informative)
The motivation is that most deep learning neural networks use the hardware inefficiently. [youtube.com] You can have a bunch of Tensor processing units, but they only get 10%-50% utilization for most NN calculations. So if these guys can figure out how to make it work better, that would be a 10x speedup over current hardware.
After Google created their TPU (and let's be honest, Nvidia and their graphics cards), a lot of people have had ideas on how to make it better. So now VCs are throwing money at them, hoping something sticks. If the technology is good enough, it could end up being in every cell phone.
In every cell phone ?!?!? (Score:5, Insightful)
Why in the name of Phuck Almighty would you want an hardware-based AI in your cellphone?
So that the "You make-a me so hawny, wanna suckie suckie you big White round-eyes cock" anime pr0n can be tailored to the peccadilloes of each individual incel?
Will it cum with a $19.95 per month subscription to (((the pr0n cloud)))?
Re: In every cell phone ?!?!? (Score:2)
Re: (Score:2)
Data compression is likely to be AI-driven in the future. It can self-optimize for the content. The compressed stream consists of data and neural network parameters for the decoder.
That's just another snake oil application. Data compression needs to use every bit available, and reproduce a bit-accurate copy of the input, and neural network parameters are anything but that, and even if the net would work, it's very hard to achieve a compression better than with the current algorithms, just now you also added a bunch of overhead.
The lossless compression algorithms we have today already achieve results very close to the theoretical maximum (or minimum, as you wish). There is no silver bu
Re: (Score:2)
Re: (Score:2)
Re: (Score:2)
Re: (Score:2)
Re: (Score:1)
Re: (Score:2)
Usually when someone comes up with some clever way of optimizing a problem it rapidly becomes obsolete as more general hardware overtakes it. Being more general means high volumes and lower costs so the advantage of more expensive specialised hardware quickly disappears.
GPUs are a good example. Lots of clever ideas but in the end raw polygon pushing power won. I expect it will be the same with AI, someone will get their strong AI working and then all the effort will go into optimising for that method and ev
Re: (Score:2)
Re: (Score:2)
I know just enough about hardware multipliers and the internal FPGA fabric to suspect that nobody is ever going to program a multiplier using and-or trees. The fabric latency on a gate-efficient tree would be crazy, while a fat tree would almost certainly nuke your thermal budget.
In all likelihood, the "logic" is everything else provided by the FPGA in the vicinity of some specific hardware multiplier, which are distributed ubiquitously, like transuranic diplumonium 119 in a nano-crystalline lattice. This
Re: (Score:2)
Transmeta actually produced useful processors... But the time to market was too long at the current churn rate so they weren't cost-effective. This sort of stuff only makes sense when you can't get performance by any other means, and improvement has stalled. We're not there yet. This is even nuttier than Transmeta.
Re: (Score:2)
GPU shaders are basically the same thing, but more general: you compile your shader code, distribute it to the shaders, and then they run for a while.
This is kind of an intermediate thing between an ASIC and a
Re: (Score:2)
"So if they are lucky, their product will replace GPUs for deep-learning."
How are they going to compete with such a high-volume product, though?
Re: (Score:2)
Re: (Score:2)
GPUs are cheap because of volume, etc.
Re: (Score:2)
Re: (Score:2)
Like a comedy from the cusp of the dotcom era (Score:2)
Shooting Fish (1997) [imdb.com]
Neural computers (Score:4, Informative)
Neural Computers (NOT neural net) are _likely_ the real break out on Von Neuman machines. There are several startups bringing real silicon to the market this year in sampling, test board quatitities. How you program these is a working problem. But what is know is they have energy compute efficiencies, without exageratiion, 3 to 4 orders of magnitude better than Von Neuman machines, without losing any speed. It appears they will outperform any forseable quantum computer for decades.
Re: Neural computers (Score:2)
Already done by Cray, IBM and Fujitsu (Score:2)
Re: (Score:2)
Re: (Score:2)
One you strip away the handwaving and jargon, it sounds like they're trying to make a TPU. Like Google did. Already.
Re: (Score:2)
You could achieve most of the benefits by adding a little bit of "content addressable memory" to your conventional Von Neumann processor. This allows you to do table lookups (and therefore switch statements) in a single memory cycle instead of strolling down the table saying "is it this one?".
We were discussing this over 30 years ago. it was doable then, and its doable now.
However, unless you live in silicon valley and have access to people who are investing other people's money while takin
Re: (Score:2)
Re: Yawn (Score:1)
Re: (Score:1)
investors seem to like generation ICO as i like to call it , but i cant find any here who will give me their billions for me sayi
Old is new again (Score:5, Interesting)
Re: (Score:1)
I thought the same thing. "Deja-vu all over again."
Re: (Score:2)
Yeah, I was waiting to read about "code-morphing hardware" or some such (cf. Transmeta's code-morphing software [wikipedia.org]).
Re: (Score:2)
More like a modern version of the Xilinx XC6000 series FPGA. They had fine-grain run-time reconfiguration and standard, and may not have been far off being self reconfigurable.
Prior art (Score:2)
So... we might eventually design something equally capable to a stem cell's design.
Better late than never...
Re: (Score:2)
Re: (Score:2)
Re: (Score:2)
I don't think it's even an FPGA. When "software 2.0" people talk about "code" they really mean the parameter values in a deep learning model. So "rewriting the code" actually means changing the values of M in Mx+b.
So you create a chip that does that can do that computation massively in parallel, you structure it so that it's very fast to load the M, x, and b, and probably throw in your nonlinearity function of choice as well, and bam, you've got yourself a Software 2.0 non-vonNeumann architecture computer
Re: (Score:2)
So you create a chip that does that can do that computation massively in parallel, you structure it so that it's very fast to load the M, x, and b, and probably throw in your nonlinearity function of choice as well, and bam, you've got yourself a Software 2.0 non-vonNeumann architecture computer of the future.
Tesla's self driving computers do this. Google's TPUs do to a certain extent too. As do the bazillion ASICs people have designed to accelerate neural networks, many of which you can buy integrated into
Re: (Score:2)
Yeah, I didn't list GPUs specifically since they're not actually specialized for neural network computation (well, the newer ones are going that direction). They're pretty good at linear algebra, of course, but they're more multipurpose than a TPU or a Tesla chip.
Re: (Score:2)
Yeah, I didn't list GPUs specifically since they're not actually specialized for neural network computation (well, the newer ones are going that direction). They're pretty good at linear algebra, of course, but they're more multipurpose than a TPU or a Tesla chip.
To some extent. They've also got texture sampling units and rasterisers which are unnecessary for many neural networks. Qualcomm's NN stuff is just their VLIW+SIMD DSP. Nothing really NN specific, it's just got a lot of data parallel compute stuff
Re: (Score:2)
The "software 2.0" stuff seems to be more like the old fixed function pipelines. The hardware does some fixed computations, and you upload the "program" (your vertices) and away you go. The GPU manufacturers moved away from this model by adding programmable shaders, then programmable geometry, then decided GPGPU was a thing.
wut? (Score:1)
Re: (Score:2)
To really attract investors (Score:3)
you need to add how valuable this is to blockchain operations.
That should send the VC vultures soaring.
Re:To really attract investors (Score:5, Funny)
It leverages the synergy between blockchain and agile to deliver a new IoT user experience paradigm.
Re: (Score:1)
And stealth, you gotta have stealth, says right there in the headline. They are very stealthy. I still can't see them.
Re: To really attract investors (Score:2)
Re: (Score:2)
Bingo!
Re: (Score:1)
Re: (Score:2)
Now if they can make it quantum....
Gotta Show proof! (Score:2, Interesting)
This is likely a patent gimmick.. put together something that "sounds" like it might work and patent it. Just another example problem with our IP system.
If these guys have proof, which is not likely, then we finally have one of the pieces needed for constructing actual AI. AI has a requirement for being able to rewire itself... not just able to re-code itself. Humans, rewire, re-code, create, destroy, repair, and replicates... until machines are capable of this we cannot get started. There are likely m
Re: (Score:2)
Re: (Score:2)
>>This is likely a patent gimmick.. put together something that "sounds" like it might work and patent it. Just another example problem with our IP system.
If these guys have proof, which is not likely, then we finally have one of the pieces needed for constructing actual AI.
What proof are you talking about? I mean, they have working silicon... and there are plenty of FPGAs which "rewire" themselves the same way.
The point of this is to increase the utilization of the math units, it doesn't do any other
Re: (Score:2)
AI has a requirement for being able to rewire itself... not just able to re-code itself. ...until machines are capable of this we cannot get started.
Your statement is based off of meaningless semantics. What this article is discussing must still be a Turing Machine, and nothing more. Thus everything they are doing can be simulated purely in code. The only difference is run-time performance, and as I've stated many times over many years, performance is not the reason we do not have demonstrable true AI or self awareness / conscience. Whether an AI takes 2 days or 2 seconds to formulate a response to a human interaction or question is inconsequential in
Not fundamentally new (Score:2)
As other comments say, FPGA, Transmeta.
Further, this is how all modern computers work. We use high level programming languages with instructions which are translated into groups of smaller instructions, and then we pass those instructions to processors which decompose them into yet smaller instructions which are executed by the functional units... Often in parallel.
One day when/if we have holographic, three-dimensional optical computing blocks then reconfiguration might actually make sense. But we're still
Re: (Score:2)
To be fair though, a lot of advancements in tech come out of reexamining old ideas and seeing if modern techniques or horsepower are enough to overcome their initial limitations.
Re: (Score:2)
I too, am fluent in buzzwords (Score:2)
Not a Von Nuemann Machine (Score:5, Informative)
A Von Neumann Machine is a theorized machine that self replicates. Those don't yet exist outside of scifi.
Funny how much the meaning changes when someone gets one word wrong.
Re: (Score:2)
a theorized machine that self replicates. Those don't yet exist outside of scifi.
That pile of old PCs in my garage which seems to grow on its own begs to differ.
Re: (Score:2)
Sorry to blast your bubble.
There are two "von Neumanns"
The one who is name sake of the "von Neumann" architecture, which is every processor that has a fetch, decode, execute cycle of instructions.
The other one is the one who coined the self replicating machines.
Uh (Score:2)
Rather than stuffing a program's instructions into a fixed set of logic gates permanently etched into the processor, the processor re-arranges its circuits, perhaps every clock cycle, to variably manipulate large amounts of data that "flows" through the chip....
No matter how I read or think about this I can't see it as anything other than gibberish that sounds vaguely like new-age mysticism. At the end of the day your processor is simply going to be a state machine that has to have well-defined steps from one state to another. Even quantum computing has this requirement.
There is always the issue of how to do this more efficiently. How to get from one state to another useful state with fewer steps or consuming less power. But there ain't no magic and if there
Re: (Score:2)
It's possible. It's just not efficient. Think of it this way. Would you like a transformer plane that turns into a robot? Well maybe. Except it's not going to be as good at being a plane as a dedicated plane, or as good at being a robot as a dedicated robot.
Similarly, while I'm sympathetic to the concept, this just isn't going to as good at any one thing as dedicated hardware that does that one thing. Transformable chips are only any good if you don't know what you want to do with it.
Starbridge Systems again? (Score:2)
So it's Starbridge Systems again? I mean, that was ages ago that they supposedly had on the fly reconfiguring computers that you could shoot holes in an they would keep on ticking.
Not Really non-Von Neumann (Score:4, Informative)
I know it's a bit pedantic but it's kinda misleading to call chips that rearrange themselves (FPGAs etc) non-Von Neumann architectures. I mean the configuration of the chip is just another state of the machine not fundamentally different than cache coherence and ordering guarantees.
I mean either we should already be calling GPUs and (current multi-cored) CPUs non Von Neumann machines or call this one one as well.
Re: (Score:2)
This appears to not be as programmable as an FPGA, more like a programmable bus. That's why they talk about filtering and data flows; it would be optimized in specific ways that make it unsuitable to being a general purpose computer.
Re: (Score:2)
Yah, I understand that. I'm just objecting to the use of Von Neumann architecture to mean 'general-purpose programable computer.' Whether or not you are a Von Neumann machine is about whether the architecture unifies code and data (instructions are another type of data) and stuff like 'acts like a programmable bus' is too high level for this distinction.
I know that's how we appear to use it now so I'm just grumbling.
Xilinx XC6000 series FPGA (Score:2)
They had fine-grain run-time reconfiguration as standard.
Poe's law of investing? (Score:2)
data flow computers (Score:2)
Seriously?
This is so ancient stuff, I learnt about it in university, and that was when we still wrote 19 in front of our years.
Oh look, even WP is up to it: https://en.wikipedia.org/wiki/... [wikipedia.org]
Literally, I've always wondered why so few of these machines actually exist, so I wish them the best of luck, but /. should know better than to claim this is some new invention.
A new computer system (Score:1)
That's actually quite old (Score:5, Interesting)
Data-flow based computing actually is even older than our current digital computers. With pure data flows it is hard to manage. Therefore there have been numerous ways to mix this with the flexibility of general purpose computers.
One rather successful way are vector computers, where an instruction sets up a computing pipeline which in parallel gathers the data, feeds it through the ALUs and stores the results. Eventually RAM became the bottleneck so the advantages became less and less important, but in the 1980s and 1990s this was how most high performance computing was done.
Another way was the Transputer. You had lots of tiny little CPUs, each with their own RAM and high speed (for the time) Interfaces in between them. You could then make your computation spread out over a herd of those. One interesting aspect of it was that individual Transputers could be so cheap you could use them for embedded systems. This would have created sales volumes making them even more cheap.
The reason why those ideas don't matter as much as they used to is that RAM hasn't grown in speed as much as logic did. Also the control logic of simple CPUs isn't much of a deal any more as it used to be. A 32x32Bit hardware multiplier is easily more complex than a whole 6202. Going the FPGA route on the other hand, means that you will need to drive fairly long wires inside your chip. Charging and discharging those takes time and energy.
So I call BS on this, but of course in an age where banks don't pay interest any more investors will pump their money into anything.
Must be we are getting quite old (Score:1)
My guess is that the collective memory of the field spans a shorter time than the memory of some individuals still in the field.
Therefore, the same roads will be traveled over and over again.
Re: (Score:2)
Well "Great minds think alike", and it's not uncommon for multiple people to get the same idea independently.
Now if you found a start up, chances are that you haven't thought about what you are doing a lot. You'll usually be mostly occupied by dealing with investors.
AMD makes graphics chips too. (Score:1)
The software is the real issue (Score:1)
While this hardware sounds spiffy, I'm trying to envision the debugging process for a system whose configuration changes with every instruction.
It seems a bigger challenge than the hardware.
Maybe with Alexia Massalin and Transmetta team (Score:2)
NMot a new idea (Score:2)
I got a pitch for a very similar idea at SC2000 (yes, 20 years ago). There seemed to be a lot of very important but unanswered questions about how it was going to actually work. Notably, no sign of an actual production rollout.
The same questions seem unanswered now. Others here have already done a pretty good job asking those questions.
If you have to say it... (Score:2)
This startup "says" their chip is fundamentally new.
If you've got to say it in your press release, it's called hype.
If other people say it on their own, then you've got something.