ARM In the Datacenter Isn't Dead Yet (theregister.co.uk) 147
prpplague writes: Despite Linus Torvald's recent claims ARM won't win in the server space, there are very specific use cases where ARM is making advances into the datacenter. One of those is for use with software-defined storage with open-source projects like CEPH. In a recent The Register article, Softiron's CTO Phil Straw states about their ARM-based CEPH appliances: "It's a totally shitty computer, but what we are trying to do here is storage, and not compute, so when you look at the IO, when you look at the buffering, when you look at the data paths, there's amazing performance -- we can approach something like a quarter of a petabyte, at 200Gbps wireline throughput." Straw claimed that, on average, SoftIron servers run 25C cooler than a comparable system powered by Xeons." So... ARM in the datacenter might be saying, "I'm not quite dead yet!"
Not "dead yet".. It has not even grown up yet. (Score:2)
The FUD is strong in this submission ...
Re: (Score:2)
Hey its in datacenters. I bet those smart temperature/environmental monitoring units are running ARM processors. ;-)
Possibly even the wireless APs.
Re: (Score:3)
That's what makes the report of these SoftIron storage nodes (and the fact that the storage nodes are accompanied by management nodes whose architecture the article doesn't specify and 'router' nodes that expose iSCSI, NFS, and SMB; architecture also unspecified) unsurprising (if you
Re: (Score:2)
There have been vendors chasing the 'high' end, but all but Marvell have bowed out. AMD has cancelled future ARM product. Qualcomm seems to have failed to bring their offering to market. Cavium got bought by Marvell. Broadcom's rumored Xeon competitor is nowhere to be seen.
So Cavium ThunderX is the only platform to get the PCIe and bunch of DIMMs. In fact the *one* benchmark Marvell can show as compelling is the memory bandwidth compared to Xeon. It also can be seen with fairly normal 'pc-like' firmwa
Re: (Score:2)
What surprises me more is that the niche that the existence of Avoton (C2000 series) and Denverton C3000 series) parts suggests exists doesn't have m
Re: (Score:1)
Let's not forget he did take that job with Transmeta. He's a really smart guy, but I don't think he's necessarily a perfect prognosticator. A lot of this hardware is in a much more usable state than it seems just because the patches are still floating around back channels and haven't been mainlined yet.
Re: Not "dead yet".. It has not even grown up yet. (Score:1)
Let's not forget he did take that job with Transmeta.
Unless you know his own particular reasons for taking that job, mentioning it fails to demonstrate a damn thing accept that he took a fucking job.
Re: Not "dead yet".. It has not even grown up yet. (Score:1)
Re: Not "dead yet".. It has not even grown up yet (Score:5, Funny)
Well, technically you'd have to say he observes, then opines (literally "to state as one's own opinion"). If all he ever did was observe then we'd never know, would we?
Re:Not "dead yet".. It has not even grown up yet. (Score:5, Informative)
it reverses the traditional storage server layout by moving CPUs to the front of the PCB and storage drives to the back. This means cool air from the fans blows over the drives first, and then the CPUs â" which wouldn't make any sense in a compute server.
Um... What?
Modern servers cool front to back. They place the drives in front where they are cooled first. Some place the CPUs behind the drives. Others place the CPUs in parallel with the drives so that they're also cooled directly from ambient air.
No one in their right mind places the drives after the CPUs. Losing a CPU is just money. Losing the drive is everything.
Were you maybe trying to say they put the fans in front of the drives instead of behind them, pushing air instead of pulling it? That doesn't make any real difference to the cooling, but it makes hot-swap harder.
If these machines actually cool back to front, that's a bad thing. Modern data centers are laid out with a hot-aisle cold-aisle design. The wiring and exhaust side (the back) faces the hot aisle. Equipment which reverses that flow effs everything up. Cisco is especially bad about this, but servers mostly get it right.
Re: (Score:1)
If these machines actually cool back to front, that's a bad thing. Modern data centers are laid out with a hot-aisle cold-aisle design. The wiring and exhaust side (the back) faces the hot aisle. Equipment which reverses that flow effs everything up. Cisco is especially bad about this, but servers mostly get it right.
They cool back to front, but they install them in reverse.
Re: (Score:2)
Re: (Score:2, Funny)
Wait for the new generation of Amiga computers :-)
Re: (Score:3)
It's the lack of custom stuff that has held ARM back. To get really high throughput you need high end NICs and storage controllers, which in turn need proprietary drivers that need porting to ARM. It's not just a case of re-compling either, or just one or two components you can bolt on. For example even now there are not many options for ARM boards with huge numbers of PCIe lanes to feed all that stuff, but if you buy a Xeon of Threadripper/Epyc system that's standard.
Re: (Score:2)
I wonder about a hypervisor standard for ARM. That way, the actual hardware can have funky SoC components, but all that stuff is abstracted away. This way, it doesn't really matter what the hardware is, there is an environment for an OS available that doesn't require the OS to deal with funky drivers in general.
Re: (Score:2)
Re: Requires changes to software (Score:2)
It's the lack of custom stuff that has held ARM back.
Right, it's not at all about economies of scale or ommoditization; its success is dependent upon smaller-volume sales of more custom shit Snicker.*
*If I I incessantly put you down, it's not because I'm an asshole (although I obviously*am*); it's because of the dumbass shit that oozes out of that hole of yours... ;)
Re: Requires changes to software (Score:2)
Re: (Score:1)
Re: (Score:2)
Re: (Score:2)
When I can compile an "ARM" kernel, and not a kernel specifically built for a board and processor, I think it'll have a lot more of a in the server market.
Nothing is generic on these things.
Every fucking chip has its own PCIe bridge.
Re: (Score:3)
The boot situation is particularly tangled. Still, the *way* the Raspberry Pi brings up the system is a step in the right direction (see the GPU being used in the same manner as the Z80 was to bring up the 6502 on the Commodore 128). Now, that's specific to one vendor (Broadcom), but maybe ARM could take a hint and formalize that a bit?
It's trivial to drop a new kernel onto one via the SD card or serial.
Re: (Score:3)
By no means is the Raspberry Pi perfect (more RAM would be nice), but it is very easy to get started working with it. It would be nice if more ARM SBC makers agreed on a chipset standard, making the SoC modules available to (and hopefully part of) the Linux mainstream kernel.
I can see a niche for SBCs designed to be desktops, or perhaps small blades for a dense enclosure, similar to a Raspberry Pi compute model, but with a non-trivial amount of RAM. This would make for a very useful VDI structure.
Linus is more nuanced ;-) (Score:3)
Re: (Score:3)
Sometimes the best engineering, the best design, or the best science, still feels to make market share because of marketing, financials, and deliberate partnerships. Betamax vs VHS. Or HD-DVD vs BluRay or even X2 vs 56KFlex. It’s not always the best design that wins. There are patents which result in recurring revenue from licensing at stake. This usually results in a whole lot of finagling and cutting deals in board rooms regardless of which technology was superior.
Re: (Score:2)
Since when do you need a dev machine with the same architecture as the server? I write ARM code all day on an x86 laptop.
The solution is obviously .NET Core. Then the CPU architecture doesn't matter. Only half kidding.
Re: (Score:3)
Re: (Score:2)
Re: (Score:2)
Is cross-compilation not working out for you? Or something else?
I've been curious about setting up arm-based servers so I'd love to know what pitfalls you've encountered.
Re: (Score:3)
Re: (Score:2)
Linus makes the mistake though of assuming that "Compile" is a step. Lots of Serverless applications are just javascript snippets. There are no compile bug concerns between platforms you just specify a .js block of code and execute. The platform is irrelevant. Serverless cloud providers don't even tell you what underlying CPU is executing your command.
There's plenty of serverless web tasks to justify ARM if it is cheaper/more efficient/has lower startup times.
AWS ARM instances (Score:2)
ARM adoption will increase because AWS offers the a1 instance family now. You can now easily fire up servers with ARM hardware to work on your software solutions. For many applications it will be a viable solution with substantial cost savings. Watch the stories and statistics that you start seeing at the summits and reinvent from customers in 2019.
LOL shitty computer? Er, no. It's an Intel killer. (Score:1)
And that's got intel scared.
Anyone can license the cores, Apple is doing it now for laptops, workstations won't be far behind once the laptops prove they have some serious oomph.
And you can dump the legacy x86 craphole shithacks that place that old cpu way of thinking, and end up with something that is actually quite good and easy to work with.
ARM will win in the datacenter for the simple reasons such as:
1) Cost of CPU is a fraction on intel/amd offerings.
2) Cost of running the CPU in heat and power is *sig
Re: LOL shitty computer? Er, no. It's an Intel kil (Score:3, Insightful)
Lack of Intel Management Engine or other spyware built-in features, that can't be removed without a high degree of risks of permanently damaging the hardware.
Re: (Score:3)
Re: (Score:2)
I have not understood why the ARM TrustZone "worlds" isn't used with a hypervisor. It would provide a very armored attack surface, preventing malware in one VM from trying to jump to another. It also would be useful for stuff the OS wants to protect (user credentials to guard against a pass the hash attack).
Re: (Score:2)
Re: LOL shitty computer? Er, no. It's an Intel kil (Score:1)
ARM can be easily scaled to hundreds of cores maybe more without having an astronomical price and and without requiring a nuclear power station sitting on the desk right behind [gaming] PC :)
Re: LOL shitty computer? Er, no. It's an Intel kil (Score:4, Interesting)
> ARM can be easily scaled to hundreds of cores
And yet, an Android phone with 8+ cores and nominal clock speed of 2GHz+ still can't render a Javascript-heavy web site (like Amazon, Walmart, or Sears) as well as a 15 year old 700MHz Pentium III.
> without having an astronomical price
Scale an ARM-based solution up to the point where it's capable of genuinely matching the performance of an i9, and you'll find that the ARM-based solution is probably quite a bit MORE expensive.
> without requiring a nuclear power station sitting on the desk
Compared to the power and cooling requirements of a Pentium IV with 15kRPM hard drive, an i9 with RTX and SSD is practically a laptop watt-wise. 20 years ago, I literally cut a hole in the wall between my computer room and the hallway so I could put my computer in the hall & pass the cables through the wall to get the heat and noise out of my face.
Re: (Score:2)
I dunno, my 8-core Galaxy S9+ seems fine with rendering websites. After all, that's what I'm posting from and do 98% of my browsing from.
As to the Pentium IV, definitely! Perhaps 7 years ago, I was given a Pentium IV tower, and I threw it in a corner as a headless media server. It only lasted the first month, because of the $40 spike in my electric bill.
Whereas my bench at home has 20 Cortex-A53 cores on it, and the kill-a-watt doesn't creep past 65 W, including 1 TB external RAID, 2 USB hubs (why not?), el
Re: (Score:2)
You need 20 A53 cores to match the performance of 4 Xeon cores.
In some workloads, this doesn't matter, because they can be scaled well.
In a lot of workloads, it simply does matter.
I administrate around 150 servers, and we run 7 datacenters.
I already avoid the slower I know a lot of armchair computing experts like to claim that it is, but I'm sorry. The reality on the ground is th
Re: (Score:2)
Re: (Score:2)
Should have been:
I already avoid the slower clocked Xeons. Aggregate performance is simply not comparable in most workloads with highly disparate per-core performance. I know a lot of armchair computing experts like to claim that it is, but I'm sorry. The reality on the ground is that it is not. That's why we're not using AMD, and we're not using ARM. Though I promise you- we look forward to being able to some day.
Re: (Score:2)
still can't render a Javascript-heavy web site (like Amazon, Walmart, or Sears) as well as a 15 year old 700MHz Pentium III.
When was the last time you used a 15 year old 700MHz Pentium III? Eight years ago?
Re: (Score:2)
About 4 years ago, I dusted off an old Compaq Armada laptop (700MHz Pentium III, 512mb ram) and tried using it with a minimalist Linux distro. The performance of Chrome or Firefox with Amazon.com and Walmart.com was SLIGHTLY better than the performance of the same two web sites with my then-new Galaxy Note 4 (all using wifi, so the mobile network quality never entered into the equation).
The Pentium III is a great reference point, because its zenith (the 1.4GHz Pentium III Xeon) marked the point when Intel t
Re: Erry day son! (Score:2)
Re: (Score:2)
Re: (Score:2)
Imagine a language with all immutable data and no side effects. It would be the purest of all languages and completely useless. You can minimize the amount of things that need to be shared, but you can never get rid of them all. I am in no way trying to have shared data.
Concurrency is not difficult, in user space. I make no claims about kernel space. I have always found all things concurrency
Re: (Score:2)
Re: (Score:3)
> 2) Cost of running the CPU in heat and power is *significantly* less than intel/amd
ONLY true when the ARM's performance is significantly less than Intel/AMD as well. Beef an ARM up to i9 specs, and it's going to burn as much power and throw off as much total heat AS an i9 with identical raw performance.
It's like LED lighting. A single LED might throw off light with just milliwatts of power... but crank it up so it's throwing off EXACTLY the same amount of light as a 100-watt halogen lightbulb (measured
Re:LOL shitty computer? Er, no. It's an Intel kill (Score:5, Interesting)
ONLY true when the ARM's performance is significantly less than Intel/AMD as well.
ARM has historically had more performance per clock than x86 and x86-64; and modern ARM chips run like 2.4GHz at a watt of peak TDP on four cores.
Think about linear character matching ("abc" in "aaabc" -> "a=a, b!=a" -> "a=a, b!=a" -> "a=a, b=b, c=c" -> match) versus Booyer-Moore ("abc" in "aabc" -> "c:a = 3" -> "c=c, b=b, a=a" -> match). Booyer-Moore finds a string--faster with longer search strings--in large amounts of text with few comparisons, thus issues fewer CPU instructions.
CPUs can implement ALUs, decoders, and pipelines to execute the same instruction code in fewer clock cycles. Just like using a different software algorithm, you can use a different hardware approach.
Prefixed instructions and fixed-length instruction sets are core to ARM. Literally every instruction is prefixed. That means where you might compare for one cycle, then jump or not jump on the next cycle, ARM simply jumps or doesn't jump. One fewer cycle.
The decoder doesn't have to deal with figuring out instruction size or the content if it picks an instruction prefixed to only execute if ZF is set, so if you SUB r2, r1 and the result is zero, the next instruction that executes only if ZF is not set is just skipped and the decoder moves on.
Because the CPU will read ahead and cache (preload) the next several instructions (fetches from RAM are slow!), it's technically-possible to block out the next e.g. 10 instructions as IFZ [INSN], and have an ARM CPU internally identify the next several instructions are prefixed IFZ and just skip the instruction pointer ahead that many. Remember: every instruction is exactly one word wide; you don't need to know what the next instruction is to know where the following instruction starts. You don't have to decode the instructions if they won't be executed.
This feature frequently eliminates a large number of comparisons and jumps, trimming down the size of the code body (you'd think variable-length insns would do that, but that usually doesn't work out). More instructions fit into cache, and branch prediction becomes simpler (less power) and more-effective.
ARM also has 30 GPRs. x86-64 has 10 GPRs, plus source/destination/base/count pointer registers that are basically GPRs. A lot happens without using RAM as an intermediate scratch pad.
It's like LED lighting. A single LED might throw off light with just milliwatts of power... but crank it up so it's throwing off EXACTLY the same amount of light as a 100-watt halogen lightbulb (measured from every direction), with color fidelity that's at least as good as that 100-watt halogen bulb (none of this "80+ CRI" shit, or even "92+ CRI with weak R9"), and it's going to CONSUME at least 70-80 watts and throw off almost as much heat AS the original incandescent bulb
Halogen and incandescent bulbs are black-body emitters: much of their light is in the infrared range. LEDs are narrow emitters and use combinations of materials to emit in multiple ranges when providing white light. That means an LED operating on 100 watts of power emits about 80 watts of visible light, while a halogen operating at 100 watts emits about 20 watts of visible light, and an incandescent tungsten-coil bulb emits about 10 watts of visible light.
An LED emitting the same broad-spectrum visible light as a 100-watt halogen would consume 25 watts of power.
Re: (Score:2)
ONLY true when the ARM's performance is significantly less than Intel/AMD as well. Beef an ARM up to i9 specs, and it's going to burn as much power and throw off as much total heat AS an i9 with identical raw performance.
It's actually worse. You can't really have high performance and low power usage at the same time, it's a trade off. CPUs have gotten better, but the transistors themselves become more leaky when you design for higher performance, among other architectural trade offs made to increase performance. And pushing transistors optimized for low power to be faster will consume more power than transistors optimized for high performance. The gap is shrinking (pun), but it's still a practical difference.
Re:LOL shitty computer? Er, no. It's an Intel kill (Score:4, Interesting)
AMD created the x86-64 architecture, and is making inroads with Epyc. AMD also has some RISC-V work in the pipeline. I'm predicting RISC-V will be big: Intel may try to capitalize on ARM thanks to mobile space, and AMD will start shoving RISC-V (no license fees) into processors for Chromebooks and the like, then into servers running Linux for RISC-V or something.
The next Raspberry Pi might be RISC-V. It's been mentioned. Nobody's taking that seriously yet, and they're not suggesting it seriously yet.
AMD beat Intel once doing this. They invented a whole new architecture and killed IA-64.
Re: (Score:2)
Re: (Score:1)
No, I say it like there is at least inter-generational and backwards compatibility. You can take any CPU and plonk it in any mainboard with a compatible socket, and provided that on-board firmware is up to date, be reasonably sure it will work.
With ARM all bets are off, in every aspect. If you think the x86/x64 world is a mess, I can tell you've never really looked into the world of ARM for sure. It's a completely different dimension.
remember that time... (Score:5, Insightful)
remember that time when everybody said intel x86 would never make it in the data center...
Re: (Score:3)
Re: (Score:2)
Chicago Stock Exchange
Uh, the label says "Chicago Mercantile Exchange"...
Re: (Score:2)
I don't remember that time nor does anyone who modded you insightful.
Re: (Score:2)
Re:remember that time... (Score:4)
Can't seem to recall any one saying that. What I do recall is there was a significant effort for everyone to have their own custom processors, PowerPC, Sparc, PA-risc, Clipper, etc etc etc. All of them eventually gave way to the x86.
Re: (Score:3)
I can. Of course the people saying it were the people pushing "PowerPC, Sparc, PA-risc, Clipper, etc etc", but yeah, I remember the notion that 80x86 wasn't proper server hardware being expressed from time to time back in the 1980's and maybe even early 1990's.
I don't believe anyone used the term "data center", though. Was the term even invented back then? "Mainframe" and "server", more like.
Re: (Score:2)
Re: (Score:1)
remember that time when everybody said intel x86 would never make it in the data center...
Hell, just use the same stupid YotLD "Linux is everywhere" argument, ARM is everywhere inside a DC already. Practically every out of band management card and HBA is an embedded ARM chip. They probably outnumber Intel chips in many DCs.
If Linux running on a bunch of cell phones is some big win for Desktop Linux then I'm sure a bunch of ARM SoCs running busybox is a win for ARM in the DC right?
Re: (Score:2)
remember that time when everybody said intel x86 would never make it in the data center...
Ooh, I member!
So tired, just need a rest (Score:2)
I remember when ARM was the cool kid. Now I guess it's just some old geezer yellin', "Get offa my greenboard!"
Linus is wrong (Score:2)
Re: (Score:2)
Re: (Score:1)
Yes, forget the link. The show isn't worth watching.
amd epyc is good for pci-e storeage nodes with 128 (Score:3)
amd epyc is good for pci-e storage nodes with the 128 pci-e lanes.
also CEPH / ZFS like lots of ram as well.
The ARM in the datacenter... (Score:4, Funny)
Immune to speculative execution (Score:1)
Another advantage is that some ARM processors aren't affected by the speculative execution vulnerabilities. In particular the ARM Cortex-A53 [raspberrypi.org], which is used in this server
http://www.socionextus.com/products/synquacer-edge-server/ [socionextus.com]
is immune to speculative execution vulnerabilities.
Re: (Score:3)
The A72 does, however, do speculative execution. And ARM chips aren't invulnerable to cache attacks, either. That said, I really like the A53 (as implemented by the BCM2837) But what got me was this about Ceph,
The ceph-deploy utility must login to a Ceph node as a user that has passwordless sudo privileges, because it needs to install software and configuration files without prompting for passwords.
Hard pass, thanks.
Not dead yet? (Score:2)
https://en.wikiquote.org/wiki/... [wikiquote.org]
How to win big in the datacenter (Score:1)
Ram.
Power supply.
Cooling.
Networking.
That can be found, that is on sale and that people like.
A CPU is wonderful.
Now make all the other parts that support the CPU. That work for servers 24/7. At a low price.
An OS would be good too. Software?
ARM well represented in Network Hardware (Score:3)
So ARM is great anytime you don't care about CPU processing power, but still want to move data -- storage appliances and network. Which is odd given that in the mobile space the few Atom x86 Android phones to reach the market had lesser raw CPU benchmarks than their ARM contemporaries at the time, yet in actual usage felt much smoother because of wider / faster buses and superior throttling (Had a Zenfone 2 with the Atom and it's still smoother than a lot of Snapdragon 6xx midrange phones).
Since we're on this subject I need ARM suggestions (Score:2)
I need suggestions for commercially made ARM systems that will work in temperature ranges from -35F to 140F (-37C to 60C) for an engineering project. These things are going to be in metal boxes on the side of Texas Highways.
Right now we've got some very impressive Intel systems, but those are in air-conditioned boxes, I'm looking for something that can survive a non-air-conditioned box. When I look for ARM stuff I find a lot of industrial boards, but not a lot of pre-made industrial systems, especially in
Re: (Score:2)
Now that top one there might be exactly what I need. I'm going to look closer and send that to the engineer and see if they'll get me one to toy with. Thank you.
The real answer is RISC (Score:2)
The real answer is RISC.
SPARC, POWER and older brother RS/6000, MIPS*, and ARM's granddaddy DEC Alpha dominated the data center space for decades. It was the cost/performance ratio of the far less efficient Intel architectures that let them win in this space.
We could easily reduce data center footprint by 1/3 by using RISC, but that's not how a free market works.
*I have installed huge SGI servers
Re: (Score:2)
Linux is great, but it does have its short-comings when you focus on sys-ad
Re: (Score:2)
Trains are dying out because of cars.
In Europe we build more and more (high speed) railway tracks.