eBay Deploys 100TB of SSDs, Cuts Rackspace By Half 197
Lucas123 writes "eBay's QA division was facing mounting performance issues related to its exponential growth of virtual servers, so instead of purchasing more 15k rpm Fibre Channel drives, the company began migrating over to a pure SSD environment. eBay said half of its 4,000 VMs are now attached to SSDs. The changeout has improved the time it takes the online site to deploy a VM from 45 minutes to 5 minutes and had a tremendous impact on its rack space requirements. 'One rack [of SSD storage] is equal to eight or nine racks of something else,' said Michael Craft, eBay's manager of QA Systems Administration."
depends if you are IO bound or need storage (Score:3)
78% power savings - that's pretty awesome too (Score:2)
Might even pay for itself by the years end
Barring any major catastrophes - expect to see may companies with server farms to go this route soon
Re: (Score:2)
Re: (Score:2)
Re: (Score:2)
So even with our generous 10 watts saved, and assuming you can get away with this budget drives, You are still looking at over 6 years to break even.
Three to four years should break you even, since pretty much 100% of the power input for a spinning disk drive is turned into heat which must be removed by costly air conditioning.
Re: (Score:2)
Not to mention the massive savings on rack space, which can get very expensive per square foot.
Chances are, their return is a year or two once you factor everything in.
Re: (Score:2)
Re: (Score:2)
If one drive has enough storage capacity for all your data, then one drive has enough capacity for backups too.. I've just taken to using HDDs for backup now. No expensive tape drive necessary, and hard drives must easily be cheaper per GB by now too..
Re: (Score:3, Insightful)
Re: (Score:2)
Remember, RAID is used to protect against drive failure. Dual Raptors in RAID0 is just asking for trouble - you're doubling the chance of failure. They'd be much better off with dual SSDs in RAID1 - much better performance and power consumption and better protection against sudden failure.
I'm worried about your customers that "don't have time" to perform backups. What do
Re: (Score:2)
I'm worried about your customers that "don't have time" to perform backups. What do they do when their laptop is stolen? Maybe their work and thus their data doesn't have much value.
Or they do everything, including storage, online. I'll publicly admit I don't backup my work laptop... The only reason I have it is to SSH, and I'm not backing up gigs of stuff when all I basically need is "putty", which I already have on a flash drive, and is widely available on the internet.
Customs wants to search my laptop, OK, its basically a vanilla install with putty. And a lot of empty space.
Re: (Score:2)
I tend to think of hard drives as temporary storage - if the data is important, have it in multiple physical places.
Re: (Score:2)
In a workstation scenario, couldn't you RAID1 an SSD and a fast HD in case the SSD dies like hairyfeet's gamer customers?
Would the SSD be slowed down by the HD trying to keep up or would the RAID controller's Cache prevent that?
Re: (Score:2)
Write performnace would be limited by the slower mechanical drive since all writes must go to both drives.
Read performance would depend entirely on how well the RAID distributed reads between the drives.
Re: (Score:2)
Rather than using RAID, you could use the SSD as a single drive and have differentials/backups copied across to the HDD every couple of minutes or so - that'd provide some protection against disk failure without overly penalising the speed.
Re: (Score:2)
I'd rather a drive fail suddenly without warning if the alternative is "slowly and silently corrupting data until somebody notices it because it tripped a SMART threshold". Because I've had that happen, and all sorts of other nasty gradual failures.
Any enterprise setup is going to have redundancy built-in. If a drive up and fails suddenly and without warning, you swap it out, boom, done. If you've got hot or cold standby drives already in the array, icing on the cake. But if a drive starts a slow march towa
Re: (Score:2)
Re: (Score:2)
Re:depends if you are IO bound or need storage (Score:4, Insightful)
Actually the vast majority of websites, even business websites, are really low traffic and they benefit far more from storage space (especially when shared with other sites) than speed. Operating system RAM caching will usually make up any performance deficit those kinds of low traffic sites may experience, where the majority of the traffic that does go to their sites tends to be read only and directed at only a few pages on any given day. Premature optimisation adds either (or both) complexity and expense, and is unnecessary for 90+% of the web.
Scalability is a nice problem to have, and the majority of websites would do an awful lot better if they worried about driving traffic more than they worried about scalability.
Re: (Score:2)
Having finished an upgrade to SSD's for our website and esp or DB servers several months ago. One of the things I found that has really made a difference, is in the ability of RAID controllers to use and understand a mix of SSD and HDD drives. LSI makes a nice controller that will move frequently used files to the SSD drives. We didn't go this route however. We found that we could replace 14 raid 10 HDD drives with 4 raid 10 SDD drives. The power saving are amazing, the heat differences are great, and the
Re: (Score:2)
Any business dealing on the web (aka hosting) would benefit from ssd
Because SSD is web scale.
SSDs are faster in all respects. (Score:2)
you have to use ssds for ALL reads, but, hdds for all writes. since ssd life is still limited with number of write operations conducted.
Re: (Score:2)
For SLC, which is what you'd find in enterprise, the number of erase operations (which is what has a lifespan, not writes) is a non-concern.
Re: (Score:2)
And a 3.5" 15K SAS drive will do 120MB/s to 200 MB/s with an average of 170 MB/s, which is only very slightly slower than an SSD on a 3Gbps SAS or SATA2 connection will do. A very large 7200RPM drive comes pretty close too. Sequential read and write speeds is not a major factor in SSD performance advantages. That's not even a typical use case. It's when IOPS get involved that SSDs completely destroy HDDs.
Uhh.. cost? (Score:3)
Of course everyone would love to replace all of their storage with SSD if price was no object.
The closest they come to mentioning cost is:
Though SSD is typically a magnitude of order more expensive than hard disk drive storage, Craft said the Nimbus arrays were "on par" with the cost of his previous storage, which a Nimbus spokesman said was from NetApp and HP 3PAR. (Craft declined to identify the vendors).
So, cost of new SSDs was similar to whatever HDDs they bought years ago? Yeah, that's kind of how it goes...
Nimbus prices its product on a per-terabyte basis - it charges $10,000 per usable terabyte
$10,000 per terabyte. Ok, then. Sure, it's faster, if you are willing and able to pay 10x the cost of *current* HDD-based systems...
Re:Uhh.. cost? (Score:5, Informative)
Depends on your workload. (Disclosure: I work in storage for a living.)
Sometimes, what you need is raw, bulk storage. There are two serious contenders in this space: tape, and disk. You use tape if you have a lot of data you need to store, but not much that you need to access regularly: less power, and it scales to near infinite levels of storage (at the cost of very slow access for a given piece of data.) Or you use disk if you need to access most of it reasonably regularly. SSDs are not, and never will be, a contender in this space - you're paying through the nose on a per GB basis.
On the other hand, sometimes what you need is IOs per second. Database administrators are very familiar with this - you need a bit of data from over here, and a bit of data from over there, and maybe a little bit more from somewhere in the middle, and you need it five minutes ago. Traditionally, you got this performance by building a large array across many spindles, and giving up half, three quarters, or even more of your disk space, in return for that nice, fast section of disk on the outside of the platter. Lots of running hard drives, drawing lots of power, generating lots of heat, and costing a lot of money for storage space that isn't even half used - because if you throw something else on that empty space, you completely ruin the throughput of your major production database.
In that latter space, SSD is king. Sure, it's more expensive on a dollars per GB basis, but hey, guess what - GB isn't the important metric here. You figure out which bit of data is being hammered, and you move it across to the SSD. Rather like profiling an application: pick the function that takes 90% of the time in the software, optimise the wazoo out of it, and you get a significant improvement (rather than picking something at random and optimising it to billy-oh, and getting not much return for your investment.)
So yeah - SSDs aren't going to compete in raw capacity any time soon. But in random I/O performance, they make a hell of a lot of sense. In some respects, yes, they most definitely are cheaper than traditional platters of spinning rust - see the aforementioned massive RAID set across dozens of spindles.
VMs are in the IO category for sure (Score:5, Informative)
You discover modern hardware does virtualization real well. You get a good host software, like vSphere or something on new hardware and you have extremely near native speeds. The CPUs handle almost everything just like it was running as the host OS, and sharing the CPU resources works great. Memory is likewise real good, in fact VMs only use what they need at the time so they can have a higher memory limit collectively than the system RAM and share, so long as they don't all try to use it all at once.
You really do have a situation where you can divide down a system pretty evenly and lose nothing. Let's say you had an app that used 2 cores to the max all the time and 3GB of RAM. You'd find that it would run more or less just as well on VM server with 4 cores and 8GB of RAM, half assigned to each of two VMs, as it would on two 2 core 4GB RAM boxes. ...Right up until you talk storage, then everything falls down. You have two VMs heavily access one regular magnetic drive at the same time and performance doesn't drop in half, it goes way below that. The drive is not good at random access and that is what it gets with two VMs hitting it at the same time, even if their individual operations are rather sequential.
It is a bottleneck that can really keep things from scaling like the other hardware can handle.
At work I use VMs to maintain images for instructional labs (since they all have different, custom requirements). When I'm doing install work on multiple labs, I always do it sequentially. I have plenty of CPU, a hyper-threaded 4 core i7, plenty of RAM, 8GB, there's no reason I can't load up multiples. However since they all live on the same magnetic disk, it is slower to do the installs in parallel than sequential.
If I had an SSD, I'd fire up probably 3-4 at once and have them all do installs at the same time, as it would be faster.
Re: (Score:2)
Re: (Score:2)
If I had a mod point, I'd mod this response upwards.
The parent to it was considering only capacity. And from a capacity standpoint, no SSD doesn't make any sense and is about 10x the price. But for performance SSD is king and yes, cheaper than the spinning rust.
Re: (Score:3)
You figure out which bit of data is being hammered, and you move it across to the SSD. Rather like profiling an application: pick the function that takes 90% of the time in the software, optimise the wazoo out of it, and you get a significant improvement (rather than picking something at random and optimising it to billy-oh, and getting not much return for your investment.)
Or you do what eBay is apparently doing and say, screw it, we're doing 5 blades, and throw all of your storage on SSD,
Re: (Score:2)
SSDs are not, and never will be, a contender in this space - you're paying through the nose on a per GB basis.
This is one of those predictions guaranteed to fail. Eventually it will be cheaper to produce SSDs than disks with a bunch of moving parts. How eventually? Too eventually for my tastes... but someday.
Re: (Score:2)
I have to agree. There is even the possibility that with only 1 hard drive manufacturer left in the game with a limited R & D budget that gains in storage capacity on hard drives will slow down. Not to mention limited production runs will increase the cost of the media.
Whereas SSD's will become more commonplace, the cost of production will fall and capacities will rise. As long as some form of optical media does not come along and make SSD's obsolete. SSD's have a bright future in front of them.
Re: (Score:2)
Again, depends if you need bulk storage, or fast i/o. If, to get the IO throughput required, you need to purchase far more hard drive spindles than you otherwise would need for the capacity required, then the total amount of storage you get being less with SSD may not be an issue (so long as it is "enough"). e.g. (fake numbers), if you need say 100k iops and 1tb of space, this could perhaps be done with 10 magnetic disks, or 2 SSDs. the additional space provided by the magnetic disk is of no use if your
Re: (Score:2)
$10,000 per terabyte. Ok, then. Sure, it's faster, if you are willing and able to pay 10x the cost of *current* HDD-based systems...
There will always be applications where the high dollar [latest and greatest] solution will provide *vastly* better performance than the current 'standard'.
If you're replacing 8 or 9 TB of 15K drives with 1 TB of SSDs, then the accounting becomes a bit more manageable.
And raw price:performance doesn't always tell the whole story, not when paying more can save in other ways.
Re: (Score:2)
Re: (Score:2)
What are the failure rates? That seems rather relevant when youre dealing with massive RAID arrays.
Re: (Score:2)
I'm not so sure, it all depends on moore's law and how well we can keep up with it. Essentially, SSDs should go the way processors have - doubling in capacity every 18months or so.
For the sake of argument, let's say that a 120Gb SSD is $200, the same $200 in 18months time should, theoretically, get you 240Gb. In 3 years time, it should be 480Gb, 4.5 years would be nearly 1TB and in 6 years, 2TB SSDs shouldn't be that unthinkable, for the same $200 you'd spend today. It was only couple of years ago that a 2
Re: (Score:2)
Re: (Score:2)
$10,000 per terabyte. Ok, then. Sure, it's faster, if you are willing and able to pay 10x the cost of *current* HDD-based systems...
I guess you missed the part where they said that the Nimbus pricing of $10,000 was "on par" with the HDD-based storage arrays they had from NetApp and HP before? Fibre-channel HDD pricing is in the same ballparks as enterprise SSDs.
Re: (Score:2)
Re: (Score:3)
x10?
Compact (Score:3)
Re: (Score:2)
Zero details (Score:4, Informative)
TFA reads like a thinly-veiled promo for Nimbus Data Systems, which I can only guess are pushing a Linux-based SAN appliance full of SSDs. Big whoop.
What I would love to know is: Why does eBay need 4000 VMs ?
Re: (Score:3)
They need 4000 VM's so they can buy 100TB of SSDs. :)
Re:Zero details (Score:4, Funny)
"We're not a storage team. We're Windows administrators who got into virtualization ... "
Re: (Score:2)
4000VMs actually sounds interestingly low if its a global total.
Article Heading :-? (Score:2)
I got the impression ebay just terminated a hosting arrangement with Rackspace (the company) -- bringing it inhouse, and cutting Rackspace's revenues in half. :)
Re: (Score:2)
That is also what I read!
Re: (Score:2)
<AOL>Me to</AOL>
Rackspace = company that sells VMs
rack space or maybe even rackspace = space you have available in your racks.
deployment time nine times better? (Score:2)
has improved the time it takes to deploy a VM from 45 minutes to 5 minutes
uh, any logical explanation for this? SSDs are snappier and the peak I/O can be faster compared to spindle drives - but not by factor 9, or?
Re: (Score:2)
At a guess comparing unformatted hard disk drives to SSD, sneaky.
Re: (Score:2)
Yes, in the real world, data isn't laid out on the disk in the exact order it's going to be read. Especially when filesystem structures are involved.
Seeks are very expensive. I measured a desktop drive barely doing 300 KB/s when doing random reads. That's of course not very realistic either, but the real world performance is going to be somewhere in between that and the ideal contiguous read speed. Having a disk capable of 100MB/s managing only 5MB/s is quite possible.
Reliability? (Score:3)
I read a blog-post a while back stating that SSDs fail a lot more then you would expect. Somewhere around a year of heavy use seems to take most of the life out of a consumer grade ssd. Now i wonder how putting SSDs into Raid 5 (or 6, or whatever) will behave. If a certain model of SSD croaks around X write ops, then i think the nature of Raid will mean that your entire array of SSDs will go bad pretty closely together. It must suck to have two more drives go belly up while rebuilding your array after the first drive failure.
Perhaps it would make sense to stagger SSDs in different phases of their lifetime to keep simultanious failures at bay, use some burned in drives and some fresh ones.
Re: (Score:2)
That's okay. I found a blog written by someone from the future, so I sent him a request to fetch some SSD durability data from 15 years into the future.
I expect a reply any minute now.
Re: (Score:2)
It really depends on the write cycles, the problem there is that you have about 3000 write cycles on a modern consumer grade ssd, combine that with the fact that most people buy really small ones to save some bucks and they start to use it heavily by swapping on them etc... and you have a situation where you very likely can reach those 3000s within a year or two.
If you use them in a sane manner and have a decent size so that wear levelling can do its magic you should not hit that limit within the lifespan o
Re: (Score:2)
i fully realize that consumer grade SSD != enterprise SSD, but a tighter grouping of failures is still very likely due to the way SSDs work. I for one am curious to see what effect this will have on RAID strategies in the future. As for usage, people using their SSD for swap are either stupid, or like the speedup that much that they are willing to eat the cost of an SSD over a years time, but in enterprise land those drives can see very heavy use as well, especially when logging/databases are involved.
Also,
Re: (Score:2)
Well they are just replacing their VM servers, the databases are possibly elsewhere on the network so the writes to the SSD in that scenario should only occur when they update a VM. (Just guessing though).
Still, I take your point of a series of SSDs used for the same purpose are more likely to fail around the same time than ye olde HDDs.
Re: (Score:2)
Perhaps it would make sense to stagger SSDs in different phases of their lifetime to keep simultanious failures at bay, use some burned in drives and some fresh ones.
Trust me, as I guy who's run raid arrays of spinning rust for well over a decade, you REALLY need to do that with old fashioned drives too.
Worst experience in the world is having a RAID-1 with two consecutive serial number drives and both bearings let go at the same time.
Re: (Score:2)
http://www.codinghorror.com/blog/2011/05/the-hot-crazy-solid-state-drive-scale.html [codinghorror.com]
There, now read past the first line of my post and consider the point i was making about SSD failures spreadage Vs HDD failure spreadage and its effect on RAID strategies.
Performance or Price? (Score:2)
Is he talking about performance or price. I can imagine that a single rack of enterprise SSD's could easily cost the same as 9 racks of anything else.
Re: (Score:2)
Return On Investment (Score:4, Insightful)
Re: (Score:2)
And don't forget that the SSD solution uses less electricity. That means you're not only saving on the electricity, but you're also spending less money removing the heat associated with using electricity. Being able to rebuild a server in five minutes as opposed to forty five minutes probably means they can keep more backup servers slack and turn them on only in case of emergency. Since SSD use less electricity on standby, they probably can keep the SSD idling. And if it's so much smaller, you can pack mor
Yes! (Score:2)
Finally we are getting a chance of seeing real reliability stats of the SSD!
That if eBay would be kind enough to publish the data couple of years later.
Re: (Score:2)
I do not care about ROI of eBay - I want to see how SSD work in real-life compared to HDD.
As one never gets any meaningful information from vendors, the only hope is the feedback of the users. The user in the case is the eBay.
This is just great news! (Score:2)
Forget it (Score:2)
only for reads. (Score:2)
but read operations are unlimited. so, if you are going to just read files from a hard disk, ssd makes the perfect candidate. in random reads, they are approx 40 times faster than best hdd at the minimum.
so, you just put 250 ssd disks, put your VM images on it, and, as the article says, it boosts your vm deployment time to other systems from 45 minutes to 5 minutes - there is nothing
SSDs are still unreliable (Score:2)
Re: (Score:2)
Deduplication + SSD (Score:2)
They sweet-spot in enterprise storage is doing deduplication with SSD with both direct attached and networked storage, plus 15k 2.5" and 7.2K 3.5" disk for the rest. Deduplication saves a lot space and with SSD it works like cache, especially in scaled out environments.
Re:Is this a Slashvertisement? (Score:4, Informative)
essentially because SSD has far better IOPs, you need less units to get the same speed. So you can cut the size of the storage array in half. Ebay are clearly io bound rather than needing huge storage space. So for them, its a win.
For others, maybe not so much.
Re: (Score:2)
This only works out if you have lots of wasted space in your array.
This still comes down to why you would have used SSD 10 years ago before most slashbots ever even heard of the tech: You simply need the IOPS regardless of cost.
I could see eBay having this requirement. The rest of us not so much.
Re: (Score:3, Funny)
I don't know, but I would be checking ebay for a butt-load of cheap 15k fiber channel drives for sale there.
Re: (Score:2)
It's not significant at all, they're moving from FibreChannel drives which are notoriously small and expensive to SSDs. My last employer had a FibreChannel disk array. If I remember the prices, we were faced with a choice like 240 GB FibreChannel for $1000 each or 2 TB IDE drives for $100 each. It was obvious that moving to anything that wasn't FibreChannel was a good idea, because for the same price we could get either 480 GB of FibreChannel or 40 TB of IDE.
Re: (Score:2)
failure rates (Score:2)
From the paper "Disk failures in the real world: What does an MTTF of 1,000,000 hours mean to you?" [usenix.org] by Bianca Schroeder Garth A. Gibson
In our data sets, the replacement rates of SATA disks are not worse than the replacement rates of SCSI or FC disks.
So the annual failure rates are apparently similar, regardless of vendor claims
Re: (Score:2)
Yeah. Had forgotten about that part.
"Enterprise" drives tend to be small and expensive to begin with. Moving from one variant of this to another is probably not such a big change after all.
Re: (Score:2)
Because you won't get the IOPs or speeds you get with SSDs? SAS driver are still the traditional drives, so the random access is a pain.
Re: (Score:2)
Because you're talking about random access rates of a few hundred mess per second vs 10 mess per second if you're lucky seeking all over on a platter. The number of I/O ops an SSD can get through compared to even a fast HDD is insane.
Re: (Score:2)
Because for small random I/O operations, SSDs are about 100x as fast as mechanical drives, but nowhere near 100x as expensive. This is not too dissimilar to moving from spools of tape to hard drives. Tape still has better capacity per dollar than disk, but you don't see anybody booting their PC off tape, do you?
You don't have to be a mega-corp for the speed advantage to make sense, which is why usage is increasing across the board.
For example, I recently installed a relatively small terminal server farm, an
Re: (Score:2)
Just had a flashback of my TRS-80 that I had as a kid. Pressing play on the tape drive to listen to the data so I could line up on a program to load. Cassett tapes were expensive so you put lot
LOAD "*", 8, 1 (Score:2)
'nuff said. [everything2.com]
Re: (Score:2)
e2... How quaint.
Re: (Score:2)
Re: (Score:2)
No, not really.
Cost will always be an issue. There has also been a movement in the industry to support "larger and cheaper" storage. The idea is to have a "tiered" system where you can put your stuff on storage arrays with variable costs depending on your requirements.
The "archive" use case is actually a pretty big one.
Re: (Score:2)
Re: (Score:2)
The thing is, the cost-per-gig will eventually not matter, because the cost-for-enough-gigs will be sufficiently low. The point will come where people can get an SSD that's "big enough" at a price that's low enough. It won't matter that you'll be able to buy a 5TB drive for the price of your 1TB SSD, because the 1TB SSD will be big enough for the average person.
Re: (Score:2)
Re: (Score:2)
Even for home use, I figure nearly everybody scoffing at SSDs, thinking they're not needed, has never *USED* an SSD for their day-to-day computing. It's the kind of thing where you think "Nobody needs that", and then you try it, and it turns out the be the single biggest performance upgrade you can make to your computer. And then you get used to it, and think "It's not that big a deal, HDDs weren't so bad", and then if somebody takes away your SSD, your computer feels like molasses.
A while ago, I built a ne
Re: (Score:2)
Small SAS drives have really small platters, which means that their seek time is somewhere around quarter to half the seek time of a cheap drive. This means they get 2-4 times the number of seeks per second, which means 2-4 times the number of I/O operations per second (IOOPS). For a site like eBay, where people browse fairly randomly over their entire site, the IOOPS number is the most important. A cheap SSD gets 100 times as many as a cheap hard disk. More expensive ones get 1000 times as many. Unles
Re: (Score:2)
I'm slightly surprised that they didn't go with something like ZFS's L2ARC, with maybe 20TB of flash and 100TB of slower disks.
So do you suggest they go with Oracle, which is fucking evil, or with FreeBSD, which implies a whole FreeBSD support infrastructure inside your organization?
a big flash cache would give them almost the same performance at a lower price.
ARRRRRRGH I WANT DM-CACHE [fiu.edu] NOW ARRRRRGH. Urgh. Sorry, involuntary reaction.
Re: (Score:2)
Re: (Score:2)
I wouldn't be at all surprised if they just buy in bulk from a manufacturer and re-brand them, like Dell do with their drives. The last lot of Dell 2.5" SAS disks I looked at were just Seagate drives with a Dell sticker on them instead of the Seagate one (they didn't even change the model number).
Re: (Score:3)
Re: (Score:2)
Why does latency matter for them? They have large amounts of relatively small files (small images, item descriptions and so forth). These files are spread across the disks. Each file requested means the latency is counted. If a webpage needs 100 files the latency of the disk is multiplied by 100. Each user (and there are many at any given moment) n
Re: (Score:2)
i was under the (misguided?) impression that ssd's weren't, as yet, enterprise ready in terms of reliability?
People rationalize what they want. Can you believe there are people that claim Windows is enterprise ready, merely because they want to / have to use it?
Re: (Score:2)
Re: (Score:2)