Follow Slashdot stories on Twitter

 



Forgot your password?
typodupeerror
×
Data Storage

Backblaze Hard Drive Stats Q2 2019 (backblaze.com) 127

Backblaze: As of June 30, 2019, Backblaze had 110,640 spinning hard drives in our ever-expanding cloud storage ecosystem. Of that number, there were 1,980 boot drives and 108,660 data drives. This review looks at the Q2 2019 and lifetime hard drive failure rates of the data drive models currently in operation in our data centers.
This discussion has been archived. No new comments can be posted.

Backblaze Hard Drive Stats Q2 2019

Comments Filter:
  • More manufacturers (Score:4, Insightful)

    by CastrTroy ( 595695 ) on Tuesday August 06, 2019 @09:19AM (#59050294)

    I wish we had data for more manufacturers. They seem to use mostly Seagate and HGST (Hitachi) drives with a few Toshiba thrown in. I would like to see Western Digital as well. In my personal anecdotal experience, I've seen many more failures from Seagate compared to Western Digital. I'd love to see some real data to back up my experiences, or even prove them wrong.

    HGST was acquire by western digital, but it doesn't seem that they are using any Western Digital branded drives.

    • by Anonymous Coward

      Looks like you did not read the blog post.

      Anyway, I for one wish there were more manufacturers, period.

      • by brianwski ( 2401184 ) on Tuesday August 06, 2019 @11:20AM (#59050994) Homepage
        Disclaimer: I work at Backblaze

        > I for one wish there were more manufacturers, period.

        Internally, everybody at Backblaze very much agrees with you. It is scary to us to be down to only two or three suppliers.

        The good news is that within a few years, SSDs will be a viable alternative to hard drives for big data storage, and there seem to be more SSD manufacturers than hard drive manufacturers.
    • They actually talk about that in the article and how they just retired the last WD Drives. I am guessing they don't make them for their applications anymore.
    • by DRJlaw ( 946416 ) on Tuesday August 06, 2019 @09:58AM (#59050528)

      They seem to use mostly Seagate and HGST (Hitachi) drives with a few Toshiba thrown in. I would like to see Western Digital as well. In my personal anecdotal experience.

      If I recall correctly they cannot purchase Western Digital consumer drives in bulk, and so have not really purchased them since the Thailand and post-Thailand years where they resorted to buying and shucking retail drives.

      I know that their own blog reported in 2016 that they cannot purchase large bulk orders from HGST, so they're left with smaller bulk orders from them.

      Seagate is apparently not a problem - which I'd expect given press references (principally Anandtech) to Seagate's datacenter beta programs and datacenter marketing. Backblaze principally uses consumer drives, but apparently Seagate is willing to do it for those too.

    • by AmiMoJo ( 196126 ) on Tuesday August 06, 2019 @10:25AM (#59050660) Homepage Journal

      They consistently see higher failure rates for Seagate drives, year after year. They also consistently buy more Seagate drives year after year.

      It seems that for them the higher failure rate is tolerable because of Seagate's lower cost. Of course for someone without massive amounts of redundancy and high performance storage servers that might not be the case, especially as the cost difference is fairly marginal.

      • That's only part of it. They get a bulk discount going to Seagate, but they have stated repeatedly that they also have problems sourcing WD drives in quantities of 5000 at a time.

    • I wish we had data for more manufacturers. They seem to use mostly Seagate and HGST (Hitachi) drives with a few Toshiba thrown in. I would like to see Western Digital as well.

      They had Western Digital drives for many years. They were awful, mainly because of the flooding in Thailand, but they weren't good even before that and didn't improve very quickly afterward. Some of their Western Digital models were retired simply because so many of them failed they weren't able to keep homogeneous pods in operation.

      But there just aren't any more manufacturers. The HGST brand was eliminated last year after Chinese regulators were sufficiently bribed by Western Digital to let them start f

    • by samdu ( 114873 )

      As a counterpoint, I've seen probably five Western Digital drives give up the ghost in the last 8-10 years and not a single Seagate drive. Personally, I think the balance shifts back and forth. Seems like every manufacturer has a lemon model every now and again.

    • My anecdotal evidence is that hard drives are pretty reliable. Until a few years ago, I would consider them more reliable than SSDs, but that's more of a statement of how shitty some of the early SSDs were. I backup, but in the interest of not having downtime, I replace any drive that's doing anything important within 5 years. In the last 10 years, I've personally had only 2 drives fail before their time, one Hitachi and one Toshiba. Both gave warning of their impending doom, so data loss was minimal.

  • Would have been interesting to see SSD failure rate stats. Assuming they use any of course.

    • Re:SSD stats? (Score:4, Interesting)

      by gweihir ( 88907 ) on Tuesday August 06, 2019 @09:50AM (#59050478)

      Backblaze probably does not. As far as I understand their business model, they provide storage for backups and archiving, nothing intended to be used real-time.

      • Re:SSD stats? (Score:5, Informative)

        by brianwski ( 2401184 ) on Tuesday August 06, 2019 @11:26AM (#59051034) Homepage
        Disclaimer: I work at Backblaze.

        > Backblaze probably does not [use SSDs]

        For the big raw storage, you are correct. The hard drives still save us a ton of money over SSDs. However, we have quite a few SSDs in our billing servers, monitoring servers, front end caching servers, and in our large Cassandra clusters (used for resolving the "friendly" file names to raw hex storage ids/locations in our B2 service).

        I don't know that we have a large enough sample size of any one model to be statistically significant. I wish some open source project would startup where you ran a TINY little app on any computer you own that reported home once per day to a global database of drives so the world could get all this info collected and reported.
        • by gweihir ( 88907 )

          Thanks for the info. The interesting SSDs would probably be the caching ones and the ones used in the Cassandra clusters. I do agree that you need quite a lot to get statistical significance.

        • You mean some kind of SDD... telemetry? :)

        • Just curious - do you use "enterprise" SSDs for these applications, or are they also consumer drives?

  • by Anonymous Coward

    I take BackBlaze reports with a grain of salt, because they are using consumer grade drives which were not really designed for this 24/7 use. Although you can gain some insight into how one maker has a stronger showing then others in this severe use scenario. Seagate has never had very good results and so given all the drives are consumer grade you can draw some conclusions if you intend to run your own PC in this sort of cycle of use.

    • by Anonymous Coward

      It was found that the more expensive enterprise drives failed at the same rate as consumer drives. The only thing that enterprise drives benefited was the manufacturers bank account.

    • I disagree. In my humble opinion, if a consumer-grade drive can survive in non-ideal situations then it will be an excellent choice for normal (consumer) conditions.
      • by Shotgun ( 30919 )

        Especially if it is used in a ultra-raid, hot swappable system. And in this case, isn't it use for backup/archive purposes? So, it gets written, and quite possibly never accessed again? If that is so, I'd expect that their usage pattern resembles the typical consumer more than an active server.

      • by brianwski ( 2401184 ) on Tuesday August 06, 2019 @11:38AM (#59051126) Homepage
        Disclaimer: I work at Backblaze.

        > if a consumer-grade drive can survive in non-ideal situations then it will be an excellent choice for normal (consumer) conditions.

        I think it's funny people think we (Backblaze) abuse the consumer drives. When I open my home gaming box, it is filled with dust bunnies. Our datacenters have filtered air and no carpets, and we walk across sticky fly paper mats as we enter. When we open a storage pod after 3 years, it is absolutely dust free inside!

        We also don't run drive intensive applications. Our backup service you basically write the data ONCE, then leave it alone for years, and possibly read it back once to do a restore. (Well, we actually walk the drives from time to time recalculating all the checksums to make sure there is no bit-rot to repair from parity.)

        The one thing we do differently than consumers is we leave the drive "powered up" for 5 years continuously. Internally we debate a lot whether this PRESERVES the drives or HURTS THEM, we honestly don't know. There is some evidence that heating drives up, then cooling them off by powering them down, then heating them up again is a lot worse for the hard drive than just leaving it in a very nice, constant temperature for 5 years.

        But if a consumer at home leaves their hard drive powered up, I claim Backblaze is WAY NICER to the drives than a consumer. No dog hair or carpet fibers clogging up the cooling systems, we don't bump the table holding the drive with a chair every time we sit down. The cat isn't sleeping on the computer vents to keep warm. :-)
        • I claim Backblaze is WAY NICER to the drives than a consumer. The cat isn't sleeping on the [data center] computer vents to keep warm. :-)

          That you KNOW of....

  • by gweihir ( 88907 ) on Tuesday August 06, 2019 @09:48AM (#59050466)

    Considering that these are mechanical devices, even an AFR of 2.8% is pretty impressive. Looks like this has finally gotten to be a mature technology.

    Now, what I would really be interesting is comparable stats for SSDs.

    • by AmiMoJo ( 196126 ) on Tuesday August 06, 2019 @10:26AM (#59050668) Homepage Journal

      I'd like to see stats on how accurate SMART data is at predicting failures. I could tolerate a 2.8% failure rate if SMART gave me warning in advance 99% of the time.

      • by gweihir ( 88907 )

        The old estimates by IBM were just 50%. I have no idea whether that has gotten better. But I do not think backups have become obsolete, also because they protect against things like encryption-Trojans and user error as well.

      • My last two HDD (Seagate in QNAP NAS) failures were predicted by SMART and I was able to replace both before either failed fully.

    • I have a 5 year old Synology NAS. Populated it with all 4TB Western Digital Enterprise Hard Drives. Only 1 remains working, all have 7 others have been replaced over the past 5 years. 2 Have failed in the past 6 months.
      • by gweihir ( 88907 )

        That is pretty bad for "enterprise" disks. Maybe consumer disks are actually more reliable? The only enterprise disk I ever bought (Seagate) was DOA, never spent the extra money since then and never had any HDD failures since the old IBM "deathstar" drives.

        • Enterprise disks get the shit pounded out of them 24x7, data from immense number of servers striped across them when storage is virtualized. Your home disk mostly snoozes if you're typical user.

          • by gweihir ( 88907 )

            True. And the they should most definitely not fail that often, regardless of load. They should also most definitely not arrive DOA.

            I have a nagging suspicion that at least some "enterprise" disks are basically a scam. But I do not have numbers to back that up.

  • by SuperKendall ( 25149 ) on Tuesday August 06, 2019 @09:55AM (#59050500)

    I've been trying to put together a large RAID array for my home, as such although I want something reliable I also really would like to be sure I'm getting relatively quiet hard drives (this after getting a Seagate for a small set of drives, that was quite a bit louder in operation than other drives I had).

    Looking at other reviews it seems like a few years ago some reviewers used to include decibel levels, but I can't find anything current...

    From reading around it seems like Western Digital RED drives may be some of the quieter drives, but if anyone has any hard data I would love to see it.

    This is a RAID for redundancy, not speed, so I'm OK with a drive failure or two over time, just want something that is not distracting when it fires up.

    P.S. As one ingredient in this mix, I decided to get a longer Thunderbolt 3 cable which helps also, if anyone is thinking along similar lines.

    • Re: (Score:2, Insightful)

      by Anonymous Coward

      Once you get four or more in an array (you say large, so I'm guessing 12+), there is no quiet HDD. You will hear the combined whirr and click-clack of all the drives, and the fans which cool them. SSD is your only option for quiet, and it helps with the heat too.

      • Once you get four or more in an array (you say large, so I'm guessing 12+), there is no quiet HDD. You will hear the combined whirr and click-clack of all the drives

        I agree with what you are saying, I do have pretty realistic expectations that it's not noise free like an SSD would be.

        However like I said I did have this experience with a really loud Seagate drive - I have a four bay external case currently (non RAID) and when I bought this Seagtae drive, 7200 RPM like all the rest, it was really lots louder

        • by pnutjam ( 523990 )
          If you setup a linux raid and configure automounter (autofs), you could have the drives power down when they are not in use.
    • by tuffy ( 10202 )

      Western Digital Reds are cooler and quieter because they run at 5400 RPM. This isn't a problem for serving up files at home, so I've been using them for my personal RAIDs without any issues.

      • Thanks for the feedback, I was looking at the 7200 RPM Red Pro [amazon.com] drives, if anyone has any feedback about those in particular...

        I had read somewhere they were fairly quiet, though I cannot find where now. Even though I don't care about performance a lot I still care enough that I'd like to use 7200RPM drives over 5400 RPM.

        I think what I may do is buy one of a few different kinds of drives, plug them in and see what seems best, then mostly use those drives and use the noisier ones for offsite backup units.

        • by tuffy ( 10202 )

          After double-checking, it looks like WDs largest Reds and Red Pros have very similar specifications according to the spec sheets on WD's website.

          WD Red [westerndigital.com]

          WD Red Pro [westerndigital.com]

    • by lgw ( 121541 )

      I did this with HGST helium SAS drives, 3 years ago. Made a nice 6-drive RAID-10 array, since I did care about speed, with a hardware RAID controller. It's been rock solid, though I did need to put in a fan blowing directly on the RAID card. For HW RAID, make sure to get a super-capacitor backup on the card in case of power loss.

      Of course, if you don't care about performance at all, software RAID 5 is about half the price, and if you're OK with consumer drives, half again. But I'd stay away from Weste

      • by lgw ( 121541 )

        Oh, and until I added the fan in the side of the case, the server was completely inaudible most of the time - you could just barely hear drive activity in a dead quiet room, but not with anything else going on.

    • by pnutjam ( 523990 )
      I've heard they get louder over 4TB, but I couldn't say. I recently switched from one loud 3TB drive with a backup 1TB WD Red drive to 2 x 4TB WD red's. I run one as my primary with xfs and it rsyncs to the 2nd which has btrfs. Btrfs takes a snapshot and then I rsync the important stuff offsite. This "server" is in my living room running on an old desktop and it's very silent.
    • From reading around it seems like Western Digital RED drives may be some of the quieter drives, but if anyone has any hard data I would love to see it.

      I can't provide data, I can only provide an anecdote from someone who is very noise sensitive. I had 2 arrays, one with 4x 1.5TB seagate drives and the other with 4x 4TB WD drives. Two of the seagate drives started throwing SMART errors so I looked at replacing the entire array.

      I ended up going WD drives *because* that array was significantly quieter. Now the caveat there is I was comparing apples to oranges (1.5TB drives vs 4TB drives) purchased over a year apart, but in general I was happier with the WD p

  • by symes ( 835608 )

    The post provides data on the total number of failures over a given period. There is some data on average age, but I think it would be much more interesting to look at failure over time in something like a survival curve. How the failures are distributed over time.

  • by ewhac ( 5844 ) on Tuesday August 06, 2019 @05:46PM (#59053892) Homepage Journal
    Since we seem to have some Backblaze people chiming in:

    I recently-ish built a new NAS with six 4TB HGST "CoolSpin" drives in a single RAID-Z2 vdev -- HGST because reliable; CoolSpin because quiet, lower power, and fast enough for GigE.

    However, when I was selecting drives, the most annoying aspect was concretely determining whether a given drive supported TLER [wikipedia.org]. The datasheets almost never disclose this, and you have to dive into various support and discussion fora to find out.

    Now, the number of variables I need to keep track of has increased, and manufacturers seem even less interested in documenting them. Specifically:

    • Unrecoverable read error rate: This used to regularly appear in datasheets, but is now disappearing. For drives that still talk about it, this seems to have not budged from 10**14 bits, which is about 10TB, even on drives which are 10TB in size. Statistically speaking, this means you could read a brand new 10TB drive fresh out of the anti-static bag front to back and you will take an unrecoverable read error. Only a handful of "enterprise" drives quote 10**15 bits.
    • Recording Method (shingled vs. perpendicular): Shingled drives have performance implications for highly active volumes. Yet I can almost never discover whether a given drive uses it. My current rule of thumb is to assume that everything 8TB and larger is SMR.
    • TB/Year workload rating and "archival" drives: Where the hell did this come from? Is this an indication of lower quality mechanics/electronics, or just gratuitous market segmentation?

    I wonder if any Backblazers would be willing to say a few words (informally, of course) on how they rate the importance of these characteristics, and how they discover what a given drive supports without having to buy a pallet of them first.

The sooner all the animals are extinct, the sooner we'll find their money. - Ed Bluestone

Working...