Want to read Slashdot from your mobile device? Point it at m.slashdot.org and keep reading!

 



Forgot your password?
typodupeerror
×
Data Storage Hardware

Are RAID Controllers the Next Data Center Bottleneck? 171

storagedude writes "This article suggests that most RAID controllers are completely unprepared for solid state drives and parallel file systems, all but guaranteeing another I/O bottleneck in data centers and another round of fixes and upgrades. What's more, some unnamed RAID vendors don't seem to even want to hear about the problem. Quoting: 'Common wisdom has held until now that I/O is random. This may have been true for many applications and file system allocation methodologies in the recent past, but with new file system allocation methods, pNFS and most importantly SSDs, the world as we know it is changing fast. RAID storage vendors who say that IOPS are all that matters for their controllers will be wrong within the next 18 months, if they aren't already.'"
This discussion has been archived. No new comments can be posted.

Are RAID Controllers the Next Data Center Bottleneck?

Comments Filter:
  • with things like Haadop and cloudstore, pNFS, Lustre, and others storage will be distributed. There will no longer be the huge EMC, Netapp, Hitachi etc central storage devices. There's no reason to pay big bucks for a giant single point of failure when you can use the Linus method of upload to the internet and let it get mirrored around the world. (In a much more localized manor.)

    • Re: (Score:3, Insightful)

      by bschorr ( 1316501 )

      That's fine for some things but I really don't want my confidential client work-product mirrored around the world. Despite all the cloud hype there is still a subset of data that I really do NOT want to let outside my corporate walls.

      • Re: (Score:3, Informative)

        by Ex-MislTech ( 557759 )

        This is correct, there are laws on the books in most countries that prohibit the exposure of medical and other data
        to risk by putting it out in the open. Some have even moved to private virtual circuits, and the SAN's with fast
        access via solid state storage of active files works fine, and it moves less accessed data to drive storage,
        but none the less quite fast and SAS technology is faster than SCSI tech in throughput.

        • Re: (Score:3, Informative)

          by Ex-MislTech ( 557759 )

          An example of SAS throughput pushing out 6 Gbps.

          http://www.pmc-sierra.com/sas6g/performance.php [pmc-sierra.com]

        • Re: (Score:3, Informative)

          by lgw ( 121541 )

          SAS technology is faster than SCSI tech in throughput

          "SCSI" does not mean "parallel cable"!

          Sorry, pet peev, but obviously Serial Attached SCSI [wikipedia.org] (SAS) is SCSI. All Fibre Channel storage speaks SCSI (the command set) all USB storage too. And iSCSI? Take a wild guess. Solid state hard drives that plug directly into PCIe slots with no other data bus? Still SCSI command set. Fast SATA drives? The high end ones often have a SATA-to-SCSI bridge chip in front of SCSI internals (and SAS can use SATA cabling anyhow these days).

          Pardon me, I'll just be over here

          • by sjames ( 1099 )

            SCSI started life as a command set AND a physical signaling specification. The physical has evolved several times, but until recently was easily recognizable as a natural evolution of the original parallel SCSI. At the cost of a performance degradation and additional limitations (such as nimber of devices), the generations of scsi have interoperated with simple adapters.

            SaS uses the same command set, but the physical is a radical departure (that is, it bears no resemblance) from the original SCSI and it's d

            • by lgw ( 121541 )

              I remember the days when people reading Slashdot wanted to use precise terminology about technology - don't you? Sure you do. But go on with your "Serial attached SCSI drives are not SCSI" and your "I double-clicked on the the internet, but it's broken" and so on. Those of us who are still nerds will pendanticly point out that all these storage technologies are "really SCSI drives, if you look closely" and we'll be right. Grumble grumble grumble.

            • So, no, SaS drives are not obviously SCSI any more than a USB drive is obviously SCSI.

              In my world the T10 Technical Committee defines SCSI and to them SAS is a SCSI protocol. QED.

              BTW, wtf is SaS?

      • by SuperQ ( 431 ) *

        Uhh, are you dense? Distributed storage doesn't mean you use someone else's servers. The software mentioned above is for internal use. Hadoop is used by yahoo for their internal cloud, and Lustre is used by a number of scientific labs that do military work.

      • by sjames ( 1099 )

        That's what strong crypto is for.

      • by jon3k ( 691256 )
        So you think your network is more secure than storing data on Google's servers. Interesting.
    • eh, properly designed systems using the big disk arrays certainly don't have a single point of failure. And their data is replicated to other big disk arrays in other locations. That's why they cost "the big bucks". Your cloud is fine for relatively low-speed low-security read-mostly data, but not for high-volume financial and healthcare systems

    • In a much more localized manor

      We're going to start putting data centers in big houses now?

  • by mpapet ( 761907 ) on Saturday July 25, 2009 @11:38AM (#28819481) Homepage

    Hardware RAID's are not exactly hopping off the shelf and I think many shops are happy with fiberchannel.

    Let's do another reality check: this is enterprise class hardware. Are you telling me you can get SSD RAID/SAN in a COTS package that is cost approximate to whatever is available now? Didn't think so....

    Let's face it, in this class of hardware things move much more slowly.

    • by jon3k ( 691256 )
      Cost per IOPS yes, several vendors are selling SSD now. Cost per terrabyte, no, SSD isn't even close. What we're seeing now is a Tier 0 storage using SSD's. It fits in between RAM cache in SAN controller nodes and on-line storage (super fast, typically fiber channel storage vs near line).

      So previously it looked like (slowest to fastest): SATA (near-line), Fiber Channel (online) -> RAM cache

      Now we'll have: SATA -> FC -> SSD -> RAM

      And in a few years after the technology gets better a
    • by dbIII ( 701233 )
      Not dead just ready to be overtaken by something else. The bizzare idea of a redundant array of fileservers with parallel NFS is already done by Panasas - but I'm too scared to get a price in case I have a heart attack.
  • BAD MATH (Score:5, Interesting)

    by adisakp ( 705706 ) on Saturday July 25, 2009 @11:57AM (#28819631) Journal
    FTA Since a disk sector is 512 bytes, requests would translate to 26.9 MB/sec if 55,000 IOPS were done with this size. On the other end of testing for small block random is 8192 byte I/O requests, which are likely the largest request sizes that are considered small block I/O, which translates into 429.7 MB/sec with 55,000 requests

    I'm not going to believe an article that assumes that because you can do 55K IOPS for 512Byte reads, you can do the same number of IOPS for 8K reads which are 16X larger and then just extrapolate from there. Especially since most SSD's (at least SATA ones) right now top out around 200MB/s and the SATA interface tops out at 300MB/s. Besides there are already real world articles out there where guys with simple RAID0 SSD's are getting 500-600 MB with 3-4 drives using Motherboard RAID much less dedicated harware RAID.
    • Re:BAD MATH (Score:5, Insightful)

      by fuzzyfuzzyfungus ( 1223518 ) on Saturday July 25, 2009 @12:47PM (#28820033) Journal
      "simple RAID0 SSD's are getting 500-600 MB with 3-4 drives using Motherboard RAID much less dedicated harware RAID."

      The last part of that sentence is particularly interesting in the context of this article. "Motherboard RAID" is, outside of the very highest end motherboards, usually just bog-standard software raid with just enough BIOS goo to make it bootable. Hardware RAID, by contrast, actually has its own little processor and does the work itself. Of late, general purpose microprocessors have been getting faster, and cores in common systems have been getting more numerous, at a substantially greater rate than hardware RAID cards have been getting spec bumps(outside of the super high end stuff, I'm not talking about whatever EMC is connecting 256 fibre channel drives to, I'm talking about anything you could get for less than $1,500 and shove in a PCIe slot). Perhaps more importantly, the sophistication of OS support for nontrivial multi-disk configurations(software RAID, ZFS, storage pools, etc.) has been getting steadily greater and more mature, with a good deal of competition between OSes and vendors. RAID cards, by contrast, leave you stuck with whatever firmware updates the vendor deigns to give you.

      I'd be inclined to suspect that, for a great many applications, dedicated hardware RAID will die(the performance and uptime of a $1,000 server with a $500 RAID card will be worse than a $1,500 server with software RAID, for instance) or be replaced by software RAID with coprocessor support(in the same way that encryption is generally handled by the OS, in software; but can be supplemented with crypto accelerator cards if desired).

      Dedicated RAID of various flavors probably will hang on in high end applications(just as high end switches and rouers typically still have loads of custom ASICs and secret sauce, while low end ones are typically just embedded *nix boxes on commodity architectures); but the low end seems increasingly hostile.
      • The current problems with "motherboard RAID" are:

        1. They can't take a BBU, so you either leave write caching turned on on the drives and lose data on an unexpected shutdown (possibly corrupting your array)

        OR

        Turn write caching off on the drives and have incredibly poor write speeds.

        2. The software (and probably the hardware) are no where near smart enough. They might tell you a drive is failing, they might not. If they do they might rebuild the array successfully or may just corrupt it (and if it's your boot

      • by afidel ( 530433 )
        You miss the real world cost of using software raid, licensing. Most applications that justify $500 RAID controllers are licensed on a per cpu or per core model which means using the cpu to accelerate I/O is really freaking expensive. Oracle enterprise lists for $60k/2 cores on linux and windows which means if you max out a core doing I/O you have spent $30k to do I/O, you better be getting some phenomenal rates for that kind of money. Plus hardware raid supplies something that no software raid implementati
    • by SuperQ ( 431 ) *

      It doesn't matter that SATA can do 300MB/s. That's just the interface line rate. Last I did benchmarks of 1T drives (seagate ES.2) they topped out at around 100MB/s. Drives still have a long way to go before they saturate the SATA bus. The only way that happens is if you are using port multipliers to reduce the number of host channels.

      • Re: (Score:3, Interesting)

        by jon3k ( 691256 )
        You forgot about SSDs, consumer versions of which are already doing over 250MB/s reads for less than $3.00/GB. And we're still essentially talking about second generation products (Vertex switched from JMICRON to Indilinx controllers and Intel basically just shrunk down to 34nm for their new ones, although their old version did 250MB/s as well).

        I'm using a 30GB OCZ Vertex for my main drive on my windows machine and it benchmarks around 230MB/s _AVERAGE_ read speed. It cost $130 ($4.30/GB) when I bought
        • by afidel ( 530433 )
          Consumer drives can peak at those kind of rates but they would fall over in a month or six if they were asked to sustain writes at anywhere near those rates. The only inexpensive SSD I would (and have) trusted my data to is the Intel x-25e and it comes in at $15/GB which doesn't compare very favorably for most loads with the still expensive $2/GB for FC HDD's. Things may change in a few generations when SLC rates come down to the $3/GB range, but high end HDD's will probably be at 25c per GB by then.
          • by jon3k ( 691256 )
            Please show me a SAN vendor that sells their certified fiber channel drives for $2/GB. Try more like $6-$10/GB. Let's also not forget that a single SSD can put out over 10 times as many random 4K IOPS than even the fastest fiber channel drives. So, if it comes down to $/IOPS (and don't forget space/power/cooling), SSDs are already blowing fiber channel out of the water. It depends on your application requirements, but I can see a lot of instances in which a couple of shelves of SSDs could come in really
            • by afidel ( 530433 )
              I have a quote that we're about to execute that includes a bunch of 450GB 15k FC drives for ~$1,000 which is within spitting distance of $2/GB. Yes, on a $/IOP basis SLC does in fact beat FC drives even today, but the percentage of my storage that needs IOPS over everything is very low so I have built a balanced array which gives me the best combination of features/price/storage/IOPS. Other people will have other requirements, and there are some that needs IOPS no matter the cost, but I think those are stil
              • by jon3k ( 691256 )
                The last quote I got from EMC for 400GB 15K FC drives was $4,000 PER DRIVE. Let me repeat that, four thousand dollars per drive. EMC wanted $1,000 for a 146GB 15K FC drive. This is their "certified" drives that they've "tested and validated"

                /me makes jerking-off motion

                Here's an HP 146GB 15K FC drive for over $1,000 [cdwg.com].

                Are you sure you don't have your prices mixed up?

                I'm sorry I dont maybe I wasn't clear enough. I totally agree it isn't for every workload, I think it works in a lot of instances,
                • by afidel ( 530433 )
                  Nope, my prices are not mixed up, we are buying an EVA with a significant number of shelves and are hence getting a decent discount, that brings the price per 450GB drive down to ~$1k. When you are buying something in that price range paying anything near list is just not doing your homework. We had competitive bids from all of the big boys except EMC (don't like to micromanage storage and their 5 year costs are so obscene as to be non-starters) and we let all the vendors know it was a competitive bid. HP c
    • by jon3k ( 691256 )
      Vertex (with Indilinx controllers) and Intel (even the "cheap" MLC drives from both vendors that are less than $3.00/GB) are seeing 250MB/s-270MB/s actual real world results for reads. The actual throughput of SATA 3G is actually slightly less than 300MB/s so essentially we're at the limitation of SATA 3G, or very very close -- too close for comfort.
      • Remember that Intel entered this market as a tiger out for blood with their *first* SSD throwing data at just under the SATA300 cap. This isnt a coincedence.

        When SATA600 goes live, expect Intel and OCZ to jump right up to the 520MB/sec area as if it was trivial to do so... (because it is!)
        ioFusion has a PCIe flash solution that goes several times faster than these SATA300 SSD's. The problem is SATA. The problem is SATA. The problem is SATA.
    • by drsmithy ( 35869 )

      Besides there are already real world articles out there where guys with simple RAID0 SSD's are getting 500-600 MB with 3-4 drives using Motherboard RAID much less dedicated harware RAID.

      It is unlikely "dedicated hardware RAID" would be meaningfully faster.

  • enterprise storage (Score:4, Insightful)

    by perlchild ( 582235 ) on Saturday July 25, 2009 @11:57AM (#28819635)

    Storage has been the performance bottleneck for so long, it's a happy problem if you actually must increase the bus speeds/cpu processors/get faster memory on raid cards to keep up. Seems to me the article(or at least the summary) was written by someone hadn't been following enterprise storage for very long...

    • Re: (Score:3, Interesting)

      by ZosX ( 517789 )

      That's kind of what I was thinking too. When you really start pushing the 300mb/s sata gives its hard to find something to complain about. Most of my hard drives max out at like 60-100mb a second and even the 15,000k drives are not a great deal faster. Low latency, fast speeds, increased reliability. This could get interesting in the next few years. Heck why not just build a raid 0 controller into the logic card with a sata connection and break the ssd into a bunch of little chunks and raid 0 them all max p

      • by jon3k ( 691256 )
        "Heck why not just build a raid 0 controller into the logic card with a sata connection and break the ssd into a bunch of little chunks and raid 0 them all"

        Cost mostly, you'd need tons of controllers, cache, etc. Plus you can already nearly saturate SATA 3G with any decent SSD (Intel, Vertex, etc) so it's kind of pointless. The new Vertex and Intel SSDs are benchmarking at 250MB/s. Not point it making them much faster until we have SATA 6G.
      • by drsmithy ( 35869 )

        Heck why not just build a raid 0 controller into the logic card with a sata connection and break the ssd into a bunch of little chunks and raid 0 them all max performance right out of the box so you get the performance advantages of raid without the cost of a card and the waste of a slot?

        Because an error anywhere nukes the whole shebang.

    • by HockeyPuck ( 141947 ) on Saturday July 25, 2009 @12:51PM (#28820073)

      Ah... pointing the finger at the storage... My favorite activity. Listening to DBAs, application writers, etc point the finger at the EMC DMX with 256GB of mirrored cache and 4Gb/s FC interfaces. You point your finger and say, "I need 8Gb FibreChannel!. Yet when I look at your hba utilization over a 3mo period (including quarter end, month end etc..) I see you averaging a paltry 100MB/s. Wow. Guess I could have saved thousands of dollars with going with 2Gb/s HBAs. Oh yeah, and you have a minimum of two HBAs per server. Running a nagios application to poll our switchports for utilization, the average host is running maybe 20% utilization of the link speed, and as you beg, "Gimme 8Gb/s FC", I look forward to your 10% utilization.

      We've taken whole databases and loaded them into dedicated cache drives on the array, and surprise, no performance increase. DBAs and application writers have gotten so used to yelling, "Add Hardware! That they forgot how to optimize their applications and sql queries."

      If storage was the bottleneck, I wouldn't be loading up storage ports (FAs) with 10-15 servers. I find it funny that the only devices on my 10,000 port SAN that can sufficiently drive IO are media servers and the tape drives (LTO-4) that they push.

      If storage was the bottleneck there would be no oversubscription in the SAN or disk array. Let me know when you demand a single storage port per HBA, and I'm sure my EMC will take us all out to lunch.

      I have more data than you. :)

      • by Anonymous Coward on Saturday July 25, 2009 @01:14PM (#28820205)

        Ah... pointing the finger at the storage... My favorite activity. Listening to DBAs, application writers, etc point the finger at the EMC DMX with 256GB of mirrored cache and 4Gb/s FC interfaces. You point your finger and say, "I need 8Gb FibreChannel!. Yet when I look at your hba utilization over a 3mo period (including quarter end, month end etc..) I see you averaging a paltry 100MB/s. Wow. Guess I could have saved thousands of dollars with going with 2Gb/s HBAs. Oh yeah, and you have a minimum of two HBAs per server. Running a nagios application to poll our switchports for utilization, the average host is running maybe 20% utilization of the link speed, and as you beg, "Gimme 8Gb/s FC", I look forward to your 10% utilization.

        You do sound like you know what you're doing, but there is quite a difference between average utilization and peak utilization. I have some servers that average less than 5% usage on a daily basis, but will briefly max out the connection about 5-6 times per day. For some applications, more peak speed does matter.

        • by Spit ( 23158 )

          What's the cost-benefit analysis of buying hardware that has headroom for those .1% peak events, vs data housekeeping and app/sql profiling? This is a management problem, not a technical one.

          • by jon3k ( 691256 )
            "What's the cost-benefit analysis of buying hardware that has headroom for those .1% peak events, vs data housekeeping and app/sql profiling? This is a management problem, not a technical one."

            Depends on the business requirements and the number of end-points, but I wouldn't rule it out completely. For example, production companies moving large amounts of video for short periods of time, it might be worth the difference between 4Gb and 8Gb fiber channel, I don't know. You're also assuming those peak eve
      • by swb ( 14022 )

        In my experience, DBAs and their fellow travelers in the application group like to point their finger at SANs and virtualization and scream about performance, not because the performance isn't adequate but because SANs (and virtualization) threaten their little app/db server empire. When they no longer "need" the direct attached storage, their dedicated boxes get folded into the ESX clusters and they have to slink back into their cubicles and quit being server & networking dilettantes.

      • by Slippy. ( 42536 ) on Saturday July 25, 2009 @01:58PM (#28820557)

        Sort of true, but not entirely accurate.

        Is the on-demand response slow? Stats lie. Stats mislead. Stats are only stats. The systems I'm monitoring would use more I/O if they could. Those basic read/write graphs are just the start. How's the latency? Any errors? Pathing setup good? Are the systems queuing i/o requests while waiting for i/o service response?

        And traffic is almost always bursty unless the link is maxed - you're checking out a nice graph of the maximums too, I hope? That average looks mighty deceiving when long periods are compressed. At an extreme over months or years, data points can be days. Overnight + workday could = 50%. No big deal on the average.

        I have a similiar usage situation on many systems, but the limits are generally still storage dependent issues like i/o latency (apps make a limited number of requests before requests start queuing), poorly grown storage (a few luns there, a few here, everything is suddenly slowing down due to striping in one over-subscribed drawer), and sometimes unexpected network latency on the SAN (switch bottlenecks on the path to the storage).

        Those graphs of i/o may look pitiful, but perhaps that's only because the poor servers can't get the data any faster.

        Older enterprise SAN units (even just 4 or 5 years ago) kinda suck performance wise. The specs are lies in the real world. A newer unit, newer drives, newer connects and just like a server, you'll be shocked. What'cha know, those 4Gb cards are good for 4Gb after all!

        Every year, there's a few changes and growth, just like in every other tech sector.

      • Well boy did I not expect this kind of reaction... I'm kinda on your side, really. I meant, here's someone that's saying that SSDs means you're no longer starving for spindles... And I say "well that's good, they were holding us back, we can do something better now, that's not a problem." On the other hand, it seems it's a lot more loaded politically in places that don't do this with just three admins, and no dedicated storage admins, so I'll just shut up now cuz I hate politics. You guys have a nice day

      • by markk ( 35828 )

        Who cares about average use? The cost is driven by the PEAK use. That is why the average use for HBA's is almost nothing, but you are paying double the money or more because of the 8 hours a month you need to smoke. And woe betide the Architect who suggest postponing a business meeting for 48 hour every month so he can save $20 million a year. Seriously.

      • Well, that was unexpected for me too. And you know, you are right. Real world applications behave quite differently from how academical models say they would, that is because the models didn't model teams limitations and the unavoidable mistakes (from the techies and from the HR) that add into some very siginificant amount on any project.

        Too bad I didn't let that academical misconception go yet. That is why I was surprized.

  • The first question is really, why RAID a SSD? It's already more reliable than a mechanical disk, so that argument goes out the window. You might get some increased performance, but that's often not a big factor.

    The second question is, with processors coming with 8 cores, why have some separate specialized controller that handles RAID and not just do it in software?

    • Re: (Score:3, Informative)

      The second question is, with processors coming with 8 cores, why have some separate specialized controller that handles RAID and not just do it in software?

      I much prefer s/ware raid (Linux kernel dm_mirror), it removes a complicated piece of h/ware which is just another thing to go wrong. It also means that you can see the real disks that make up the mirror and so monitor it with the smart tools.

      OK: if you do raid5 rather than mirroring (raid1) you might want a h/ware card to offload the work to, but for many systems a few terabyte disks are big and cheap enough to just mirror.

    • by adisakp ( 705706 )

      The first question is really, why RAID a SSD? It's already more reliable than a mechanical disk, so that argument goes out the window. You might get some increased performance, but that's often not a big factor.

      The second question is, with processors coming with 8 cores, why have some separate specialized controller that handles RAID and not just do it in software?

      RAID0 for Speed. SSD's in RAID0 can perform 2.5-3X faster [hothardware.com] than a single drive. A RAID SSD array can challenge the speed of a FusionIO card [hothardware.com] that is several thousand dollars.

      Now that the new faster 34nm Intel SSD's can be preordered for under $250, it's reasonable for an enthusiast to buy 3-4 of them and thrown them in a RAID0 array. Also, software (or built-in MB RAID) is fine -- a lot of the sites have shown that 3 SSD drives is the sweet point for price/performance using standard MB RAID controllers.

      • Don't forget about the "Battleship MTRON" guys that raided up 8 MTRON SSD's (the fastest SSD's at the time) several years ago and then had a lot of trouble actualy finding a raid controller than could handle the bandwidth. This years SSD's are twice as fast, and expect performance to double again within 12 months.
      • by jon3k ( 691256 )
        I'll take 4 drives in RAID10 please :)
    • by Yert ( 25874 )
      Because I can't buy a 26TB SSD drive, but I can put 52 500GB SSD drives in two CoreRAID chassis and mount them as one filesystem...as opposed to the 2 Sun storage arrays we use now, that are fiber attached and starting to get a little ... slow. SSDs would give us 10x the IO overhead.
    • In enterprise, it doesn't matter if the disk has a less likely chance of failing; redundancy for HA is worth the extra cost.

      If someone is spending the money on SSD then performance had better be a big factor!
      • That's only true to a point. If the reliability of the SSD gets to the point where it's about as likely as the RAID controller to fail, then the RAID controller is just an extra point of failure that will not increase your availability at all. However, AFAIK SSDs aren't that reliable yet so the RAID controllers are still worth it.

        • by jon3k ( 691256 )
          "However, AFAIK SSDs aren't that reliable yet so the RAID controllers are still worth it."

          Please stop spreading baseless FUD.
          • What FUD? I said AFAIK. I haven't been following them closely. If I'm wrong please feel free to correct me instead of jumping down my throat.

            • by jon3k ( 691256 )
              "FUD" is fear, uncertainty and doubt. By saying unproven, un-researched things like this you continue to promote a commonly held misconception about solid state drives. It's like two women at the hair salon gossiping, stop it. Either research it and come up with a useful contribution to the thread or just don't reply.
    • the on board chips are not build for high speed / useing all the ports at the max at one time.

    • by Helmholtz ( 2715 )

      This is where ZFS has some potential to become even more important than it already is.

      The reason you RAID a SSD is to protect against silent data corruption, which SSDs are not immune from. While you don't necessarily need RAID for this with ZFS, it certainly makes it easier.

      The point about the insane abundance of CPU power is one that ZFS specifically takes advantage of right out of the starting gate.

    • I have data sets spanning multiple terabytes. One recent PhD graduate in the lab I support accumulated 20 TB of results during his time here. Even if I had highly reliable SSDs that never failed, I'd still toss the SSDs together in a zpool to get the capacities I need to accommodate a single data set. RAID is not just about redundancy. With SSDs, I'd probably use RAID5 instead of RAID6 just in case I had a freak bad drive, but RAID in some form is here to stay.
    • by drsmithy ( 35869 )

      The second question is, with processors coming with 8 cores, why have some separate specialized controller that handles RAID and not just do it in software?

      Transparency and simplicity. It's a lot easier dealing with a single device than a dozen.

  • Multiple interfaces and lots of block servers.

    Does anyone actually still use NFS?

     

    • Re: (Score:2, Informative)

      by Anonymous Coward

      Of course. NFS provides an easy to use concurrent shared filesystem that doesn't require any cluster overhead or complication like GFS or GPFS.

    • Re: (Score:3, Informative)

      by drsmithy ( 35869 )

      Does anyone actually still use NFS?

      Of course. It's nearly always fast enough, trivially simple to setup, and doesn't need complicated and fragile clustering software so that multiple systems can access the same disk space.

    • by dstar ( 34869 )

      Where I work, we've only got a few petabytes of NFS storage. And it's only used for mission critical (in the literal meaning of the term -- no access to data, no work gets done, literally $millions lost if a deadline is blown) data.

      NetApp doesn't seem to be having any trouble selling NFS, either.

      So no, I don't think anyone uses NFS anymore.

    • by guruevi ( 827432 )

      Not everybody (hardly anyone) needs a single block device in a work environment. You might as well hang the hard drive in their systems if that's all you need, cheaper, faster and simpler. Also block devices don't separate very well. You have to assign and reserve a certain block of data no matter whether it's used.

      NFS is much more granular that way, you put everything on a large block device, give it some permissions and you're good to go. Also for shared data, sharing block devices might not be a good ide

  • Not quite (Score:4, Informative)

    by greg1104 ( 461138 ) <gsmith@gregsmith.com> on Saturday July 25, 2009 @12:11PM (#28819763) Homepage

    There may need to be some minor rethinking of controller throughput for read applications on smaller data sets for SSD. But right now, I regularly saturate the controller or bus when running sequential RW tests against a large number of physical drives in a RAID{1}0 array, so it's not like that's anything new. Using SSD just makes it more likely that will happen even on random workloads.

    There are two major problems with this analysis though. The first is that it presumes SSD will be large enough for the sorts of workloads people with RAID controllers encounter. While there are certainly people using such controllers to accelerate small data sets, you'll find just as many people who are using RAID to handle large amounts of data. Right now, if you've got terabytes of stuff, it's just not practical to use SSD yet. For example, I do database work for living, and the only place we're using SSD right now is for holding indexes. None of the data can fit, and the data growth volume is such that I don't even expect SSDs to ever catch up--hard drives are just keeping up with the pace of data growth.

    The second problem is that SSDs rely on volatile write caches in order to achieve their stated write performance, which is just plain not acceptable for enterprise applications where honoring fsync is important, like all database ones. You end up with disk corruption if there's a crash [mysqlperformanceblog.com], and as you can see in that article once everything was switched to only relying on non-volatile cache the performance of the SSD wasn't that much better than the RAID 10 system under test. The write IOPS claims of Intel's SSD products are garbage if you care about honoring write guarantees, which means it's not that hard to keep with them after all on the write side in a serious application.

    • The fact that SSD perf drops like a rock when you actually need to be absolutely sure the data makes it to disk is huge factor in enterprise storage. No enterprise storage customer is going to accept the possibility their data goes down the bit-bucket just because somebody tripped over the power cord. Enterprise databases are built around the idea that when the storage stack says data has been written, it has, in fact, been written. Storage vendors spend a great deal of money, effort, and complexity guar

    • Re: (Score:3, Insightful)

      The second problem is that SSDs rely on volatile write caches in order to achieve their stated write performance, which is just plain not acceptable for enterprise applications where honoring fsync is important, like all database ones. You end up with disk corruption if there's a crash, and as you can see in that article once everything was switched to only relying on non-volatile cache the performance of the SSD wasn't that much better than the RAID 10 system under test. The write IOPS claims of Intel's SS

      • Re: (Score:3, Insightful)

        by greg1104 ( 461138 )

        You can't turn fsync into a complete noop just by putting a cache in the middle. A fsync call on the OS side that forces that write out to cache will block if the BBWC is full for example, and if the underlying device can't write fast enough without its own cache being turned on you'll still be in trouble.

        While the cache in the middle will improve the situation by coalescing writes into the form the SSD can handle efficiently, the published SSD write IOPS numbers are still quite inflated relative to what y

        • by jon3k ( 691256 )
          "The other funny thing you discover if you benchmark enough of these things is that a regular hard drive confined to only use as much space as a SSD provides is quite a bit faster too."

          Yes there's an improvement, but to compare read IOPS from an enterprise SSD to a short-stroked SATA disk on a purely performance basis isn't even close. We're talking orders of magnitude slower.

          I think SSDs really shine when you get into situations where your performance requirements vastly outweigh our capacity requir
    • by AllynM ( 600515 ) *

      First a quick clarification: Intel X25 series SSDs do not use their RAM as a data writeback cache. Intel ships racks full of both M and E series drives, with those drives living in a RAID configuration. They couldn't pull that off if the array was corrupted on power loss. The competition had to start using large caches to reduce write stutters and increase random write performance, mostly in an attempt to catch up to Intel.

      The parent article is a bit 'off' as far as bandwidth vs. IOPS on RAID controllers

    • There are two major problems with this analysis though. The first is that it presumes SSD will be large enough for the sorts of workloads people with RAID controllers encounter. While there are certainly people using such controllers to accelerate small data sets, you'll find just as many people who are using RAID to handle large amounts of data. Right now, if you've got terabytes of stuff, it's just not practical to use SSD yet. For example, I do database work for living, and the only place we're using SSD right now is for holding indexes.

      That's probably true for your databases, but are databases that measure in terabytes really the norm?

      None of the data can fit, and the data growth volume is such that I don't even expect SSDs to ever catch up--hard drives are just keeping up with the pace of data growth.

      The latest SSD drives of Intel already has room for 320 GB. These are low end consumer disks. Once these things get popular you'll see a sharp increase in production volume. The growth *rate* of flash SSD is very, very high. They haven't caught up yet but I'm quite sure that they will, if only because the hard disks only seem to have these three advantages (size, price and many years of experience with them)

  • All wrong. (Score:3, Informative)

    by sirwired ( 27582 ) on Saturday July 25, 2009 @12:16PM (#28819797)

    1) Most high-end RAID controllers aren't used for file serving. They are used to serve databases. Changes in filesystem technology don't affect them one bit, as most of the storage allocation decisions are made by the database.
    2) Assuming that a SSD controller that can pump 55k IOPS w/ 512B I/O's can do the same w/ 4K I/O's is stupid and probably wrong. That is Cringely math; could this guy possibly be as lame?
    3) The databases high-end RAID arrays get mostly used for do not now, and never have, used much bandwidth. They aren't going to magically do so just because the underlying disks (which the front-end server never even sees) can now handle more IOPS.

    All SSD's do is flip the Capacity/IOPS equation on the back end. Before, you ran out of drive IOPS before ran out of capacity. Now, you get to run out of capacity before you run out of IOPS on the drive side.

    Even if you have sufficient capacity (due to the rapid increase in SSD capacity), you are still going to run out of IOPS capacity on the RAID controller before you run out of IOPS or bandwidth on the drives. The RAID controller still has a lot of work to do with each I/O, and that isn't going to change just because the back-end drives are now more capable.

    SirWired

    • by jon3k ( 691256 )
      "All SSD's do is flip the Capacity/IOPS equation on the back end. Before, you ran out of drive IOPS before ran out of capacity. Now, you get to run out of capacity before you run out of IOPS on the drive side."

      Thank you so much for summarizing that point so succinctly, I'm stealing that line, hope you don't mind :)
    • Re: (Score:3, Interesting)

      by AllynM ( 600515 ) *

      Well said. I've found using an ICH-10R kills that overhead, and I have seen excellent IOPS scaling with SSDs right on the motherboard controller. I've hit over 190k IOPS (single sector random read) with queue depth at 32, using only 4 X25-M G1 units. The only catch is the ICH-10R maxes out at about 650-700 MB/sec on throughput.

      Allyn Malventano
      Storage Editor, PC Perspective

No spitting on the Bus! Thank you, The Mgt.

Working...