Follow Slashdot blog updates by subscribing to our blog RSS feed

 



Forgot your password?
typodupeerror
×
Data Storage Intel Technology

Long-Term Performance Analysis of Intel SSDs 95

Vigile writes "When the Intel X25-M series of solid state drives hit the market last year, there was little debate that they were easily the best performing MLC (multi-level cell) offerings to date. The one area in which they blew away the competition was with write speeds — initial reviews showed consistent 80MB/s results. However, a new article over at PC Perspective that looks at Intel X25-M performance over a period of time shows that write speeds are dramatically reduced from everyday usage patterns. Average write speeds are shown to drop to half (40MB/s) or less in the worst cases, though the author does describe ways that users can recover some of the original drive speed using standard HDD testing tools." Reader MojoKid contributes related SSD news that researchers from the University of Tokyo have developed a new power supply system which will significantly reduce power consumption for NAND Flash memory.
This discussion has been archived. No new comments can be posted.

Long-Term Performance Analysis of Intel SSDs

Comments Filter:
  • Damn (Score:1, Funny)

    by Anonymous Coward

    Where is everyone? Oh right. Friday night.

  • Why? (Score:4, Interesting)

    by IamGarageGuy 2 ( 687655 ) on Friday February 13, 2009 @11:20PM (#26852839) Journal
    I didn't see anything that answered the question of why this would happen. I may be slow but shouldn't it either fail or work? Is storage being lost and therefore getting less with more time used to find a good area? Please don't mod me as troll for not knowing (maybe flamebait for being stupid, I guess)
    • well, as the cells age it has to work harder and harder to find cells that are working to write on. probably involves some kind of write then test scenario, and the more times it has to rewrite data due to bad cells the slower it gets to complete a singe write.
      • Re:Why? (Score:5, Informative)

        by Dr. Ion ( 169741 ) on Saturday February 14, 2009 @01:08AM (#26853393)

        Um... no.

        When cells age, they take longer to erase. This happens over 5,000, 10,000 cycles or longer. It's not dramatic, and eventually the cells fail in a way more severe than can be corrected by the ECC.

        Because there is a (software) process to bring full speed back to the drive, we can safely conclude that none of the slowdown is related to cell aging or other cell-level issues. It's more of an organization and fragmentation issue.

        • It brings it back to full speed, but I'd bet that it returns to the slow state much faster.

          • Re: (Score:2, Informative)

            by AllynM ( 600515 ) *

            There was no difference in how long it took to fragment. If we wrote a nasty enough mix of smaller file sizes to the drive, performance would drop right at the point where all flash was written to at least once (i.e. just over the 80GB mark).

            After running HDDErase on the drive, it went the same *exact* 80 MB/sec write speed each and every time. Additionally, running successive software secure erasures (writing 0's across all 80GB) showed 0 drop in speed even after 10 passes.

            In testing several different SS

            • I'm curious -

              How much is the speed difference between a new Hard Disk Drive and a fragmented drive? Do HDDs also slowdown with age?

              • No they dont. The filesystem can get fragmented, but thats a different matter which also applies to flash.
                While they are still working, they will always pump out the same speed.

    • Re:Why? (Score:5, Informative)

      by SpazmodeusG ( 1334705 ) on Friday February 13, 2009 @11:41PM (#26852991)
      The Intel SSD (and all SSDs) are made up of big addressable blocks. On the Intel SSD these are 64KiB in length. When you read or write to the drive the internal controller actually reads or writes an entire 64KiB block.
      A simple change to 1 byte means a read of the entire 64KiB block that byte is in, a change of the data and then a write of 64KiB.
      If the filesystem isn't flash-aware you can suffer a theoretical performance hit of being 65536 times slower because of this.

      So what you really need is a filesystem that stores files in 64KiB blocks and groups reads and writes to the same blocks together as 1 operation.
      • Re: (Score:3, Informative)

        I thought it could flip a bit too 1 at will without having to rewrite the whole block but if you want to write a 0 it needs to read the whole block, wipe it out and rewrite it with the same data but with the bit flipped to 0. But I wouldn't really know, I'll use SSDs when they cost about the same as hard drives.

        • This is technically correct, but you'd need to massively redesign an operating system's block device and caching layers to take advantage of it. The smallest write that a piece of software will ever do is one byte, and the probability of such a write only changing 0s to 1s is very low. If you assume an equal probability of every prior and new value, then, for every bit, you have four equally likely transitions: 0 to 0, 0 to 1, 1 to 0, 1 to 1. The only one that is a problem is 1 to 0, so you have a 0.25 p
          • Wouldn't the probability be heavily weighted in favor of "1" to "1"? Think of all the minor edits that typically get done to existing files...
            • by maxume ( 22995 )

              Unchanged bytes might be slightly more common than changed bytes, but deleting a single byte might require the file to be rewritten. The 'only 0's get flipped' case is not worth giving special consideration.

          • by sjames ( 1099 )

            The idea is to avoid erases or at least batch them.

            The idea is to allocate metadata blocks somewhere to manage the storage. The metadata for a block includes a replacement pointer set initially to all ones. If the block changes, write the replacement block number into that pointer. If a block is released, write 0 to the pointer.

            This can be done in the drive logic or can be handled by the OS. The catch is that the device gets slower as time goes on. Eventually, you are forced to defrag the flash management

      • Re:Why? (Score:5, Informative)

        by tlhIngan ( 30335 ) <slashdot.worf@net> on Saturday February 14, 2009 @01:08AM (#26853395)

        The Intel SSD (and all SSDs) are made up of big addressable blocks. On the Intel SSD these are 64KiB in length. When you read or write to the drive the internal controller actually reads or writes an entire 64KiB block.
        A simple change to 1 byte means a read of the entire 64KiB block that byte is in, a change of the data and then a write of 64KiB.
        If the filesystem isn't flash-aware you can suffer a theoretical performance hit of being 65536 times slower because of this.

        So what you really need is a filesystem that stores files in 64KiB blocks and groups reads and writes to the same blocks together as 1 operation.

        Actually, NAND flash comes in 2 block sizes - small block (16kiB/block, 512bytes/page, 32 pages/block), and large block (128kiB/block, 2048bytes/pages, 64 pages/block).

        Also, in NAND flash, a "write" operation can turn a "1" bit to a "0" bit. An "erase" operation turns a "0" bit into a "1" bit. Writes can work at the bit level, erases at the block level. (Though, large block NAND can NOT be partial-page programmed, so you must write 2048 bytes at once, but you can read all 2048 bytes, flip one bit, then write it all back). This characteristic is used by the flash management routines in order to manage the flash block. Marking pages as "discard" or "ready for erase" is done by flipping a 1 bit to 0 since that's easy. You can write a block partially, so you don't have to incur a huge 128kiB write always.

        Given this, it's a block device, so you can't write 1 byte anyhow - you must write the sector size, which is emulated as 512 bytes. What normally happens is that the SSD will mark a page as "dirty" to indicate it's not to be used, and remap that page's contents onto a new page elsewhere, thus only performing a 2048 byte write (plus 64 out of band bytes).

        Now, what happens when all the blocks are used? The flash routines have to erase a block, but before erasing a block, it has to make sure all the pages within it are "dirty". If there are non-dirty pages, they're copied to another block, and when all non-dirty pages are copied, that block is erased. If your access pattern is such that all the blocks have non-dirty pages, it takes a little while to actually move all the data around to get blocks that can be erased. Do enough random I/O, and this can happen quite easily.

        • Re: (Score:3, Informative)

          by Dr. Ion ( 169741 )

          Older flash devices allowed multiple writes to one page, but new ones do not.

          The higher-density MLC devices do not allow you to read a page, flip a bit to 0 and overwrite it. They require that pages be written just one, and in order.

          This is causing no end of frustration for the Microsoft mobile filesystems, which frequently overwrote pages to flag them.

      • Closer, but.. no. (Score:5, Informative)

        by Dr. Ion ( 169741 ) on Saturday February 14, 2009 @01:11AM (#26853403)

        NAND blocks are *erased* in large blocks, probably 128KB or larger in this case.

        However, the read and write operations occur at a *page* level, not block. NAND pages today are typically 2K or 4KB in size.

        So you can read and write in smaller units than 128KB.

        However, to erase any byte of the NAND, you have to relocate the preserved data and erase a whole block.

        Because these drives operate on huge aggregate arrays of NAND, their block structure may be much larger, or they may have very complicated and smart algorithms to re-map write new data while waiting to perform erases much later.

      • This isn't Wikipedia, we call them "KB" here. Thanks.

    • Re:Why? (Score:5, Informative)

      by Rockoon ( 1252108 ) on Friday February 13, 2009 @11:58PM (#26853081)
      It happens because flash drives write on very large (512KB) "blocks", but they still pretend like they have average sized (4KB) "sectors".

      Essentialy what the intel write-combining technology is doing is combining multiple small (4KB) writes into a single block, and letting the old block become fragmented (having a bunch of 4KB holes in it.)

      The scenario in the nutshell:

      You have a 1MB file and a program which modifies a single 4KB chunk of it. Intels technology marks the original 4KB chunk within its original "block" as erased, and then allocates a new block (using the wear leveling algorithm) to hold the new version of the 4KB chunk and additionally combines it with any other small writer operations that may have recently occured or will recently occur. Up to 128 such 4KB writes can be combined into a single block write.

      After this is done many hundreds of thousands of times, however, the drive begins to be in a state where nearly every "block" is only partialy used. The write combiner itself is stuck with whatever the wear leveling algorithm handed it, which is now a partialy used block instead of a fully virgin block. It can no longer combine 128 small 4KB writes together, but maybe only has space to combine 10 of them, or in the worst case scenario.. 1 of them.
      • by Dr. Ion ( 169741 )

        Again, it's only the ERASE unit that is huge -- 64KB, 128KB, or 256KB on the device itself.

        You can't erase 4KB alone.

        It gets more complicated when you consider huge parallel arrays of NAND, and the complex logical remapping that goes on to give the appearance of a typical 512-byte sector device.

    • by AllynM ( 600515 ) *

      Nothing is lost, it just goes slower. Due to the fragmentation caused by write combining, the drive has to shuffle more data around when you write to those same areas later on. The flash is still being written at full speed as it copies the data internally, but the drive can only accept new data at the reduced speeds seen.

      We're working with Intel to help them reproduce the more significant issue we saw.

      Allyn Malventano
      PCPer Editor

  • With the fix for this problem being essentially "nuke the drive and reinstall periodically", there's really no fix until you get a flash-aware filesystem. Too many virtualization layers between your app doing the write, and the bits being flipped.

    This could be useful for ETL jobs or other heavy 'batch' type work, as the nature of the access will essentially 'reset' the drive for the next pass.

    • Something like LFS, which predates ZFS by almost two decades, would also work. Everything old is new again.
  • TL:DR (Score:4, Insightful)

    by Nursie ( 632944 ) on Friday February 13, 2009 @11:33PM (#26852937)

    That article is a multi-page annoyance, the grammar is bad and we already have flash-aware filesystems like jffs2.

    • Re:TL:DR (Score:4, Informative)

      by bcrowell ( 177657 ) on Saturday February 14, 2009 @12:11AM (#26853145) Homepage

      That article is a multi-page annoyance, the grammar is bad and we already have flash-aware filesystems like jffs2.

      As far as I can tell from some quick googling and checking on Wikipedia, jffs2 isn't much of a competitor at this point, e.g., it's apparently not really usable on flash chips bigger than 512 Mb. Maybe UBIFS or LogFS? None of them seem to be really mature.

      • Re:TL:DR (Score:5, Interesting)

        by TheRaven64 ( 641858 ) on Saturday February 14, 2009 @10:02AM (#26855489) Journal
        Not sure about the Linux world, but LFS on NetBSD counts as mature. It's been sitting in the BSD tree since 4.4BSD (1990, a year before the first Linux release) and is well supported by NetBSD, although the other BSDs dropped it from their trees in the intervening decades because it didn't provide major benefits on rotating mechanical disks. With flash becoming cheap, suddenly it's seeing a lot more interest...
    • Re: (Score:2, Interesting)

      by mcbridematt ( 544099 )

      Will the traditional flash file systems (jffs2) etc. still work when we have SSD's interfacing over SATA? USB sticks don't work with it because they 'pretend' to be a hard drive over USB, and same for the SSD's over SATA. jffs wants the flash device (MTD) interface.

      Intel employee Matthew Wilcox spoke at linux.conf.au about some kernel performance improvements related to the Intel SSD drives - redundant ATA calls that have been removed, and allowing larger sector sizes under ATA 8 [lwn.net], so maybe the authors of th

    • Re: (Score:3, Insightful)

      by renoX ( 11677 )

      Flash-aware filesystem currently only works on embeded setup where there is direct access to the Flash.
      Given the need for compatibility, SSD will always have a controller showing the SSD as a disk, but I agree that it'd be nice if they would add additionnal lower level access in the case the computer is able to use Flash-aware filesystem.

    • by cuby ( 832037 )
      Dont't be a troll!
      The article was very useful and, as far I can tell, the author was systematic and did a good research. This conclusions are very useful for everyone using SSD. Did you ever used jffs2? Do you trust your files to it? Come on... At present there is no production reliable, SSD oriented file system available.
  • Fortunately for us, flash based storage has access times nearly as fast as RAM

    My 300MHz DDR bus had 300M x 2 x 8 == 600M x 8 == 4.8GB/s sequential read access speed.

    • by Aluvus ( 691449 )
      I assume you realize that "maximum theoretical speed of the interface" and "real-world performance" are not the same thing.
      • Consider that's 300MHz DDR, and DDR3 can hit 12.8 gig peak. Peak performance and real-world performance will be analogous; a slower chip's peak performance and real-world performance will be slower.
    • Re: (Score:1, Informative)

      by Anonymous Coward

      Access time != bandwidth. But you're right, RAM still has much quicker access times. Still, both seem instantaneous to humans; is a 0.2 ms access time really so bad for most applications?

    • by Dr. Ion ( 169741 )

      Access time != sequential bulk read throughput.

      Think hard drive vs flash drive.

      Flash does have "access time" close to RAM, since it doesn't have to seek or do complex addressing.

      When you have these huge banks of flash acting as one drive, then "access time" becomes a computational problem of how fast you can look up the physical location of the user's data, based on a logical sector address.

      Still faster then mechanically moving a drive head, of course.

  • The problem, according to the article is writing small files over and over again, like the journal in journaling file systems, so would not using a journaling file system reduce the problem?
    • I'm confused. Is the problem the deteriorating hardware or an unsuitable filesystem?

      No I didn't RTA..
      • Re: (Score:3, Insightful)

        by Dr. Ion ( 169741 )

        One of the biggest challenges of the coming years will be finding and developing filesystems (logical data stores) that take advantage of the strengths of flash memory while deminishing the weaknesses of it.

        Our approach today is mapping large banks of Flash to look like a hard drive, and then using a filesystem that is optimized to reduce seek activity. (Cyl/Hds/Tracks-per-Sector..)

        EXT3 on SSD, FAT on huge SD cards, it's just shoe-horning our old filesystems onto new media. It makes about as much sense as

        • Or we could use an old filesystem, like LFS from 1990, which was part of the last BSD release from UCB. It works best on devices where random read and linear writes are very cheap. Devices like flash.
  • by bogaboga ( 793279 ) on Friday February 13, 2009 @11:45PM (#26853017)

    I am patiently waiting for these SSDs and plan to test them on a MythTV distro box. I will get a fully compatible Linux SSD notebook onto which a MythTV distro will be installed.

    Then with 3 TV cards, I will see how these SSDs measure up on reading/writing/transcoding etc. My intention is to work the SSD for about a week. Watch this space for results.

    I do not think that Intel will deliver the "golden" SSD. I think Samsung's SSD [samsung.com] effort will bear results faster. Those videos say a lot.

  • SLC vs MLC (Score:4, Interesting)

    by w0mprat ( 1317953 ) on Saturday February 14, 2009 @12:25AM (#26853195)
    Isn't the problem partly MLC? SLC has consistently better small random write performance. Many cheap SSDs use MLC for obvious reasons, it fairs well in benchmarking -MLC has relatively high read performance- but write performance hurts real bad in real world usage. You may get noticeable micro-lag anytime the OS writes to storage. Application loading may be snappy for example, but the whole system slows down while writes are done. It's good to see the truth coming out amongst all the benchmarketing

    It's early days for SSDs. I'll be sticking with my power guzzling magnetic frisbe stacks for a while yet.
    • Many cheap SSDs use MLC for obvious reasons,

      You hit the nail on the head. This high-price Intel SSD is just a cheap MLC unit with a big brand name and inflated expectations.

      I'm usually quick to adopt new tech, but I'm still not satisfied with SSDs. I can't think of a good reason to use one in a desktop PC right now. They're slow, they die young, and they have perverse design quirks like oversized pages that result in internal fragmentation, and oh yeah, they cost an arm and a leg. Once someone releases an SSD that solves ALL of those sticky point

      • Re: (Score:2, Informative)

        by Anonymous Coward

        Once someone releases an SSD that solves ALL of those sticky points, and ideally delivers enough random-access throughput to saturate the 300MB/s SATA line (or whatever bus is mainstream by then), that's when I'll jump on board.

        Well, like myself, you will be waiting for a non-flash based SSD then.

        Inevitably, something like PRAM [wikipedia.org] will displace Flash, and it can't happen soon enough. Until then, I would much rather see some of that fab capacity reclaimed for DRAM production.

      • Re: (Score:3, Interesting)

        by Kneo24 ( 688412 )

        I've been using an Intel SSD as a boot drive and I think it's worth every penny so far. I have a few programs and games on the boot drive and they all load up considerably faster than the alternatives. I don't care about write speeds. Their size alone means they're not really meant for storage yet, so using it as such is a bit retarded. If you're doing a lot of write operations to your SSD, you should probably think about moving that file(s) to a different storage device.

      • by afidel ( 530433 )
        Intel already sells most of what you want, it's just spendy. It's called the X25-e and it's an SLC based SSD that will do 170MB/s read and 150MB/s write even at 4KB block sizes. It costs ~$700 for a 32GB drive which is why they aren't all that popular with the ricer desktop set, but if you have a database server that can use a wicked fast log file drive and you don't already have a SAN then it's currently the way to go.
    • MLC brings more density to the table. That's the only reason they do it. Smaller die size and storage density means more MB per dollar

      SLC would be a much smaller capacity drive for the same money. It would be faster at writing, but probably too expensive or too small to have many adopters.

      Same reason SLC is all but unheard-of in thumbdrives. (IronKey being one exception.)

    • I'm trying to snatch up some last-gen Mtron SSDs. The Mobi 3000 looks especially good for the price. They're old and only 16GB, but they're also SLC server SSDs, with about 100MB/sec read and 80MB/sec write.

      Cant wait for them to drop under $100!

    • I think a major hidden problem with MLC is not so much read or write speed, but data integrity. If an MLC chip is storing 4-bits per cell, that's 16 discrete VT levels that need to be detected to resolve the stored info. Couple that with increasingly smaller cell sizes and it would seem to me that even very low levels of gate leakage could lead to bit errors.

      I cringe at the thought of using an MLC based SSD to store important data and then having it basically bit rot due to gate leakage (an effect which i

      • To me, MLC has a conceptual problem of going against the fine tradition of binary computing, which is all about data integrity. Why don't we go back to analog computers for even higher densities, while we're at it.
        • Re:SLC vs MLC (Score:5, Insightful)

          by Kjella ( 173770 ) on Saturday February 14, 2009 @09:19AM (#26855289) Homepage

          To me, MLC has a conceptual problem of going against the fine tradition of binary computing, which is all about data integrity. Why don't we go back to analog computers for even higher densities, while we're at it.

          Says someone that obviously has never seen the raw output of a HDD read head, or the optical laser in a DVD reader. The real world was always very ugly and analog, there's a helluva lot going on to give you a 0 or 1 answer.

        • Re:SLC vs MLC (Score:5, Informative)

          by cnettel ( 836611 ) on Saturday February 14, 2009 @09:30AM (#26855327)
          Do you detest Gigabit Ethernet and disk-based drives as well? Pure binary protocols for signals on media tend to be inefficient. The technology is still digital. Whether data integrity is a priority or not is really a matter of proper error correction, to rely on avoiding on single-bit errors is a flawed strategy.
  • This is unfortunately another case to show that SSDs are not ready for prime-time. With that said, I'm anxiously awaiting the ability to buy a super-fast 120GB+ SLC drive once prices drop below $400.

    I just hope that Microsoft and Apple come up with some great software enhancements for handling SSDs ASAP.

    It is hard for me to believe that the two OS giants can't release their upcoming software in a way that is totally SSD optimized. They are kidding themselves if they don't think that conventional mechanica

    • For some time now all my storage needs are satisfied in their entirety by SSDs and I have no HDDs now. Certainly much better than my previous 10,000 RPM hard disks, so I think they are ready for the prime time.

    • I believe there is so much misconception out there about flash memory performance, it's astonishing. There just isn't a good understanding of how all the layers of cache in the OS work.

      SSD's are not slow, do not "die young". I just built a new system with 3 SSD's in RAID0 and I'm getting 350MB/sec sequential read performance, and nearly 250MB/sec sequential write. In fact, I'm less worried adding additional drives in RAID0, because they fail by total wear, not a single point of a failure, and if the wear is

      • by lmnfrs ( 829146 )

        In a 3 drive RAID 0 array 350MB/s isn't that impressive. That's 116 2/3 MB/s per drive which is only slightly higher than the 1TB SATA spindles that were on sale months ago for less than a SSD. (Those speeds were maximums, though. Min was in the upper 60's I believe.)

        That performance is cool but its price premium is huge for a drive that provides only a fraction of spindles' capacity. Until they drop in price or are handled more efficiently to improve their performance, they may be ready for prime time but

        • The benchmarks I've seen on 1TB drives show about 80MB/sec average, and it goes from 60MB to 100MB/sec depending on the location of the read on the drive, SSD's don't have this non-linearity. Also, the 350MB/sec is reaching the limitations of the RAID controller, a single drive is about 150MB/sec, so it starts scaling down a little when you add more. I will probably try a PCI express adapter that has more bandwidth in the future.

          Still, the main speed boost is in the latency. 0.2ms vs 8ms is a huge leap. Als

  • by Anonymous Coward

    Looking at the big picture, I'd rather have a slow SSD than keep dealing with the data losses of (criminally unreliable) HDs.

  • A boost converter is FAR from new.

    http://en.wikipedia.org/wiki/Boost_converter [wikipedia.org]

    They have not invented a new power supply system. They are just suggesting it be applied to NAND and the high voltage needed be fed into the chip from a central supply, instead of having a charge pump (switching capacitors) in each NAND chip.

    I'm not certain this will save power, but it will reduce peak currents because when the charge pumps switch on in the NAND chips, it creates a huge (but short) current spike. And if you write

  • As people push for smaller laptops with longer battery life and as flash memory continues to drop in price and power requirements and to gain in raw performance, it makes less and less sense for people to use mechanical hard drives in laptops. But as this article shows, the drive's logic can only do so much to try to maintain performance while appearing to the OS to be just a regular hard drive. Using a direct flash interface and a flash file system like UBIFS, YAFFS2, or LogFS should provide Linux netbooks

  • Hardware CAN'T know what areas are used and what aren't. So those hw workarounds for not using a real flash filesystems can't work well.
    I can't understand why those people are still spending money in producing such complex brain damaged products. Give us complete access to flash chips, let the OS do the right thing. Legacy operating systems like Windows? Just give a driver and the flash file system in bundle! I'm sure performance and cost outweight the inconvenience of installing it. Maybe use a replaceable

    • This is what happens when your OS is closed/proprietary.

    • by maxume ( 22995 )

      You do realize that there is software in the hardware that is lying to the software at higher layers, right?

      Even spinning disks present a virtual interface to the hardware.

      • by chiui ( 1120973 )

        You do realize that there is software in the hardware that is lying to the software at higher layers, right?

        Even spinning disks present a virtual interface to the hardware.

        I understand that there IS a layer even on hard disks.
        The problem is that the layer for SSD is currently the wrong one (the hard disk one)
        In fact, the linux layer for MTD devices is completely different and more complex, and flash filesystems are very specific for that layer so you can't use them for normal block devices.
        I don't say you don't need a layer. I say that using the wrong layer and reversing it in hardware is excessively WRONG.

  • This guy presented SSDs and they're are laid out. Seems like quite a bit of unknowns now.

    http://www.youtube.com/watch?v=WcO7xn0wJ2I [youtube.com]

  • Okay, I just ran this benchmark on my 3 RAID0 SSD array....

    From 0.5KB to 128KB write performance (in KB/sec)
    3928
    7368
    12579
    19931
    48306
    83492
    143772
    233510
    252352

    The reason for the low performance on small block sizes is the option called "Direct I/O" on ATTO Disk Benchmark. What this probably does is turns of your system's caching capability, so of course you are going to get ridiculously slow rates. It's good for comparison, but to say you're system is going to be slow because of it is ridiculous because in the r

  • The Intel X25-M series of drives are the best performing MLCs offerings to date.

For God's sake, stop researching for a while and begin to think!

Working...