Performance Showdown - SSDs vs. HDDs 259
Lucas123 writes "Computerworld compared four disks, two popular solid state drives and two Seagate mechanical drives, for read/write performance, bootup speed, CPU utilization and other metrics. The question asked by the reviewer is whether it's worth spending an additional $550 for a SSD in your PC/laptop or to plunk down the extra $1,300 for an SSD-equipped MacBook Air? The answer is a resounding No. From the story: "Neither of the SSDs fared very well when having data copied to them. Crucial (SSD) needed 243 seconds and Ridata (SSD) took 264.5 seconds. The Momentus and Barracuda hard drives shaved nearly a full minute from those times at 185 seconds. In the other direction, copying the data from the drives, Crucial sprinted ahead at 130.7 seconds, but the mechanical Momentus drive wasn't far behind at 144.7 seconds."
bad test (Score:5, Insightful)
Re:bad test (Score:5, Interesting)
Skimming the article, it seems very likely that the person responsible has read just enough to be dangerous (they know the physics of why seeking is slow), but not enough to have a clue what kind of behavior would trigger seeking. The one measure was boot time, during which they acknowledge that Vista does a bunch of background stuff after boot, but don't measure it.
He did get one thing right, though -- they are not exactly living up to their potential. For one thing, there are filesystems explicitly designed for flash media, but you need to actually access it as flash (and the filesystem does its own wear leveling) -- these things pretend to be a hard disk, and are running filesystems optimized for a hard disk, so the results are not going to be at all what they could be.
Re: (Score:3, Funny)
Re: (Score:2)
Re:bad test (Score:5, Interesting)
XP IO subsystem is pretty OK.
The problem with SSD is that flash based storage has much much higher block size.
While conventional HDDs have block size 512 bytes, actual SSDs have block size of 64 kilobytes.
Not only Flashes write relatively slow, but if file system has e.g. cluster size of 8K, every write to it in worst case would also (re)write redundantly 64K-8K=56K.
Test is realistic - if you want to see how bad most applications can be with SSDs. But that's going to change with SSD becoming more and more common place.
If they really wanted to test SSD performance they would have taken Linux with jffs2 or newer logfs. Though this two have their own problems.
Solutions for hybrid disk setups? (Score:3, Interesting)
If they really wanted to test SSD performance they would have taken Linux with jffs2 or newer logfs.
Does anybody have a decent solution for using a flash drive to boost performance of a regular drive?
I just ordered a new laptop, and it has an ExpressCard slot into which I could drop 4 or 8 GB of solid-state disk at a reasonable price. That could serve as a giant cache, one that unlike RAM could be safely used as a write cache.
It seems like there would be a clever way to treat the SSD plus the regular hard drive as one unit so that the hard drive could be spun down for hours of normal working situations,
Untested performance... (Score:4, Interesting)
Re:Untested performance... (Score:4, Interesting)
Definitely not enough data (Score:2)
Although some data from the Palm LifeDrive (featuring a mecanical Microdrive CF module) could answer the drop-survivability in small form factor.
So, in short, they managed to produce only 1 single data i.e. bulk speed (well, not exactly. They also mentioned random access from a synthetic test, but no actual real-world application) when users would need ab
Re:And some drop tests, and airport scaner tests, (Score:4, Informative)
1) Mine have been formatted NTFS, running Windows XP (and additionally Apple HFS Journaled recently when experimenting with OS X). I do not defragment the SSD, there is no point. Read speeds have always been better than write speed, but I see no difference in performance over time.
2) Both of the drives I have are fully functional, even though I abused the 32GB one mercilessly. That laptop has only 1GB of RAM and I would run so many programs that things were swapping constantly for the past year.
3) The 32GB SSD has been through airport scanners approximately 50 times now, no problems. The 64GB is too new, only travelled a few times so far.
4) My laptops are always on the go, brought into many factories as a consultant. While in my bag it has taken falls down sets of stairs. The laptop itself (a Fujitsu P1610) has been dropped from a height of 3.5 to 4 feet onto a metal catwalk while running with no adverse affects (other than a few scuffs and dents on the corners).
5) Not sure how well they stand up to static, but it has stood up well to a variety of high EM fields, and high/low temperatures. No data loss. I have had regular hard disks die from working next to large transformers (and their magnetic fields) for an afternoon.
Hope that helps you. For my line of work, they have been incredible. I used to go through 3 or 4 laptop hard disks per year due to various issues. Now the only reason I bought the 64GB SSD is increased storage capacity.
Re: (Score:2)
Is file fragmentation really that big of a problem?
I know at one time I used to defragment a lot, but the difference has always been negligible for me. I only did it with the thought of keeping it "in tune", but even once a year doesn't make much apparent difference in computer performance.
Re: (Score:2)
Re: (Score:2)
If objective tests show no real advantage, then any subjective improvement in value is basically self-delusion... probably in an attempt to justify the extra cost to yourself.
If the test DO show an improvement, then I agree it is more up to the customer to determine if the extra cost is worth the extra performance.
=Smidge=
Re: (Score:3, Interesting)
Noise? Heat? (Score:3, Interesting)
Re: (Score:2)
Those Who Write (Score:2)
Re: (Score:2)
Re: (Score:2)
I haven't used any swapspace for years on my desktops, memory is so cheap now that there's no point. On my servers, of course, but then again it's 99.9% unused.
For example, this thinkpad has 1.25GB RAM, and I've seen at most 300MB used. Then again, I don't run Vista.
Re: (Score:3, Funny)
I haven't used any swapspace for years on my desktops, memory is so cheap now that there's no point. On my servers, of course, but then again it's 99.9% unused.
For example, this thinkpad has 1.25GB RAM, and I've seen at most 300MB used. Then again, I don't run Vista.
Buzz (Score:2)
Not very good reasons... (Score:5, Insightful)
But of course not the metrics that really matter, which SSD's vastly excel at and make them worth the price for many people: MTBF, power consumption, ruggedness and noise level.
Re:Not very good reasons... (Score:5, Insightful)
I know it's not a car analogy, I humbly beg the forgiveness of the
Re: (Score:2)
Re: (Score:2)
Re: (Score:2)
Re: (Score:2)
Similar story over at StorageMojo [storagemojo.com] and Robin draws a similar conclusion.
MTBF - Infant failures about the same as discs, return rates higher
Power - Flash already near the bottom of the power curve, drives appear to have room to drop
Ruggedness - No moving parts a plus, perhaps countered by whole-block rewrites on write. Not enough data here
Noise - Flash wins, no contest
Bottom line? Not enough improvement to justify the cost, except in certain
Noise level on new 2.5" drives ... (Score:2)
I can't hear the hd on my laptop, and I rarely hear the fan. The newest 2.5" drives are super-quiet.
Power Consumption (Score:4, Insightful)
Re: (Score:3, Interesting)
Performance is not the key to SSD (Score:3, Insightful)
But on the performance front, they compared with 7200RPM hard drives, last time I checked (admittedly a while ago) most laptop are outfitted with 5400RPM drives.
Re:Performance is not the key to SSD (Score:4, Insightful)
Re: (Score:2)
Another concern, which I forgot from my original post is heat. And I am sure heat is key concern with a laptop like the Mac Air
Re: (Score:2)
The stock hdd is 4200rpm so even that 5400 figure you had was over the stock drive speed. So they should have compared those two options as well as what they did to get a good idea. As well as including drives with 8mb and
Why a "drive"? (Score:4, Interesting)
Re: (Score:2)
Re:Why a "drive"? (Score:4, Informative)
Re: (Score:3, Interesting)
Re: (Score:2)
Stupid Test (Score:5, Informative)
SSD works best when accessing files randomly.
Re: (Score:2)
Rex: Go back, go back, you missed it.
Hamm: Too late, I'm already on the 40's, gotta go around the horn, it's faster.
Re: (Score:3, Insightful)
This is like a hybrid vehicle vs normal gas shootout, with each vehicle towing something. It's irrelevant.
He boiled down all the variables and performance profiles into just one - the one that favors traditional drives. There is NO WAY this should have been published as-is.
I can't attribute this to malice, but basically Bill O'Brien of Computerworld DOESN'T KNOW WHAT HE'S DOING, and neither does his editor for letting this slide. This was probably a case of a traditional drive maker whispering in his ear
I've seen this before. (Score:3, Insightful)
Now a good purpose for these might be in desktop bound short-stack storage arrays instead of that large tera-byte drive array. They're just quick enough for data retention backups off of the mechanical drives in the client PC.
Another use is small-scale server apps that usually are bound into hardware in some form of internet controllable appliance. Speed isn't really a major factor here for this and these would potentially work well.
Just my opinion. Subject to change.
SSDs are ideal for servers (Score:5, Insightful)
According to the article the Crucial SSD has an access time of 0.4 ms which equates to 2500 IOs/s as compared to the Barracuda HDD with 13.4 ms access time which equates to a mere 75 IOs/s.
So for servers SSDs are 33 times better!
Bring them on
Re: (Score:2)
Re: (Score:2)
Re: (Score:2, Interesting)
If your filesystem is designed to distribute the writes properly, the failure time is comparable to the MBF of hard drives.
Though personally, I think the way to go on servers is to use 64GB of RAM and put most of it as a RAM disk. Depending on your application, you can either have a shell script copy the data back to a hard drive for persistent data, or use that kernel driver to mirror the data to a hard drive. Software RAID 1 would work, too.
Re: (Score:2)
Re: (Score:2)
If your 2500 IOs/s hit the same sector, your server SSD is fried in 7 min. SSD are distinctively NOT server suitable if you have a lot of write cycles (probably less of an issue if it's just answering read requests).
Re: (Score:2)
One would think that would actually be an ideal scenario. Your cache hits would be through the roof. Even if it wrote the sector back to the flash drive once every 2 seconds, that would be 5000 IO's worth of updates in one write op.
Factor in drive wear leveling (so that it moves the data sector around on the empty space on the physical disk rather than in the same physical place each time), and the disk would probably last nearly for
Re: (Score:2, Informative)
I don't know why everyone keeps repeating flash "problems" from the mid/late 90s. We're in 2008, flash has been widely used in h
Re: (Score:2)
More specifically, I'm talking about CompactFlash cards, since they can be used as an IDE drive with a simple 10$ adapter.
Re: (Score:2)
Similar for using it as a specialty device for read heavy applications; it was the general "ideal server device" that I had a problem with.
When I first read about SSD it sounded like the second coming of sliced bread, it was the "devil in the detail" that soured me, especially the write limitations that seem to be a physical limitation, not something you can eng
Re: (Score:2)
A good example where an SSD might be a good solution is one of the database servers at my office.
A record gets created and then updated maybe twice. It then my get read a few hundred thousand times.
So yes for some servers it might be a really good thing. Lots of databases are very very read heavy and write light.
Re: (Score:2, Insightful)
A modern SSD is able to handle write intensive database application with reliability on par with HDDs. The SSD logic spreads the writes around the disk to prevent premature wear so a record updated a million times might well never be written twice over any given flash cells. And even at 512 bytes each, there's 195 millions of them on a 100GB SSD. Each of which has about a 1 million cycle life and there's normally spare cells to handle failures. I'll take that over a SCSI disk.
Re: (Score:2)
Re: (Score:2)
What about the power usage? (Score:2)
Re: (Score:3, Informative)
Re: (Score:2)
it's all about the battery (Score:2)
SSD's performance boost is in battery life due to its lower power consumption from zero moving parts. Flash-based storage has always had a problem with writing; don't forget about the fact that it can only be written to ~1000 times.
Furthermore, SSD is just temporary relief for batteries; I envision a laptop with both SSD and HDD that almost never writes to the SSD; on Windows, C:\WINDOWS and C:\Program Files would live in SSD while C:\Documents & Settings would live on HDD and C:\WINDOWS\Temp (or whe
Re: (Score:2)
Swap is the main concern here - the solution is to give the machine enough RAM that you can turn swap off.
Re: (Score:2)
With a 64 gigabyte drive with a block size of 256 ki
Re: (Score:2)
Not surprising or bad to me. (Score:3, Informative)
What about battery life? (Score:2)
Admittedly the article described itself as a performance showdown, but I'm disappointed that the reviewer made no attempt to compare power consumption and battery life.
If nothing else, I would have thought a solid state drive would eliminate that annoying pause when a hard drive awakes from sleep and spins up, and that this would feel like a worthwhile "
Re: (Score:2)
Re: (Score:2)
http://www.codinghorror.com/blog/archives/000562.html [codinghorror.com] is the link I used for a few of those figures
Seek Times Make the Difference (Score:5, Interesting)
- p
Vista for performance testing? (Score:2)
That just seems silly. I'd like to see performance tests on a system where the disk's performance affects the end result, rather than all of the results being homogenized by the operating system's poor I/O capability. Given Vista's adoption, it's not even a test of what disk performance will be like "in the real world."
What the reviewer did wrong. (Score:2)
When asked about ignoring the 20:1 advantage SSDs have in seek speed, responded:
But keep in mind that it's only one component of the overall operation. These were all freshly formatted drives so fragmentation shouldn't be an issue and the longer the operation under that condition, the less it tends to matter.
SSDs might even slow down slightly because some are built intelligently enough to not write to the same location each time (and thus prematurely "wear out" segments of memory which are, after all, limited use within context).
Anything else to add?
Why are the results so bad? (Score:2)
BOOT UP TIMES (IN SECONDS)
Yeah I know synthetic tests are problematic, but the two tests gives contrary results.
Is it because a MS Vista boot and reboot doesn't involve much random R/W and therefore doesn't shows the appearrent strength of SSD's? Or is it because an extremely lo
Is Limited Number of Writes to SSDs a Problem? (Score:2)
However, in real-world situations, do SSD write limitations ever pose a problem or is it a total non-issue these days?
Ron
Reliability and shock resistance (Score:2)
The real improvement - resource contention (Score:2)
Just think back to when you moved to a dual-core CPU how much more responsive it was. Now take that same jump to I/O, which is always the performance bottleneck. We're leaving the age of simple increases in horsepower - Mhz, RPM, and throu
Flash memory not true SSD tech (Score:2)
These flash based drives are l
Performance RAID (Score:2)
If someone could put together a convenient RAID type package, the extra cost might actually result in extra, noticable speed improvements, even for writes. And two 64GB SSD units arranged in a performance RAID package would give a more usable 128GB "hard disk" to store things on anyway.
MTBF (Score:2)
All I care about is MTBF. I am so sick and tired of trying to get data off of crashed drives and restoring computers for family members (and myself) . Even with current backups, it's a hassle and disks fail at the most inconvenient time.
My wife wanted a laptop recently and I made her spend the extra money for an SSD.
Re: (Score:2)
All I care about is MTBF. I am so sick and tired of trying to get data off of crashed drives and restoring computers for family members (and myself) . Even with current backups, it's a hassle and disks fail at the most inconvenient time.
I have had much more frequent and less predictable failures from flash drives than from hard drives. Granted it is not a completely fair comparison since the flash drives were being used in portable applications (i.e. cameras) and the hard drives were sitting comfortably in my computer case. However I am less than impressed with what I have seen of flash drive reliability. I have been using the same Seagate hard drive for a system drive for the last 10 years. Can any frequently used flash drive ever hope t
Native command queuing with newer SATA drives (Score:2)
I recently built myself a new system. The new processor (Xeon E3110, aka Core 2 Duo E8400) certainly did make boot time somewhat faster, but not dramatically so. Likewise for initial login -- the KDE desktop came up somewhat faster, but it wasn't overwhelming.
Then it occurred to me to move my root and home directory partitions from an older 250 GB 1.5 Gb/sec SATA drive to my newer 500 GB 3.0 Gb/sec compatible drive. There
Test protocol (Score:2)
Crappy SSDs to test (Score:2)
Speed? WTF? (Score:2)
SSD is supposed to be about power savings, which should be one of your top priorities when designing a portable device (see also Nintendo GameBoy vs Sega GameGear).
What about Altitude (Score:2)
Using SSDs on our portables, we can now go to extreme altitudes in unpressureized enclosures without the fear of drive failure.
Bill
How Flash SSDs Work and How They Can Work Better (Score:4, Informative)
The reason that Flash SSDs act "wierd" in benchmarks is that they have asymmetric performance patterns when reading and writing. Particularly with random operations, this asymmetry is huge. Here are a couple of example "drives":
* Mtron 7000 series: >14,000 4K random reads. ~130 4K random writes.
* SanDisk 5000 series: ~7,000 4K random reads. 13 4K random writes.
* Cheap CF card or USB stick: ~2,500 4K random reads. 3.3 4K random writes.
This is a 100:1 performance deficit when doing random writes versus the random reads. This has some really weird impacts on system performance. For example, if you run Outlook and tell it to "index" your system, it will build a 1-4 GB index file in-place with 100% random writes. If you do this on a hard disk, the job takes a long time and drags down your laptop, but the operation is still pretty smooth. Do the same think on an SSD and the system slugs to molasses. One of our customers described it as "totally unusable" with 2+ minutes to bring up task manager. What happens is that the fast reads allow the application to dirty write buffer faster and this then swamps system RAM, you get a 100+ deep write queue (at 13/sec), and you want to throw the machine off of a bridge.
This fix as some have described it is not some magic new controller glue or putting the flash closer to the CPU. It is organizing the write patterns to more closely match what the Flash chips are good at. Numerous embedded file systems like JFFS do this, but they are really designed for very small devices and are more concerned with wear and lifespan issue than performance.
Now here comes the advert (flames welcome). A little over 2 years ago, I wrote a "block translation" layer for use with Flash storage devices. It is somewhat similar to a LogFS, but it is not really a file system and it does not play be all of the rules of a LogFS. It does however remap blocks and linearize writes. Thus it plays well with Flash. It also appears to be an "invention", and thus my patent lawyer is well paid.
The working name of the driver layer itself is "Fast Block Device" (fbd) and the marketing name is "Manged Flash Technology". And what this does is to transparently map one block device into another view. You can then put whatever file system you want into the mix.
In terms of performance, it is all about bandwidth. Build a little raid-5 array with 4 Mtron drives and you will get over 200 MB/sec of sustained write throughput. With MFT in place, this directly translates into 50,000 4K random writes/sec. Even better, you tend to end up with something that is much closer to symmetric in terms of random read/write performance.
MFT is production on Linux (it has actually been shipping since last summer) and is in Beta test on Windows. It works with single drives as well as small to medium sized arrays. It does work with large arrays, but the controllers don't tend to keep up with the drives, so large arrays are useful for capacity but don't really help performance a lot. Once you get to 50,000 IOPS it is hard for the controllers to go much faster.
Consumer testing with MFT tends to produce some laughable results. We ran PCMark05's disk test on it and produced numbers in the 250K range. This was with a single Mtron 3025. Our code is fast, but we fooled the benchmark in this case.
There are several white papers on MFT posted in the news link of our website:
http://managedflash.com/ [managedflash.com]
My apologies for the advert, but I see a lot of talk about SSDs without actually knowing what is going on inside.
I am happy to answer any questions on-line of off.
Doug Dumitru
EasyCo LLC
610 237-2000 x43
http://easyco.com/ [easyco.com]
http://managedflash.com/ [managedflash.com]
http://mtron.easyco.com/ [easyco.com]
2nd Gen SSD benchmarks. (Score:3, Informative)
The second generation SSDs would cost you more than a whole notebook, but have significant performance improvements:
Memoright GT vs Mtron vs Raptor vs Seagate [benchmarkreviews.com]
Memoright nails it. It is easily twice as fast as what Mac puts in their notebooks.
If you *really* want an SSD, buy one separately and install it yourself. You will not be disappointed.
BTW the file indexing that causes SSDs to slow cause HDDs to slow as well. Many people have reported unbearable slowdown, and that is with HDDs. I am sure anything slower than that would make you want to return the whole thing, but this can be fixed. Most people will tell you to just turn it off [4sysops.com]. Google has also complainted about Microsoft pre-installing an indexing system that sucks [nwsource.com].
Re: (Score:2)
Re: (Score:3, Informative)
Apples to Oranges (Score:2)
Re: (Score:3, Insightful)
Re: (Score:2)
Re: (Score:2)
Maybe if your top professionals got some counseling, they wouldn't be tossing their laptops all over the place. Just who do you work for anyway? Microsoft?
Re: (Score:2)
Or not.
Re: (Score:2)
Re: (Score:2)
Re: (Score:2)
Re: (Score:2)
Re: (Score:2)
Re: (Score:2)
FUD (Score:3, Informative)