Subject: General Tech, Storage | September 21, 2014 - 08:41 PM | Scott Michaud
Tagged: ssd, Samsung, kingston hyper x, kingston, endurance, corsair neutron gtx, corsair, 840 pro
Many drives have died over the last year and a bit. The Tech Report has been torturing SSDs with writes until they drop. Before a full petabyte of data was written, three of the six drives kicked the bucket. They are now at 1500TB of total writes and one of the three survivors, the 240GB Corsair Neutron GTX, dropped out. This was a bit surprising as it was reporting fairly high health when it entered "the petabyte club" aside from a dip in read speeds.
The two remaining drives are the Samsung 840 Pro (256GB) and Kingston HyperX 3K (240GB).
Two stand, one fell (Image Credit: Tech Report)
Between those two, the Samsung 840 Pro is given the nod as the Kingston drive lived through uncorrectable errors; meanwhile, the Samsung has yet to report any true errors (only reallocations). Since the test considers a failure to be a whole drive failure, though, the lashings will persist until the final drive gives out (or until Scott Wasson gives up in a glorious sledgehammer apocalypse -- could you imagine if one of them lasted a decade? :3).
Of course, with just one unit from each model, it is difficult to faithfully compare brands with this marathon. While each lasted a ridiculously long time, the worst of the bunch putting up with a whole 2800 full-drive writes, it would not be fair to determine an average lifespan for a given model with one data point each. It is good to suggest that your SSD probably did not die from a defrag run -- but it is still a complete waste of your time and you should never do it.
Investigating the issue
** Edit ** (24 Sep)
We have updated this story with temperature effects on the read speed of old data. Additional info on page 3.
** End edit **
** Edit 2 ** (26 Sep)
New quote from Samsung:
"We acknowledge the recent issue associated with the Samsung 840 EVO SSDs and are qualifying a firmware update to address the issue. While this issue only affects a small subset of all 840 EVO users, we regret any inconvenience experienced by our customers. A firmware update that resolves the issue will be available on the Samsung SSD website soon. We appreciate our customer’s support and patience as we work diligently to resolve this issue."
** End edit 2 **
** Edit 3 **
The firmware update and performance restoration tool has been tested. Results are found here.
** End edit 3 **
Over the past week or two, there have been growing rumblings from owners of Samsung 840 and 840 EVO SSDs. A few reports scattered across internet forums gradually snowballed into lengthy threads as more and more people took a longer look at their own TLC-based Samsung SSD's performance. I've spent the past week following these threads, and the past few days evaluating this issue on the 840 and 840 EVO samples we have here at PC Perspective. This post is meant to inform you of our current 'best guess' as to just what is happening with these drives, and just what you should do about it.
The issue at hand is an apparent slow down in the reading of 'stale' data on TLC-based Samsung SSDs. Allow me to demonstrate:
You might have seen what looks like similar issues before, but after much research and testing, I can say with some confidence that this is a completely different and unique issue. The old X25-M bug was the result of random writes to the drive over time, but the above result is from a drive that only ever saw a single large file write to a clean drive. The above drive was the very same 500GB 840 EVO sample used in our prior review. It did just fine in that review, and at afterwards I needed a quick temporary place to put a HDD image file and just happened to grab that EVO. The file was written to the drive in December of 2013, and if it wasn't already apparent from the above HDTach pass, it was 442GB in size. This brings on some questions:
- If random writes (i.e. flash fragmentation) are not causing the slow down, then what is?
- How long does it take for this slow down to manifest after a file is written?
Subject: Storage | September 18, 2014 - 07:10 PM | Jeremy Hellstrom
Tagged: micron, M600, SLC. MLC, DWA
Micron's M600 SSD has a new trick up its sleeve, called dynamic write acceleration which is somewhat similar to the HDDs with an NAND cache to accelerate the speed frequently accessed data can be read but with a brand new trick. In this case SLC NAND acts as the cache for MLC NAND but it does so dynamically, the NAND can switch from SLC to MLC and back depending on the amount of usage. There is a cost, the SLC storage capacity is 50% lower than MLC so the larger the cache the lower the total amount of storage is available. As well the endurance rating is also higher than previous drives, not because of better NAND but because of new trim techniques being used. This is not yet a retail product so The Tech Report does not have benchmarks but this goes to show you there are plenty more tricks we can teach SSDs.
"Micron's new M600 SSD can flip its NAND cells between SLC and MLC modes on the fly, enabling a dynamic write cache that scales with the drive's unused capacity. We've outlined how this dynamic write acceleration is supposed to impact performance, power consumption, and endurance."
Here are some more Storage reviews from around the web:
- Adata's Premier SP610 @ The Tech Report
- A SCORCHIO fatboy SSD: Samsung SSD850 PRO 3D V-NAND @ The Register
- Silicon Power Blaze B06 64GB USB 3.0 Flash Drive Review @ NikKTech
- SanDisk Ultra II SSD Review (240GB) - TLC Memory becomes Mainstream @ The SSD Review
- Thecus N2310 2-bay NAS @ Kitguru
- QNAP TS-451 @ techPowerUp
- Kingston HyperX FURY 64GB USB 3.0 Flash Drive Review @ OCC
Introduction, Specifications and Packaging
It seems a lot of folks have been incorporating Silicon Motion's SM2246EN controller into their product lines. We first reviewed the Angelbird SSD wrk, but only in a 512GB capacity. We then reviewed a pair of Corsair Force LX's (256GB and 512GB). ADATA has joined the club with their new Premier SP610 product line, and today we are going to take a look at all available capacities of this new model:
It's fortunate that ADATA was able to sample us a full capacity spread, as this will let us evaluate all shipping SSD capacites that exist for the Silicon Motion SM2246EN controller.
Subject: Storage, Shows and Expos | September 16, 2014 - 02:29 PM | Allyn Malventano
Tagged: ssd, slc, sata, mlc, micron, M600, crucial
You may already be familiar with the Micron Crucial M550 line of SSDs (if not, familiarize yourself with our full capacity roundup here). Today Micron is pushing their tech further by releasing a new M600 line. The M600's are the first full lineup from Micron to use their 16nm flash (previously only in their MX100 line). Aside from the die shrink, Micron has addressed the glaring issue we noted in our M550 review - that issue being the sharp falloff in write speeds in lower capacities of that line. Their solution is rather innovative, to say the least.
Recall the Samsung 840 EVO's 'TurboWrite' cache, which gave that drive a burst of write speed during short sustained write periods. The 840 EVO accomplished this by each TLC die having a small SLC section of flash memory. All data written passed through this cache, and once full (a few GB, varying with drive capacity), write speed slowed to TLC levels until the host system stopped writing for long enough for the SSD to flush the cached data from SLC to TLC.
The Micron M600 SSD in 2.5" SATA, MSATA, and M.2 form factors.
Micron flips the 'typical' concept of caching methods on its head. It does employ two different types of flash writing (SLC and MLC), but the first big difference is that the SLC is not really cache at all - not in the traditional sense, at least. The M600 controller, coupled with some changes made to Micron's 16nm flash, is able to dynamically change the mode of each flash memory die *on the fly*. For example, the M600 can place most of the individual 16GB (MLC) dies into SLC mode when the SSD is empty. This halves the capacity of each die, but with the added benefit of much faster and more power efficient writes. This means the M600 would really perform more like an SLC-only SSD so long as it was kept less than half full.
As you fill the SSD towards (and beyond) half capacity, the controller incrementally clears the SLC-written data, moving that data onto dies configured to MLC mode. Once empty, the SLC die is switched over to MLC mode, effectively clearing more flash area for the increasing amount of user data to be stored on the SSD. This process repeats over time as the drive is filled, meaning you will see less SLC area available for accelerated writing (see chart above). Writing to the SLC area is also advantageous in mobile devices, as those writes not only occur more quickly, they consume less power in the process:
For those worst case / power user scenarios, here is a graph of what a sustained sequential write to the entire drive area would look like:
Realize this is not typical usage, but if it happened, you would see SLC speeds for the first ~45% of the drive, followed by MLC speeds for another 10%. After the 65% point, the drive is forced to initiate the process of clearing SLC and flipping dies over to MLC, doing so while the host write is still in progress, and therefore resulting in the relatively slow write speed (~50 MB/sec) seen above. Realize that in normal use (i.e. not filling the entire drive at full speed in one go), garbage collection would be able to rearrange data in the background during idle time, meaning write speeds should be near full SLC speed for the majority of the time. Even with the SSD nearly full, there should be at least a few GB of SLC-mode flash available for short bursts of SLC speed writes.
This caching has enabled some increased specs over the prior generation models:
Note the differences in write speeds, particularly in the lower capacity models. The 128GB M550 was limited to 190MB/sec, while the M600 can write at 400MB/sec in SLC mode (which is where it should sit most of the time).
We'll be testing the M600 shortly and will come back with a full evaluation of the SSD as a whole and more specifically how it handles this new tech under real usage scenarios.
Subject: Storage, Shows and Expos | September 16, 2014 - 12:49 PM | Allyn Malventano
Tagged: ram, NVMe, IOPS, idf 2014, idf, ddr4, DDR
The Intel Developer Forum was last week, and there were many things to be seen for sure. Mixed in with all of the wearable and miniature technology news, there was a sprinkling of storage goodness. Kicking off the show, we saw new cold storage announcements from both HGST and Western Digital, but that was about it for HDD news, as the growing trend these days is with solid state storage technologies. I'll start with RAM:
First up was ADATA, who were showing off 64GB DDR3 (!) DIMMs:
Next up were various manufacturers pushing DDR4 technology quite far. First was SK Hynix's TSV 128GB DIMMs (covered in much greater depth last week):
Next up is Kingston, who were showing a server chassis equipped with 256GB of DDR4:
If you look closer at the stats, you'll note there is more RAM in this system than flash:
Next up is IDT, who were showing off their LRDIMM technology:
This technology adds special data buffers to the DIMM modules, enabling significantly higher amounts of installed RAM into a single system, with a 1-2 step de-rating of clock speeds as you take capacities to the far extremes. The above server has 768GB of DDR4 installed and running!:
Moving onto flash memory type stuff, Scott covered Intel's new 40 Gbit Ethernet technology last week. At IDF, Intel had a demo showing off some of the potential of these new faster links:
This demo used a custom network stack that allowed a P3700 in a local system to be matched in IOPS by an identical P3700 *being accessed over the network*. Both local and networked storage turned in the same 450k IOPS, with the remote link adding only 8ms of latency. Here's a close-up of one of the SFF-8639 (2.5" PCIe 3.0 x4) SSDs and the 40 Gbit network card above it (low speed fans were installed in these demo systems to keep some air flowing across the cards):
Stepping up the IOPS a bit further, Microsoft was showing off the capabilities of their 'Inbox AHCI driver', shown here driving a pair of P3700's at a total of 1.5 million IOPS:
...for those who want to get their hands on this 'Inbox driver', guess what? You already have it! "Inbox" is Microsoft's way of saying the driver is 'in the box', meaning it comes with Windows 8. Bear in bind you may get better performance with manufacturer specific drivers, but it's still a decent showing for a default driver.
Now for even more IOPS:
Yes, you are reading that correctly. That screen is showing a system running over 11 million IOPS. Think it's RAM? Wrong. This is flash memory pulling those numbers. Remember the 2.5" P3700 from a few pics back? How about 24 of them:
The above photo shows three 2U systems (bottom), which are all connected to a single 2U flash memory chassis (top). The top chassis supports three submodules, each with eight SFF-8639 SSDs. The system, assembled by Newisys, demonstrates just how much high speed flash you can fit within an 8U space. The main reason for connecting three systems to one flash chassis is because it takes those three systems to process the full IOPS capability of 24 low latency NVMe SSDs (that's 96 total lanes of PCIe 3.0!)!
So there you have it, IDF storage tech in a nutshell. More to come as we follow these emerging technologies to their maturity.
Subject: Storage | September 12, 2014 - 05:30 PM | Jeremy Hellstrom
Tagged: SM2246EN, S3C, mlc, Apotop
The Apotop S3C SSD uses the same controller as the Angelbird drive Al reviewed recently. It uses synchronous MLC NAND with the 4 channel present on the Silicon Motion controller and is able to provide more than the specified 490 MB/s read and 275 MB/s write in some benchmarks. It can often read faster than the wrk SSD but the writes cannot always keep up though it is not something likely to be noticeable in real usage scenarios. The MSRP is very attractive with the 512GB model expected to be released at $200. Silicon Motion is likely to start appearing in a lot more SSDs in the near future with this mix of price and performance. Read the full review at Kitguru.
"The new Apotop S3C SSD features the Silicon Motion 2246EN controller which we first reviewed in the Angelbird 512GB wrk SSD back in August this year. The controller impressed us, so we have already high hopes for the Apotop S3C."
Here are some more Storage reviews from around the web:
- PNY XLR8 120 GiB SSD Review @ Hardware Secrets
- Plextor M6 Pro 256GB @ Kitguru
- Plextor M6 PRO 256GB @ eTeknix
- We lift the lid on Intel's Pro 2500 SSD. Shock, horror: It doesn't use its own NAND chips @ The Register
- HGST Ultrastar He6 6TB SAS HDD Review @ NikKTech
- TB (2.5-inch) Hard Drives @ SPCR
- SPYRUS WorkSafe Pro WTG Secure Flash Drive @ The SSD Review
- Synology DiskStaion DS115j @ Legion Hardware
- QNAP TurboNAS TS-451 NAS Server Review @ NikKTech
- Icy Box FlexCage MB973SP 2B Trayless 3-in-2 SATA Backplane @ Kitguru
- Thermaltake BlacX Urban Wi-Fi Edition HDD Docking Station Review @ TechwareLabs
- Angelbird SSD2go Pocket External SSD @ The SSD Review
Subject: General Tech, Storage | September 12, 2014 - 04:08 PM | Scott Michaud
Tagged: sandisk, sdxc, sdhc, sd card, 512GB
Assuming your camera, card reader, or other device fully conforms to the SDXC standard, Sandisk has developed a half-terabyte (512GB) memory card. Beyond being gigantic, it can be read at up to 95 MB/s and written at up to 90 MB/s, which should be enough to stream 4K video. Sandisk claims that it is temperature proof, shock proof, water proof, and x-ray proof. It also comes with a lifetime warranty and "RescuePRO Deluxe" recovery software but, honestly, I expect people would just use PhotoRec or something.
It should be noted that the SDXC standard covers memory cards up to 2TB so it will probably not be too long before we see another standard get ratified. What is next? SDUC? SDYC? SDALLTHEC? Blah! This is why IEEE assigns names sequentially.
The SanDisk Extreme PRO UHS-I SDHC/SDXC 512GB memory card should be available now, although I cannot yet find them online, for $799.99 MSRP.
Subject: Storage, Shows and Expos | September 10, 2014 - 03:34 PM | Allyn Malventano
Tagged: TSV, Through Silicon Via, memory, idf 2014, idf
If you're a general computer user, you might have never heard the term "Through Silicon Via". If you geek out on photos of chip dies and wafers, and how chips are assembled and packaged, you might have heard about it. Regardless of your current knowledge of TSV, it's about to be a thing that impacts all of you in the near future.
Let's go into a bit of background first. We're going to talk about how chips are packaged. Micron has an excellent video on the process here:
The part we are going to focus on appears at 1:31 in the above video:
This is how chip dies are currently connected to the outside world. The dies are stacked (four high in the above pic) and a machine has to individually wire them to a substrate, which in turn communicates with the rest of the system. As you might imagine, things get more complex with this process as you stack more and more dies on top of each other:
16 layer die stack, pic courtesy NovaChips
...so we have these microchips with extremely small features, but to connect them we are limited to a relatively bulky process (called package-on-package). Stacking these flat planes of storage is a tricky thing to do, and one would naturally want to limit how many of those wires you need to connect. The catch is that those wires also equate to available throughput from the device (i.e. one wire per bit of a data bus). So, just how can we improve this method and increase data bus widths, throughput, etc?
Before I answer that, let me lead up to it by showing how flash memory has just taken a leap in performance. Samsung has recently made the jump to VNAND:
By stacking flash memory cells vertically within a die, Samsung was able to make many advances in flash memory, simply because they had more room within each die. Because of the complexity of the process, they also had to revert back to an older (larger) feature size. That compromise meant that the capacity of each die is similar to current 2D NAND tech, but the bonus is speed, longevity, and power reduction advantages by using this new process.
I showed you the VNAND example because it bears a striking resemblance to what is now happening in the area of die stacking and packaging. Imagine if you could stack dies by punching holes straight through them and making the connections directly through the bottom of each die. As it turns out, that's actually a thing:
Subject: Storage, Shows and Expos | September 9, 2014 - 04:51 PM | Allyn Malventano
Tagged: WDC< Western Digital, WD, idf 2014, idf, hdd, Cold, Archival, Ae
We talked about helium filled, shingled HDD's from HGST earlier today. Helium may give you reduced power demands, but at the added expensive of hermetically sealed enclosures over conventional HDD's. Shingling may give added capacity, but at the expense of being forced into specific writing methods. Now we know Western Digital's angle into archival / cold storage:
..so instead of going with higher cost newer technologies, WD is taking their consumer products and making them more robust. They are also getting rid of the conventional thinking of capacity increments and are moving to 100GB increments. The idea is that once a large company or distributor has qualified a specific HDD model on their hardware, that model will stick around for a while, but be continued at an increased capacity as platter density yields increase over time. WD has also told me that capacities may even be mixed an matched within a 20-box of drives, so long as the average capacity matches the box label. This works in the field of archival / cold storage for a few reasons:
- Archival storage systems generally do not use conventional RAID (where an entire array of matching capacity disks are spinning simultaneously). Drives are spun up and written to individually, or spun up individually to service the occasional read request. This saves power overall, and it also means the individual drives can vary in capacity with no ill effects.
- Allowing for variable capacity binning helps WD ship more usable platters/drives overall (i.e. not rejecting drives that can't meet 6TB). This should drive overall costs down.
- Increasing capacity by only a few hundred GB per drive turns into *huge* differences in cost when you scale that difference up to the number of drives you would need to handle a very large total capacity (i.e. Exabytes).
So the idea here is that WD is choosing to stick with what they do best, which they can potentially do for even cheaper than their consumer products. That said, this is really meant for enterprise use and not as a way for a home power user to save a few bucks on a half-dozen drives for their home NAS. You really need an infrastructure in place that can handle variable capacity drives seamlessly. While these drives do not employ SMR to get greater capacity, that may work out as a bonus, as writes can be performed in a way that all systems are currently compatible with (even though I suspect they will be tuned more for sequential write workloads).
Here's an illustration of this difference:
The 'old' method meant that drives on the left half of the above bell curve would have to be sold as 5TB units.
With the 'new' method, drives can be sold based on a spec closer to their actual capacity yield. For a given model, shipping capacities would increase as time goes on (top to bottom of the above graphic).
To further clarify what is meant by the term 'cold storage' - the data itself is cold, as in rarely if ever accessed:
Examples of this would be Facebook posts / images from months or years ago. That data may be rarely touched, but it needs to be accessible enough to be browsed to via the internet. The few second archival HDD spinup can handle this sort of thing, while a tape system would take far too long and would likely timeout that data request.