Subject: Storage, Shows and Expos | September 16, 2014 - 02:29 PM | Allyn Malventano
Tagged: ssd, slc, sata, mlc, micron, M600, crucial
You may already be familiar with the Micron Crucial M550 line of SSDs (if not, familiarize yourself with our full capacity roundup here). Today Micron is pushing their tech further by releasing a new M600 line. The M600's are the first full lineup from Micron to use their 16nm flash (previously only in their MX100 line). Aside from the die shrink, Micron has addressed the glaring issue we noted in our M550 review - that issue being the sharp falloff in write speeds in lower capacities of that line. Their solution is rather innovative, to say the least.
Recall the Samsung 840 EVO's 'TurboWrite' cache, which gave that drive a burst of write speed during short sustained write periods. The 840 EVO accomplished this by each TLC die having a small SLC section of flash memory. All data written passed through this cache, and once full (a few GB, varying with drive capacity), write speed slowed to TLC levels until the host system stopped writing for long enough for the SSD to flush the cached data from SLC to TLC.
The Micron M600 SSD in 2.5" SATA, MSATA, and M.2 form factors.
Micron flips the 'typical' concept of caching methods on its head. It does employ two different types of flash writing (SLC and MLC), but the first big difference is that the SLC is not really cache at all - not in the traditional sense, at least. The M600 controller, coupled with some changes made to Micron's 16nm flash, is able to dynamically change the mode of each flash memory die *on the fly*. For example, the M600 can place most of the individual 16GB (MLC) dies into SLC mode when the SSD is empty. This halves the capacity of each die, but with the added benefit of much faster and more power efficient writes. This means the M600 would really perform more like an SLC-only SSD so long as it was kept less than half full.
As you fill the SSD towards (and beyond) half capacity, the controller incrementally clears the SLC-written data, moving that data onto dies configured to MLC mode. Once empty, the SLC die is switched over to MLC mode, effectively clearing more flash area for the increasing amount of user data to be stored on the SSD. This process repeats over time as the drive is filled, meaning you will see less SLC area available for accelerated writing (see chart above). Writing to the SLC area is also advantageous in mobile devices, as those writes not only occur more quickly, they consume less power in the process:
For those worst case / power user scenarios, here is a graph of what a sustained sequential write to the entire drive area would look like:
Realize this is not typical usage, but if it happened, you would see SLC speeds for the first ~45% of the drive, followed by MLC speeds for another 10%. After the 65% point, the drive is forced to initiate the process of clearing SLC and flipping dies over to MLC, doing so while the host write is still in progress, and therefore resulting in the relatively slow write speed (~50 MB/sec) seen above. Realize that in normal use (i.e. not filling the entire drive at full speed in one go), garbage collection would be able to rearrange data in the background during idle time, meaning write speeds should be near full SLC speed for the majority of the time. Even with the SSD nearly full, there should be at least a few GB of SLC-mode flash available for short bursts of SLC speed writes.
This caching has enabled some increased specs over the prior generation models:
Note the differences in write speeds, particularly in the lower capacity models. The 128GB M550 was limited to 190MB/sec, while the M600 can write at 400MB/sec in SLC mode (which is where it should sit most of the time).
We'll be testing the M600 shortly and will come back with a full evaluation of the SSD as a whole and more specifically how it handles this new tech under real usage scenarios.
Subject: Storage, Shows and Expos | September 16, 2014 - 12:49 PM | Allyn Malventano
Tagged: ram, NVMe, IOPS, idf 2014, idf, ddr4, DDR
The Intel Developer Forum was last week, and there were many things to be seen for sure. Mixed in with all of the wearable and miniature technology news, there was a sprinkling of storage goodness. Kicking off the show, we saw new cold storage announcements from both HGST and Western Digital, but that was about it for HDD news, as the growing trend these days is with solid state storage technologies. I'll start with RAM:
First up was ADATA, who were showing off 64GB DDR3 (!) DIMMs:
Next up were various manufacturers pushing DDR4 technology quite far. First was SK Hynix's TSV 128GB DIMMs (covered in much greater depth last week):
Next up is Kingston, who were showing a server chassis equipped with 256GB of DDR4:
If you look closer at the stats, you'll note there is more RAM in this system than flash:
Next up is IDT, who were showing off their LRDIMM technology:
This technology adds special data buffers to the DIMM modules, enabling significantly higher amounts of installed RAM into a single system, with a 1-2 step de-rating of clock speeds as you take capacities to the far extremes. The above server has 768GB of DDR4 installed and running!:
Moving onto flash memory type stuff, Scott covered Intel's new 40 Gbit Ethernet technology last week. At IDF, Intel had a demo showing off some of the potential of these new faster links:
This demo used a custom network stack that allowed a P3700 in a local system to be matched in IOPS by an identical P3700 *being accessed over the network*. Both local and networked storage turned in the same 450k IOPS, with the remote link adding only 8ms of latency. Here's a close-up of one of the SFF-8639 (2.5" PCIe 3.0 x4) SSDs and the 40 Gbit network card above it (low speed fans were installed in these demo systems to keep some air flowing across the cards):
Stepping up the IOPS a bit further, Microsoft was showing off the capabilities of their 'Inbox AHCI driver', shown here driving a pair of P3700's at a total of 1.5 million IOPS:
...for those who want to get their hands on this 'Inbox driver', guess what? You already have it! "Inbox" is Microsoft's way of saying the driver is 'in the box', meaning it comes with Windows 8. Bear in bind you may get better performance with manufacturer specific drivers, but it's still a decent showing for a default driver.
Now for even more IOPS:
Yes, you are reading that correctly. That screen is showing a system running over 11 million IOPS. Think it's RAM? Wrong. This is flash memory pulling those numbers. Remember the 2.5" P3700 from a few pics back? How about 24 of them:
The above photo shows three 2U systems (bottom), which are all connected to a single 2U flash memory chassis (top). The top chassis supports three submodules, each with eight SFF-8639 SSDs. The system, assembled by Newisys, demonstrates just how much high speed flash you can fit within an 8U space. The main reason for connecting three systems to one flash chassis is because it takes those three systems to process the full IOPS capability of 24 low latency NVMe SSDs (that's 96 total lanes of PCIe 3.0!)!
So there you have it, IDF storage tech in a nutshell. More to come as we follow these emerging technologies to their maturity.
Subject: Storage | September 12, 2014 - 05:30 PM | Jeremy Hellstrom
Tagged: SM2246EN, S3C, mlc, Apotop
The Apotop S3C SSD uses the same controller as the Angelbird drive Al reviewed recently. It uses synchronous MLC NAND with the 4 channel present on the Silicon Motion controller and is able to provide more than the specified 490 MB/s read and 275 MB/s write in some benchmarks. It can often read faster than the wrk SSD but the writes cannot always keep up though it is not something likely to be noticeable in real usage scenarios. The MSRP is very attractive with the 512GB model expected to be released at $200. Silicon Motion is likely to start appearing in a lot more SSDs in the near future with this mix of price and performance. Read the full review at Kitguru.
"The new Apotop S3C SSD features the Silicon Motion 2246EN controller which we first reviewed in the Angelbird 512GB wrk SSD back in August this year. The controller impressed us, so we have already high hopes for the Apotop S3C."
Here are some more Storage reviews from around the web:
- PNY XLR8 120 GiB SSD Review @ Hardware Secrets
- Plextor M6 Pro 256GB @ Kitguru
- Plextor M6 PRO 256GB @ eTeknix
- We lift the lid on Intel's Pro 2500 SSD. Shock, horror: It doesn't use its own NAND chips @ The Register
- HGST Ultrastar He6 6TB SAS HDD Review @ NikKTech
- TB (2.5-inch) Hard Drives @ SPCR
- SPYRUS WorkSafe Pro WTG Secure Flash Drive @ The SSD Review
- Synology DiskStaion DS115j @ Legion Hardware
- QNAP TurboNAS TS-451 NAS Server Review @ NikKTech
- Icy Box FlexCage MB973SP 2B Trayless 3-in-2 SATA Backplane @ Kitguru
- Thermaltake BlacX Urban Wi-Fi Edition HDD Docking Station Review @ TechwareLabs
- Angelbird SSD2go Pocket External SSD @ The SSD Review
Subject: General Tech, Storage | September 12, 2014 - 04:08 PM | Scott Michaud
Tagged: sandisk, sdxc, sdhc, sd card, 512GB
Assuming your camera, card reader, or other device fully conforms to the SDXC standard, Sandisk has developed a half-terabyte (512GB) memory card. Beyond being gigantic, it can be read at up to 95 MB/s and written at up to 90 MB/s, which should be enough to stream 4K video. Sandisk claims that it is temperature proof, shock proof, water proof, and x-ray proof. It also comes with a lifetime warranty and "RescuePRO Deluxe" recovery software but, honestly, I expect people would just use PhotoRec or something.
It should be noted that the SDXC standard covers memory cards up to 2TB so it will probably not be too long before we see another standard get ratified. What is next? SDUC? SDYC? SDALLTHEC? Blah! This is why IEEE assigns names sequentially.
The SanDisk Extreme PRO UHS-I SDHC/SDXC 512GB memory card should be available now, although I cannot yet find them online, for $799.99 MSRP.
Subject: Storage, Shows and Expos | September 10, 2014 - 03:34 PM | Allyn Malventano
Tagged: TSV, Through Silicon Via, memory, idf 2014, idf
If you're a general computer user, you might have never heard the term "Through Silicon Via". If you geek out on photos of chip dies and wafers, and how chips are assembled and packaged, you might have heard about it. Regardless of your current knowledge of TSV, it's about to be a thing that impacts all of you in the near future.
Let's go into a bit of background first. We're going to talk about how chips are packaged. Micron has an excellent video on the process here:
The part we are going to focus on appears at 1:31 in the above video:
This is how chip dies are currently connected to the outside world. The dies are stacked (four high in the above pic) and a machine has to individually wire them to a substrate, which in turn communicates with the rest of the system. As you might imagine, things get more complex with this process as you stack more and more dies on top of each other:
16 layer die stack, pic courtesy NovaChips
...so we have these microchips with extremely small features, but to connect them we are limited to a relatively bulky process (called package-on-package). Stacking these flat planes of storage is a tricky thing to do, and one would naturally want to limit how many of those wires you need to connect. The catch is that those wires also equate to available throughput from the device (i.e. one wire per bit of a data bus). So, just how can we improve this method and increase data bus widths, throughput, etc?
Before I answer that, let me lead up to it by showing how flash memory has just taken a leap in performance. Samsung has recently made the jump to VNAND:
By stacking flash memory cells vertically within a die, Samsung was able to make many advances in flash memory, simply because they had more room within each die. Because of the complexity of the process, they also had to revert back to an older (larger) feature size. That compromise meant that the capacity of each die is similar to current 2D NAND tech, but the bonus is speed, longevity, and power reduction advantages by using this new process.
I showed you the VNAND example because it bears a striking resemblance to what is now happening in the area of die stacking and packaging. Imagine if you could stack dies by punching holes straight through them and making the connections directly through the bottom of each die. As it turns out, that's actually a thing:
Subject: Storage, Shows and Expos | September 9, 2014 - 04:51 PM | Allyn Malventano
Tagged: WDC< Western Digital, WD, idf 2014, idf, hdd, Cold, Archival, Ae
We talked about helium filled, shingled HDD's from HGST earlier today. Helium may give you reduced power demands, but at the added expensive of hermetically sealed enclosures over conventional HDD's. Shingling may give added capacity, but at the expense of being forced into specific writing methods. Now we know Western Digital's angle into archival / cold storage:
..so instead of going with higher cost newer technologies, WD is taking their consumer products and making them more robust. They are also getting rid of the conventional thinking of capacity increments and are moving to 100GB increments. The idea is that once a large company or distributor has qualified a specific HDD model on their hardware, that model will stick around for a while, but be continued at an increased capacity as platter density yields increase over time. WD has also told me that capacities may even be mixed an matched within a 20-box of drives, so long as the average capacity matches the box label. This works in the field of archival / cold storage for a few reasons:
- Archival storage systems generally do not use conventional RAID (where an entire array of matching capacity disks are spinning simultaneously). Drives are spun up and written to individually, or spun up individually to service the occasional read request. This saves power overall, and it also means the individual drives can vary in capacity with no ill effects.
- Allowing for variable capacity binning helps WD ship more usable platters/drives overall (i.e. not rejecting drives that can't meet 6TB). This should drive overall costs down.
- Increasing capacity by only a few hundred GB per drive turns into *huge* differences in cost when you scale that difference up to the number of drives you would need to handle a very large total capacity (i.e. Exabytes).
So the idea here is that WD is choosing to stick with what they do best, which they can potentially do for even cheaper than their consumer products. That said, this is really meant for enterprise use and not as a way for a home power user to save a few bucks on a half-dozen drives for their home NAS. You really need an infrastructure in place that can handle variable capacity drives seamlessly. While these drives do not employ SMR to get greater capacity, that may work out as a bonus, as writes can be performed in a way that all systems are currently compatible with (even though I suspect they will be tuned more for sequential write workloads).
Here's an illustration of this difference:
The 'old' method meant that drives on the left half of the above bell curve would have to be sold as 5TB units.
With the 'new' method, drives can be sold based on a spec closer to their actual capacity yield. For a given model, shipping capacities would increase as time goes on (top to bottom of the above graphic).
To further clarify what is meant by the term 'cold storage' - the data itself is cold, as in rarely if ever accessed:
Examples of this would be Facebook posts / images from months or years ago. That data may be rarely touched, but it needs to be accessible enough to be browsed to via the internet. The few second archival HDD spinup can handle this sort of thing, while a tape system would take far too long and would likely timeout that data request.
Subject: Storage, Shows and Expos | September 9, 2014 - 02:00 PM | Allyn Malventano
Tagged: ssd, SMR, pcie, NVMe, idf 2014, idf, hgst, hdd, 10TB
It's the first day of IDF, so it's only natural that we see a bunch of non-IDF news start pouring out :). I'll kick them off with a few announcements from HGST. First item up is their new SN100 line of PCIe SSDs:
These are NVMe capable PCIe SSDs, available from 800GB to 3.2TB capacities and in (PCI-based - not SATA) 2.5" as well as half-height PCIe cards.
Next up is an expansion of their HelioSeal (Helium filled) drive line:
Through the use of Shingled Magnetic Recording (SMR), HGST can make an even bigger improvement in storage densities. This does not come completely free, as due to the way SMR writes to the disk, it is primarily meant to be a sequential write / random access read storage device. Picture roofing shingles, but for hard drives. The tracks are slightly overlapped as they are written to disk. This increases density greatly, but writting to the middle of a shingled section is not possible without potentially overwriting two shingled tracks simultaneously. Think of it as CD-RW writing, but for hard disks. This tech is primarily geared towards 'cold storage', or data that is not actively being written. Think archival data. The ability to still read that data randomly and on demand makes these drives more appealing than retrieving that same data from tape-based archival methods.
Further details on the above releases is scarce at present, but we will keep you posted on further details as they develop.
Subject: Storage | September 5, 2014 - 03:06 PM | Jeremy Hellstrom
Tagged: roundup, ssd
The SSD Review has put a quick overview of what they feel are the best SSDs released this summer in several classes, though picking the Intel P3700 PCIe SSD which is not slated for release until the end of September might be considered cheating a bit. It is no surprise that the Samsung 850 Pro is the Enthusiast recommendation or the Crucial MX100 being recommended for those with a tight budget. They also list M.2, mSATA and even USB recommendations so head on over to see the full round up.
"Summer has come and gone, and over the past few months, there have been quite a few SSDs released into the market, and the question of, "Which SSD should I buy?" seems to still come up a lot around forums. Usually, there are some predetermined recommended favorite in each."
Here are some more Storage reviews from around the web:
- Crucial MX100 256 GB @ techPowerUp
- Plextor M6S 256GB @ eTeknix
- Plextor M6M 128GB mSATA @ eTeknix
- Transcend SSD370 SSD @ The SSD Review
- Samsung 850 Pro 512GB SSD Review @ NikKTech
- MyDigitalSSD Super Cache 2 128GB SATA III M.2 Drive Review @ Legit Reviews
- ioSafe 214 Fire and WaterProof NAS Video Review @ Madshrimps
- Synology DS115j @ HardwareHeaven
- QNAP TurboNAS TS-470 Pro NAS Server Review @ NikKTech
- WD My Passport Wireless 1TB Storage Drive Review @ Legit Reviews
- Western Digital My Passport Ultra Review @ TechwareLab
- WD My Passport Wireless Review – Your Own Hand-Held Personal Cloud @ Techgage
- Corsair Voyager Air 2 1TB Wireless Hard Drive @ eTeknix
- Western DIgital My Cloud EX2 Review @ TechwareLabs
- Silicon Power Armor A60 2TB USB 3.0 Portable Hard Drive Review @ NikKTech
- Corsair Flash Voyager GTX USB 3.0 128GB Flash Drive Review @ Madshrimps
Subject: Motherboards, Processors, Chipsets, Memory, Storage | September 5, 2014 - 01:21 PM | Ryan Shrout
Tagged: X99-Deluxe, SSD 730, Intel, Haswell-E, ddr4, asus, 5960X
Okay, I'll be the first to admit that I didn't know what I was getting into. When a couple of packages showed up at our office from Intel with claims that they wanted to showcase the new Haswell-E platform...I was confused. The setup was simple: turn on cameras and watch what happens.
So out of the box comes...a containment chamber. A carefully crafted, wood+paint concoction that includes lights, beeps, motors and platforms.
Want to see how Intel promotes the Core i7-5960X and X99 platform? Check out this video below.
Our reviews of products included in this video:
Introduction, Specifications and Packaging
We first looked at the Silicon Motion 2246EN controller in our Angelbird SSD wrk review. In that review, we noted the highest sequential performance seen in any SATA SSD reviewed to date. Eager to expand our testing to include additional vendors and capacities, our next review touching on this controller is the Corsair Force LX series of SSDs. The Force LX Series is available in 128GB, 256GB, and 512GB capacities, and today we will look at the 256GB and 512GB iterations of this line: