RAID Not Dead But Future Looks Small
By Jerome Wendt, president and analyst of DCIG
By Jean Jacques Maleval | November 2, 2010 at 3:34 pmHere is an article written by Jerome M. Wendt, president and lead analyst of DCIG, Inc.
RAID is Certainly not Dead But Its Future Looks Small
The month of October saw a sizeable uptick in the readership of a blog entry that appeared nearly two years ago on DCIG’s website on the topic of data loss on SATA storage systems. While this blog entry received a fair amount of interest when it was first published, exactly what prompted a resurgence of interest in this topic this month is unclear.
Maybe it is just an anomoly driven by the whimsical interests of Internet users who are for whatever reason searching on this topic, finding this blog entry and then reading it.
However it may be a more ominous indication that SATA disk drives, which became popular 2 – 3 years ago, are wearing out and that the traditional RAID technologies used to protect them are failing. As a result, users are looking for information as to why RAID, in some circumstances, is not doing the job in their environment.
The death of RAID (or at least RAID 5) has previously been forecast by some analysts. But even now, when I look at the features of new storage arrays, the number of RAID options that they support is always prominently mentioned.
A good example is earlier this week Overland Storage announced its new SnapSAN S1000. It offered at least 10 different possible ways that RAID could be configured (including RAID 5) on a storage array that starts under $10K in price so do not tell me that RAID is dead or even on its last legs.
But these is no disputing that the capacities of SATA disk drives are forecast and expected to cross the 4, 8, 16 and 32 TB thresholds over the next decade. As that occurs, it becomes questionable if current RAID technologies are adequate to protect these size disk drives. If the increased interest in DCIG’s 2008 blog entry is any indication, it would appear no.
So am I predicting the death of RAID? Clearly I am not. RAID technology is as much a part of the storage landscape as tape and odds are that innovation will continue to occur in RAID that make it a relevant technology for the foreseeable future.
Yet it was clear from speaking to a few users and storage providers in attendance at Storage Networking World (SNW) in Dallas, TX, earlier this month, that new approaches to protecting data stored on larger capacity SATA disk drives are going to be needed in the next decade in order to meet their anticipated needs.
One specific company that I met with at length while at SNW was a company called Amplidata. It is already innovating in this space to overcome two of the better known limitations of RAID to include:
- The increasing length of time to rebuild larger capacity drives. Rebuild times for 2 TB drives are already known to take four hours or longer to complete though I have heard that in some cases, depending on how busy the storage system is, it can take days for a rebuild of a disk drive of this size to finish.
- The need to keep all disks in its RAID group spinning so no power savings can be realized. Spin down is likely to become more important in the years to come as more data is archived to disk with it likely becoming a function of the storage array to intelligently manage and place the archived data on these drives as opposed to the software to facilitate the spin down of these drives.
So what Amplidata’s AmpliStor does is distribute and store data redundantly across a large number of disks. The algorithm that AmpliStor uses first puts the data into an object and then stores the data across multiple disks in the AmpliStor system. By storing the data as an object, Amplidata can reconstruct the original data from any of the disks on which the data within the object resides.
This technique eliminates the growing concerns about the rebuild times associated with large disk drives since the original data can be retrieved and reconstructed even if a one, two or even more disks fail. Also, should disk drives in the system be spun down to save energy, they do not need to be spun up to retrieve needed data since the data can be retrieved and reconstructed from other spinning disks on the system.
While it unlikely that AmpliStor or its underlying technology will be widely adopted in the next few year, the simple fact is that increasing capacities of disk drives will eventually make technologies like what AmpliStor has embedded inside of it a prerequisite in almost any high capacity enterprise storage system.
So in the same way that enterprise storage vendors started to adopt RAID 6 about five years ago to prevent the loss of data should two SATA drives fail, look for some variation of the technology that Amplidata has implemented in its AmpliStor to begin to find its way into enterprise storage systems over the next decade to prevent the loss of data on these ever larger disk drives. At the same time, expect RAID to find a new home on smaller storage arrays where the level of protection and speed of recovery that RAID provides should be more than adequate.
Comments
There are several other companies using a technology similar to Amplidata, with blocks mirrored on several disks on a network rather than several HDDs using RAID controller. They include Cleversafe, IBM (XIV), Isilon, Permabit and Scality. It's clearly an innovative idea that has great future since it can now be implemented because the bandwidth of the networks has been largely pushed with 1GbE and 10GbE, or IB. Some people called it RAIN (Redundant Array of Independent Nodes) that partitions storage space across multiple nodes in a network. A similar approach can also be applied on private or public clouds (Wuala and others).
We see specifically five advantages:
- No controller is needed at all.
- The level of protection is practically infinite, it just depends on the number of copies you want to do for each block: 1, 2, 3, 4 or more.
- There is no supplementary HDDs to store the parity calculated or for disk spare (like in RAID-6)
- Also fewer time is wasted to reconstruct a failed disk array, a heavy difficulty with current high-capacity HDDs. Henry Newman Henry Newman, CTO of Instrumental Inc. and a regular Enterprise Storage Forum contributor, has calculated that it takes 22 hours to reconstruct a disk array with 1.5TB disks.
- When a disk array is full, you have either to substitute disks by higher-capacity ones or to add an expansion unit. On a RAIN, you just add a new node automatically configured to work with all the other ones.