The external storage industry is undergoing considerable churn right now spurned by waves of disruptive developments from flash to hyperconvergence to cloud. As enterprises seek to modernize storage, they are increasingly looking at scale-out NAS to meet pressing critical needs for lots of file storage.
For years the epitome of big storage for most data centers was the storage area network (SAN) array. A foundation for data center consolidation, the SAN array started as a platform for block-level storage over a dedicated fibre channel (FC) network. Over a decade, the SAN evolved to be an all-singing, all-dancing unified storage solution typically hosting multiple network transports (FC, IP ethernet, fibre channel over ethernet) as well as block and file storage.
Concurrent with the evolution of unified storage, Info-Tech’s enterprise storage strategy projects typically were about helping the organization assess needs and evaluate requirements for their next big SAN purchase. But a funny thing happened on the road to unified storage bliss. Many have concluded that their current SAN array is likely to be their last.
Here are three examples from recent Info-Tech engagements around storage modernization:
1. An aerospace technology manufacturer assessing their current and future storage needs found that total capacity requirement for one purpose – file storage for product engineering quality assurance requirements – was more than what was needed for all other applications and services combined.
Source: Info-Tech Research Group
2. A university was finding that research files were stored in a vast hodgepodge of PC hard drives, NAS filers, and removable media. There was no capability to set standards for availability and recoverability for research storage and no way to manage predictable capacity growth.
3. A national multi-location law firm needed a storage platform for a large and growing repository of legal documents. The need was for enterprise class storage that was also “cheap and deep.”
In all three of the above cases, the organization was looking to a scale-out NAS solution. They all have traditional SAN arrays deployed but see those eventually being replaced by storage closer to the processing, such as flash-storage-based hyperconverged appliances.
Network attached storage has been around for decades. In fact, most organizations first external storage – that is storage outside of a server or workstation where the data was accessible over a network – was likely a NAS storage array. One of the early leaders in selling these network-attached storage appliances basically named their company after the product type (NetApp).
Saying the NAS is “back” is a bit misleading for a technology that has never left. What is “back” is the focus on NAS to address critical needs. Today’s enterprise NAS solutions have evolved in a number of ways while continuing to provide all the enterprise-class resiliency bells and whistles such as hardware redundancy, replication, and data snapshotting. Some defining characteristics include:
We see a number of storage veteran and upstart providers for scale-out NAS are vying for your business. NetApp continues to be a player in the NAS game, of course. Others include Panasas and HPE Scality. For purely software-defined scale out NAS on commodity server clusters, RedHat comes up a lot with its Gluster File System (GlusterFS). An example of a recent player that has come up often in storage modernization engagements is Seattle, Washington-based Qumulo.
A leader in this field has to be Dell EMC Isilon. Isilon first came on the scene as a high-performance, high-capacity scale out NAS filer targeted at media companies. As use cases for high-capacity file storage broadened to other industries, Isilon became a go-to for enterprise NAS. All three example organizations listed above evaluated Isilon (as well as some of the others including Qumulo).
Isilon exhibits all of the characteristics listed above. It is scale-out. It has cloud gateway capability (thanks to EMC’s acquisition a few years ago of cloud gateway pioneer TwinStrata). And it can run big data Hadoop File System (HDFS) natively. It is an enterprise-class solution, so cost can be prohibitive. Storage is getting cheaper, but enterprise storage remains expensive. Isilon is always on the shortlist but it isn’t always the final choice.
Isilon was first acquired by EMC in 2010 and has now landed with Dell as part of the acquisition of EMC. Interestingly, the recent player example above, Qumulo, was founded by a group of ex-Isilon engineers.
I would be remiss when discussing big storage buckets if I didn’t also mention object storage. If the search for “cheap and deep” is leading toward complete software abstraction on lowest common denominator commodity hardware, object storage is a contender.
This is how storage is done in public clouds like Microsoft Azure or Amazon S3. Data is chunked up into file-like objects that are replicated and storage across a grid of storage devices. Some object competitors in the file storage space include Dell EMC Elastic Cloud Storage and HPE Scality (again), as well as OpenStack Swift.
In planning your future roadmap, recognize that the big SAN array may not be the single foundation for your storage.
Network-attached file storage has been around for decades, but it is having a renaissance with scale-out and software-defined technologies. Enterprises are moving away from big unified SAN storage arrays, but the need for big file repositories remains. Evaluate whether scale-out NAS or large object storage grids are in your future.