- Best iPhone, iPad Business Apps for 2014
- 14 Tech Conventions You Should Attend in 2014
- 10 Desktop Apps to Power Your Windows PC
- How to Add New Job Skills Without Going Back to School
Network World - This vendor-written tech primer has been edited by Network World to eliminate product promotion, but readers should note it will likely favor the submitter's approach.
Enterprises are challenged to keep pace with mounting unstructured file data. While network attached storage (NAS) represents the optimal choice for storing such data, the strain to scale NAS economically while preserving application performance is like trying to use your fingers to plug holes in a dam that has sprouted thousands of leaks.
This is worsened by the fact that much of today's enterprise storage is consumed by inactive data. In response, vendors have delivered storage tiers and media types optimized to each tier. But tiered storage can increase latency, and nothing can eat at performance faster or frustrate users more than adding latency.
IN DEPTH: Struggling with supersized storage
How much latency is acceptable varies greatly. In the world of visual effects and computer graphics, for example, latencies of .2ms are considered a competitive advantage, while tolerances of 1ms or less are ideal in the database domain. On the other hand, users accessing a general purpose application may tolerate much higher latencies.
Tiered storage falls short in achieving targeted latencies. Among popular file serving benchmarks, measuring very high-end NAS devices with integrated tiered storage (including SSD technology), a .4ms latency was recently achieved. This was the absolute best the high-end NAS solution being tested could deliver. As the benchmark test added incremental client load, the latency increased, reaching .8ms.
The issue isn't so much with the different media in a tiered storage device, but more with the underlying architecture. Within any NAS is a server with processors, memory and I/O. What makes NAS "NAS" and not a general purpose file server is the customized file systems and the many associated applications resting on top used to protect and/or optimize data placement.
Tiered storage architectures often execute clustered file systems, joining N number of nodes, load balancing across nodes, moving data to and from NVRAM to disks, hosting RAID controllers, performing housekeeping of file-system metadata, and executing data protection applications. The point is that NAS controllers perform many tasks that consume system bandwidth. The most sacred function is the orderly storage, retrieval and protection of user data, and not performance.
The question becomes, How do enterprises get the best of both worlds: scalable capacity and scalable performance that preserves low latency?
What is needed is a performance layer within the NAS that isolates high performing media such as Flash, a performance layer that sits alongside the NAS the enterprise already has. It is void of latency-consuming responsibilities, such as executing a file system. Its job is simply to cache hot and warm data and accelerate NFS processing such that clients get sought after data quickly. It should pass on newly written data to the back-end NAS while adding minimal latency, say as low as 10 microseconds.