Why Flash Belongs on the Server

 Technical
Brief: Why Flash Belongs on the Server Flash-based storage in enterprise IT is one of the most significant developments in the modern
datacenter. Today, almost all storage solutions supporting multiple enterprise applications utilize
flash technology. Indeed, all-flash devices are increasingly popular for certain tiers of applications.
But where does flash really belong? Should it be merely a component within an existing storage
solution, or should it be designed directly into the server architecture – closer to the applications it
serves?
Before addressing that important question, it is important to first understand the basics of
traditional storage solutions. Network attached storage (NAS) and storage area networks (SAN)
from established vendors dominate the legacy storage market scape. Storage systems from these
companies rely on storage controllers, which are not much more than x86-based servers. They are
dedicated to managing multiple shelves of disks, providing management and efficiency functions
such as snapshots, replication, compression, de-duplication, and automatically tiering data.
Each storage array commonly utilizes two storage controllers for high availability and load
balancing. Shared storage resources are connected to the server cluster through existing Ethernet
or dedicated Fibre Channel networks. With this traditional three-tier architecture, data reads and
writes travel from a virtual machine (VM) to the physical server, through the network fabric, to the
storage controller, and eventually to the disks.
How does flash fit into this model? Flash provides applications with high performance and low
latency, delivering thousands of inputs/outputs per second (IOPS), compared to just hundreds for
spinning disks. Data tiering is a common technique leveraged in today’s hybrid storage solutions,
which contain both hard disk drives and solid state drives (SSDs), to drive higher performance. The
storage controller keeps track of ‘hot data,’ or data that is most frequently used, and promotes this
data into the higher performance SSDs. When data becomes ‘cold’, it is automatically moved to
slower spinning disks.
Tiering ensures that the most commonly used data is always on the highest performance storage
media. This is very similar to how processors use DRAMs today. Since the amount of data that can
be stored in memory is limited, only the “hot data” is in the DRAM and aged data is automatically
swapped out. Why then is DRAM closer to the CPU (i.e., inside the server), whereas flash-based
storage is placed a network hop away inside a distant storage array? Does it not make sense to
keep flash closer to the server just the way DRAMs are today? After all, the PCIe bus is closer to the
CPU and provides more aggregate throughput and lower latency than Fibre Channel networking or
Ethernet over a switch.
Hyper-converged vendors, like Nutanix, who integrate compute and storage in a single 2U
appliance, have been incorporating server-side flash for the past three years. Current datacenter
architectural models are on a path to obsolescence in the new world of web-scale IT. For instance,
what will happen when there are 50 servers in the compute cluster vying to access flash-resident
data on a NAS array? SSDs capable of tens of thousands of IOPS are now connected through the
same controller on the same network. It is important to understand where the performance
bottleneck is – is it the network, the controller, or flash disks? Now, what happens if 50 or even 100
additional servers are added? Will the current model scale? It is analogous to draining water from a
barrel using a single straw. Will adding more straws solve the problem?
The problem is that flash performance is constrained by the underlying storage architecture. To
maintain performance, the end-to-end infrastructure has to scale with the demand for data.
However, anyone with datacenter experience knows that capacity planning exercises are rarely
Technical Brief: Why Flash Belongs on the Server accurate, as business and application requirements are notoriously difficult to predict. Add the
common pain points of traditional storage infrastructure, such as zoning, unmasking, LUN
provisioning, lack of VM-centric functionality, and more, and it is clear why storage administrators
struggle to maintain overall performance and availability SLAs.
Hyper-converged solutions incorporate flash disks in the server, delivering high levels of data
resiliency and pooled data storage that is equivalent to a NAS device, while allowing fast,
unfettered access to the flash storage by applications. They are designed for scalability and
parallelism, without the inherent limitations created when accessing storage through a single
controller. When deployed in clustered configurations, hyper-converged appliances avoid the
network contention that occurs when there is simultaneous demand for flash disks in a single
storage device. Instead, each server has its own flash disks, such that most read operations can be
satisfied by local, direct attached resources, and avoid network access entirely. Even for I/Os that
require the hyper-converged server to go out over the network (e.g., remote reads from adjacent
nodes, or replication of writes to other nodes), there is no single point of bottleneck. Instead, these
operations are distributed across different servers throughout the cluster.
Hyper-convergence optimizes the benefits of flash, even when cluster size grows. Further,
introducing additional storage is as easy as adding another server with the required SSD capacity
or replacing the current SSDs on the server(s) with higher capacity ones. Integrating flash into the
server enables all applications to run on common infrastructure, including workloads that are
bandwidth-hungry or sensitive to storage latency. Server-based flash also eliminates the cost and
complexity of deployment silos driven by hardware architecture.
Figure 1: Evolution of Flash from DAS to Networked Storage to Hyper-Converged Infrastructure
Flash-based storage delivers orders-of-magnitude faster performance than traditional hard disk
drives. However, inserting flash resources into central arrays and accessing them over a network
will never allow their full potential to be realized. Going forward, the emergence of non-volatile
memory express (NVMe) memory brings a lighter-weight standard for accessing SSDs, which will
only increase the demand for server-based flash. Hyper-converged architectures offer a holistic
approach to realize the potential of the modern SSD technology innovations.
About Nutanix
Nutanix is the leader in hyper-converged infrastructure, natively converging compute and
storage into a single 100% software-driven solution to drive unprecedented simplicity at
lower costs in the datacenter. Customers run any application, at any scale with predictable
performance and economics.
Learn more at www.nutanix.com or follow up on Twitter @nutanix.