Tintri VMStore Overview
Tintri VMStore Overview
irtualization is the new normal. More than 50 percent of new workloads are virtualized, and V companies are beginning to make significant investments in virtual desktop infrastructure (VDI). ommodity flash storage is quickly becoming a significant part of both local and shared storage C infrastructure.
Existing storage offerings are poorly adapted to both virtualization and flash:
n
irtualization benefits significantly from shared storage, but todays general-purpose shared V storage was designed 20 years before VMware popularized virtualization with a very different set of workloads in mind. lash storage which is about 400 times faster than disk must be treated very differently F than the rotating magnetic disks most storage systems were designed to use. As a result, most solutions use an expensive and complex bolt-on approach with flash-as-cache.
One hundred percent of flash arrays can address the second problem, but are not designed to run virtual workloads efficiently, so costly space and performance are wasted on idle data. Its the equivalent of using a commercial jet to commute 30 miles it may be slightly faster than driving, but the fuel costs are prohibitively expensive. Consequently, customers struggle with existing storage systems that are poorly adapted to both flash and virtualization, inhibiting the fundamental IT goals of lower cost and greater business agility. Systems purpose-built for both virtualization and flash can overcome these issues. As data centers move from about 30 percent virtualized to well over 50 percent virtualized, deploying storage specifically designed for these environments provides substantially more value. This paper will explore the challenges of designing storage systems using flash for virtualization, and describe Tintris approach.
ost-efficiency: Although MLC is two to four times cheaper than its cousin SLC, its still about C 20 times more expensive than SATA disks. To use flash cost-efficiently, technologies like inline deduplication and compression are critical. atency spikes: Flash drives are programmed at the page level (512B to 4KB), but can only L be erased at the block level (512KB to 2MB) sizes much larger than average IO requests. This asymmetry in write vs. erase sizes leads to write amplification which, if not managed appropriately, creates latency spikes. urability: MLC flash in particular can be vulnerable to durability and reliability problems in the D underlying flash technology. Each MLC cell can be overwritten only 5,000 to 10,000 times before wearing out, so the file system must account for this and write evenly across cells.
Disk-based systems were created more than 20 years ago to cope with a very different set of problems. Adapting these systems to use flash efficiently is comparable to attempting to adapt an old 8-bit singlethreaded operating system to efficiently use todays multicore 64-bit architectures.
ost-efficiency: By design, nearly all active data will live exclusively in flash. To maximize flash C usage, Tintri combines fast inline dedupe and compression with a hybrid file system that automatically moves only cold data to SATA. Inline dedupe and compression are also highly effective in virtualized environments where many VMs are deployed by cloning existing VMs, or have the same operating system and applications installed. Tintri VMstore flash is neither a pure read cache nor a separate preallocated storage tier. Instead, flash is intelligently assigned where its high performance will provide the most benefit.
Figure 1: Tintri employs sophisticated, patent-pending technology to eliminate both write amplification and latency spikes.
atency management: Tintri employs sophisticated patent-pending technology to eliminate both L the write amplification and latency spikes characteristic of MLC flash Technology (Figure 1). This approach delivers consistent sub-millisecond latency from the more cost-effective MLC flash. lash durability: Tintri uses an array of technologies including deduplication, compression, F advanced transactional and garbage collection techniques, combined with SMART (Self-Monitoring, Analysis and Reporting Technology) monitoring of flash devices to intelligently maximize the durability of MLC flash. Tintri also employs RAID 6, eliminating the impact of potential latent manufacturing or internal software defects from this new class of storage devices.
Tintri VMstore leverages the strengths of MLC flash, while negating its weaknesses, providing a highly reliable and durable storage system suitable for enterprise applications.
CPU
Memory
Memory
Figure 2: Traditional shared storage maps VMs to LUNs or volumes, rather than managing at the VM and vDisk level.
Storage Arrays
Platinum
Gold
Silver
This mismatch increases cost and complexity. Each new VM instance must be assigned a specific storage LUN or volume. When IO requirements and VM behavior are not well understood, this becomes a painful trial-and-error process. Storage and VM administrators must coordinate to ensure each application has not only the space it needs, but also sufficient IO performance for the expected load. Usually, multiple VMs occupy the same volume or LUN to reduce mapping complexity and space overhead; however, this complicates IO performance problems. A storage-centric view of performance data means administrators must work backward to determine which VMs are affected and which VMs are generating load. Even technologies such as auto-tiering, which seek to reduce storage management overhead, operate at the wrong level. Without the ability to report behavior on a per-VM or per-virtual disk level, advanced storage technology increases complexity and risk. Instead of the simple VM model provided by hypervisors, legacy storage responds with a blizzard of options and interfaces. In these situations, the complexity of configuring, managing and tuning traditional storage for VMs is costly, and ultimately limits the adoption of virtualization. In fact, many applications cannot be costeffectively virtualized with legacy shared storage.
Tintri
VM vDisk
Tintri incorporates advances in flash technology, file system architecture, and user interface design to make storage for virtual applications simple and efficient. Tintri VMstore is designed from the ground up exclusively for VMs by experts in both virtualization and storage. Tintri VMstore is managed in terms of VMs and virtual disks, not LUNs or volumes. The Tintri OS is built from scratch to meet the demands of a VM environment, and to provide features relevant to VMs. It is designed to use flash efficiently and reliably, leveraging technologies like deduplication, compression and automatic data placement to deliver 99 percent of IO from flash. These innovations shift the focus from managing storage as a separately configured component to managing VMs as a whole. This overcomes the performance, management and cost obstacles that prevent virtualization of more of the computing infrastructure. Tintris sharp focus on creating a better storage system for VMs enables us to build a fundamentally new type of product. Building a VM-focused management interface relies on far more than just an attractive GUI. The underlying storage system is designed to natively understand and support storage management operations such as performance and capacity monitoring, snapshots, quality of service (QoS) management, and replication at the VM level.
Tintris focus on VMs is most apparent in the VMstore management interface, which presents VMs as the basic units of management, rather than LUNs, volumes, or files. Every object in the interface is familiar to VM administrators (Figure 5). The interface is simple enough for VM administrators to manage storage directly, yet sophisticated enough for storage administrators to leverage their expertise in managing storage for huge numbers of VMs.
Figure 7: Tintri provides insight in to the historical latency statistics at the VM level.
VM Alignment
VM alignment is the daunting to-do item. And its a problem that poses real challenges as virtualization spreads into more mainstream workloads. Misaligned VMs magnify IO requests, consuming extra IOPS on the storage array. At a small scale, the impact is small. However the impact snowballs as the environment grows, with a single array supporting hundreds of VMs. At this size, performance impact estimates range from 10 percent to more than 30 percent. Every guest OS writes data to disk in logical chunks. Storage arrays also represent data in logical blocks. When a VM is created, the block boundaries on the guest OS and storage dont always align automatically. If the blocks are not aligned, guest requests span two storage blocks, requiring additional IO (see Figure 8 and Figure 9). A VM runs a guest OS that creates one or more virtual disks to store state. The guest OS typically defines the layout of each virtual disk with a common partition layout, such as a master boot record (MBR). The MBR stores information about how each virtual disk is partitioned into smaller regions, with its size and location. Except for Windows Server 2008 and Windows 7, blocks defined by the guest OS file system (NTFS, EXT3, etc.) do not typically align with the underlying datastore block layout.
Tintris VM Auto-alignment
So why are VMs misaligned? Certainly administrators attempt to address the issue by using a variety of utilities to manually align VMs and reduce performance demand. Numerous blogs, whitepapers and knowledgebases describe why VMs should be aligned and provide step-by-step instructions. But as administrators know, realigning a VM is a manual process. Worse, it generally requires substantial downtime.
Summary
Storage remains the primary obstacle to accelerating virtualization growth. Tintri VMstore allows you to overcome the complexity, performance and cost obstacles that prevent virtualization of more of your computing infrastructure. Instant bottleneck visualization and VM auto-alignment are a direct outgrowth of Tintris custom VM-aware file system. Tintris VM-aware appliance eliminates major storage issues in virtualized environments and leverages flash to provide sub-millisecond latency for hundreds of VMs on a single device.