h130021 Vnxe3200 Fastsuite WP
h130021 Vnxe3200 Fastsuite WP
h130021 Vnxe3200 Fastsuite WP
Abstract
This white paper is an introduction to the EMC® FAST™ Suite for
the VNXe3200 storage system. It describes the components of
the EMC FAST Suite: The FAST Cache and FAST VP, as well as
their implementation in the VNXe3200. Usage guidelines and
major customer benefits are also included.
February 2015
Copyright © 2015 EMC Corporation. All Rights Reserved.
For the most up-to-date listing of EMC product names, see EMC
Corporation Trademarks on EMC.com.
Audience
This white paper is intended for EMC customers, partners, and employees who are
considering the use of the FAST Cache and FAST VP features in the VNXe3200 storage
system. It assumes familiarity with VNXe storage systems and EMC’s management software.
Terminology
Cache page - Unit of allocation inside the FAST Cache, 64 KB in size.
Cache clean page - Page of the FAST Cache that is valid and contains a copy of data that
has been synchronized with the storage resource.
Cache dirty page - Page of the FAST Cache that is valid and contains the most recent copy of
data, but has not yet been synchronized with the storage resource.
Chunk - Portion of data in a particular address range (64 KB).
DRAM memory - Storage-system component used by the Multicore Cache to store data in
very fast storage media (DRAM), to serve requests for that data faster.
Policy Engine
The Policy Engine monitors the flow of IO through the FAST Cache and determines what data
gets copied to the FAST Cache. When a chunk of data on a LUN in a FAST-Cache-enabled
pool is accessed three times in a short period of time, the chunk is copied to the FAST
Cache. There are a few exceptions to this general rule that will be discussed later. The Policy
Engine also maintains statistical information about the data access.
Memory Map
The Memory Map tracks the location of all 64 KB chunks in the FAST Cache. It resides in
DRAM memory to ensure the best performance when locating data. A copy of the Memory
Map is also maintained on the drives in the FAST Cache for high availability. After querying
the Memory Map to determine whether a chunk is in FAST Cache, the system can direct IO
for that chunk accordingly.
Theory of operation
Multicore FAST Cache promotions
A FAST Cache promotion is the process by which data is copied from spinning media HDDs
and placed into the FAST Cache. The process is defined as a promotion due to the upgrade
in performance the data receives from being copied from spinning media to flash
technology. After being promoted, a copy of the data continues to reside on the spinning
media.
Drive configuration
During the FAST Cache creation, you can specify the capacity by the number of drives used.
Once the FAST Cache configuration is set, it cannot be changed. Table 2. RAID Configuration
Options shows the drive quantities, sizes, and capacities that are supported for the FAST
Cache configuration.
FAST VP licensing
When a FAST VP license is installed, users have access to all the features and management
options described in the FAST VP sections of this white paper. The following features are
enabled when a FAST VP license is installed:
Ability to create a multi-tier pool
Ability to set LUN tiering policies
Ability to access the Data Efficiency Settings page in Unisphere
Ability to access the FAST VP tabs on storage pools or storage resources
If a user installs the FAST VP license after creating storage resources, the tiering policies of
those resources will be set to “Start High then Auto-Tier” by default. Rebalancing when
expanding a single-tier pool (discussed later) will occur regardless of whether FAST VP is
Using FAST VP
Storage pools
Storage pools provide a framework that allows FAST VP to take full advantage of different
disk types. A pool is a physical collection of disks on which logical units (LUNs), VMware
datastores, and file systems are created. Pools can contain a few disks or hundreds of
disks. Because of the large number of disks supported in a pool, pool-based provisioning
spreads workloads over more resources thus minimizing planning and management efforts.
Pools can be homogeneous - containing a single disk type (flash, SAS, or NL-SAS), or
heterogeneous - containing multiple disk types.
Homogeneous pools
Homogeneous pools are recommended for applications with limited skew, such that their
access profiles can be very random across a large address range. Multiple LUNs with similar
profiles can share the same pool resources. These LUNs provide more predictable
performance based on the disk type employed. In a homogeneous pool, only one disk type
(flash, SAS, or NL-SAS) is selected during pool creation.
Configurations with higher data to parity ratios (12+1, 14+2, etc.) provide capacity savings
compared to those with lower data-to-parity ratios. The tradeoffs for higher data to parity
ratios are larger fault domains and potentially longer rebuild times. This is especially true
for RAID 5, which has only a single parity disk. EMC advises you to choose carefully between
RAID 5 configurations and decide whether robustness or efficiency is a higher priority.
Robustness is less likely to be an issue for RAID 6, because it has two parity disks.
When creating a storage pool with any of the available tiers, you must add disks in
multiples of the supported RAID configurations.
Performance tier
The Performance tier achieves a combination of performance and capacity. This tier,
composed of Serial Attached SCSI (SAS) disks, offers high levels of performance, reliability,
and capacity. SAS disks are based on industry-standardized, enterprise-level, mechanical
hard-drive technology that stores digital data on a series of rapidly rotating magnetic
platters.
Capacity tier
The Capacity tier is used to decrease the cost per GB of data. This tier, consisting of 7.2K
RPM Near-Line SAS (NL-SAS) disks, is designed for maximum capacity at a modest
performance level. Although NL-SAS disks have a slower rotational speed compared to
disks in the Performance tier, NL-SAS disks can significantly reduce energy use and free up
capacity in the more expensive and higher performing storage tiers.
In a typical system, 75-95% of application data has little IO activity. Since NL-SAS disks cost
less than performance disks on a per-GB basis, and their cost is a small fraction of the cost
of flash disks, they are the most appropriate type of media for this “cold” data. NL-SAS
disks consume 96% less power per terabyte than performance disks, and offer a compelling
opportunity for TCO improvement that considers both purchase cost and operational
efficiency.
FAST VP management
Although FAST VP is a feature that is implemented for the entire storage pool, you can
modify settings at the storage resource level. During the creation of a storage resource via
the GUI or CLI, you can define a tiering policy. A tiering policy specifies where the initial
placement of data will be and how that data will be relocated within the pool during
scheduled and manually invoked relocation periods. FAST VP bases decisions for how data
relocation occurs on performance statistics collected every hour. For more information, see
the “FAST VP Algorithm” section in this paper.
No Data Movement
If a storage resource is configured with this policy, no slices provisioned to the storage
resource are relocated across tiers. Data remains in its current position, but can still be
relocated within a tier. The system collects statistics on these slices after the tiering policy
is changed. You can only apply the “No Data Movement” policy after a storage resource is
created. Because of this, the initial data placement is set to the policy selected during
creation.
Auto-Tier Optimize for Pool Sets the initial data placement to Optimized for Pool and
Performance then relocates the LUN's data based on the LUN's
performance statistics such that data is relocated among
tiers according to I/O activity.
Start High then Auto-Tier Highest Available Tier First sets the preferred tier for the initial data placement to
(Default) the highest-performing disk drives with available space, then
relocates the LUN's data based on the LUN's performance
statistics and the auto-tiering algorithm.
Lowest Available Tier Lowest Available Tier Sets the preferred tier for the initial data placement and
subsequent data relocation (if applicable) to the most cost-
effective disk drives with available space.
No Data Movement Based on previous FAST VP Prevents any ongoing data movement for this LUN. The
policy used current placement of data is retained. Only available as a
LUN property after the LUN is created.
Relocation schedule
The FAST VP feature allows for automatic data relocation based on a user-defined relocation
schedule. This schedule defines when and how frequently the array starts data relocation
on the participating storage pools. These scheduled relocations, which take place on all
pools concurrently, can be arranged to take place during off-hours to minimize any adverse
performance impact. The Schedule tab of the FAST VP Settings page allows you to change
the data relocation schedule. You can launch this window from the storage pool properties
FAST VP tab or by navigating to Settings > Data Efficiency Settings and clicking on the
Schedule tab.
In the Unisphere FAST VP Schedule tab, you can view whether scheduled relocation is
enabled, and the current relocation schedule. You can select days of the week for
relocations to run, as well as start and end times. Multiple relocation operations can occur
during this timeframe, and the software assigns this schedule to all participating pools. This
allows FAST VP to run continuously as a low priority background task if desired. FAST VP can
also be scheduled for a smaller timeframe as a higher priority task. This allows scheduling
of data relocation for off-peak hours in an attempt to minimize the performance impact from
relocation. In the example shown in Figure 10, data relocations are allowed to run each day
of the week for an eight hour window, starting at 22:00 (10:00 PM) and concluding at 6:00
AM.
FAST VP operations
FAST VP operates by relocating the most active data directly to the highest available tier
(either the Extreme Performance or Performance tier). To ensure sufficient space in the
higher tiers, relocations attempt to reclaim 10% free space in the tiers to allow for the new
slice allocation that occurs when new storage resources are created or when thin storage
FAST VP algorithm
FAST VP uses three different strategies to improve performance, capacity, and TCO. These
techniques help identify and relocate slices to the most appropriate tiers by collecting
statistics on each slice, analyzing the data, and relocating each slice based on its activity
level.
Statistics collection
As previously noted, a slice of data is considered “hotter” (more active) or “colder” (less
active) than another slice of data based on the relative activity level of those slices. The
activity level of a particular slice is determined by counting the number of IOs (reads and
writes) bound for each slice. FAST VP maintains a cumulative IO count and “weighs” each IO
according to how recently it arrived. This weight deteriorates over time, and newer IOs are
given a higher weight. After approximately 24 hours, the weights of IOs are nearly cut in half
and continue to decrease. This statistics collection occurs continuously in the background
for all storage resources.
Analysis
FAST VP analyzes the collected data once per hour. This analysis process produces a
ranking order from “hottest” to “coldest” for each slice within the pool. Before relocation is
invoked with automatic or manual relocation, FAST VP creates a candidate list of slices to
move up, down, and within a pool. The ranking of a storage resource and its slices can be
influenced by changing the tiering policy, in which case the tiering policy takes precedence
over the activity level.
Relocation
During the user-defined relocation window, FAST VP promotes slices according to the
candidate list that it created in the analysis stage. The hottest slices are moved to the
highest tier available. During relocation, FAST VP prioritizes relocating slices to higher tiers.
Slices are generally only relocated to lower tiers if the space they occupy is required for a
higher priority slice. This way, FAST VP ensures that the higher performing disks are always
used.
However there are exceptions to this rule. Slices set to “Lowest Available Tier” may be
actively relocated to a lower tier if they are not there already. Also, as mentioned previously,
FAST VP actively demotes the “coldest” slices in a tier if needed to free up 10% of capacity
in a tier.
Interoperability
The FAST Cache and FAST VP features are designed to work not only with each other, but
also with all other VNXe features. This integration allows for ease of management, as users
do not have to worry about conflicts when using multiple data services or features.
Unified Snapshots
The Unified Snapshots feature is fully compatible with the FAST Cache and FAST VP features.
Using the Redirect on Write technology, changes to a snapped storage resource are written
to a new location in the same pool. These new blocks will have the same Tiering Policy and
initial allocation rules as the source storage resource because of the nature of block sharing
between source storage resource and snapshots. This also applies to changes made to
writable snapshots. These new blocks can be promoted and relocated the same as regular
storage resource blocks. For more information, refer to the EMC VNXe3200 Unified
Snapshots white paper on EMC Online Support.
Conclusion
The FAST Suite optimizes VNXe3200 storage systems for flash, which greatly benefits TCO
and overall system performance. The VNXe FAST Suite makes the most efficient use of flash
disk capacity by using flash disks for the most frequently accessed data in the storage
system instead of dedicating disks to particular applications. The FAST Cache and FAST VP
features achieve the goal of flash optimization in different ways, and are best suited to
different situations.
The FAST Cache quickly handles dynamic and unpredictable workload changes, while FAST
VP intelligently tiers data to the appropriate disks based on statistics collection and
analysis over time. These two solutions are complementary, because they work on different
granularity levels and with different frequencies. Implementing both FAST VP and FAST
Cache can significantly improve performance and reduce storage costs.
References
The following white papers can be found on EMC Online Support:
EMC VNXe3200 Unified Snapshots
EMC VNXe3200 File Deduplication and Compression
Using a VNXe System with CIFS File Systems