Você está na página 1de 0

White Paper

Abstract
This white paper introduces EMC

Fully Automated Storage


Tiering for Virtual Pools (FAST VP) technology and describes its
features and implementation. Details on how to work with the
product in the Unisphere

operating environment are


discussed, and usage guidance and major customer benefits are
also included.

October 2011


EMC FAST VP FOR UNIFIED STORAGE SYSTEMS
A Detailed Review


2 EMC FAST VP for Unified Storage SystemsA Detailed Review




















Copyright 2011 EMC Corporation. All Rights Reserved.

EMC believes the information in this publication is accurate as
of its publication date. The information is subject to change
without notice.

The information in this publication is provided as is. EMC
Corporation makes no representations or warranties of any kind
with respect to the information in this publication, and
specifically disclaims implied warranties of merchantability or
fitness for a particular purpose.

Use, copying, and distribution of any EMC software described in
this publication requires an applicable software license.

For the most up-to-date listing of EMC product names, see EMC
Corporation Trademarks on EMC.com.

VMware is a registered trademarks or trademarks of VMware,
Inc. in the United States and/or other jurisdictions. All other
trademarks used herein are the property of their respective
owners.

Part Number h8058.3

3 EMC FAST VP for Unified Storage SystemsA Detailed Review
Table of Contents
Executive summary ....................................................................................... 4
Audience ............................................................................................................................ 4
Introduction ................................................................................................. 5
Storage tiers ................................................................................................ 6
Extreme Performance Tier drives: Flash drives (for VNX and CX4 platforms) ........................ 6
Performance Tier drives: SAS (VNX) and Fibre Channel (CX4) .............................................. 6
Capacity Tier drives: NL-SAS (VNX) and SATA (CX4)............................................................. 7
FAST VP operations ....................................................................................... 8
Storage pools ..................................................................................................................... 8
FAST VP algorithm .............................................................................................................. 9
Statistics collection ........................................................................................................ 9
Analysis ....................................................................................................................... 10
Relocation .................................................................................................................... 10
Managing FAST VP at the storage pool level...................................................................... 10
Physically allocated capacity versus reserved capacity................................................. 12
Automated scheduler ................................................................................................... 13
Manual relocation ........................................................................................................ 14
FAST VP LUN management................................................................................................ 14
Tiering policies ............................................................................................................. 15
Initial placement .......................................................................................................... 15
Common uses for Highest Available Tiering policy ........................................................ 16
Using FAST VP for file ................................................................................... 17
Management .................................................................................................................... 17
Best practices for VNX for File ........................................................................................... 19
Automatic Volume Manager rules ..................................................................................... 20
General Guidance and recommendations ........................................................ 23
FAST VP and FAST Cache................................................................................................... 23
What drive mix is right for my I/O profile? ......................................................................... 24
Conclusion ................................................................................................ 25
References................................................................................................. 26



4 EMC FAST VP for Unified Storage SystemsA Detailed Review
Executive summary
Fully Automated Storage Tiering for Virtual Pools (FAST VP) can lower total cost of
ownership (TCO) and increase performance by intelligently managing data placement
at a sub-LUN level. When FAST VP is implemented, the storage system measures,
analyzes, and implements a dynamic storage-tiering policy much faster and more
efficiently than a human analyst could ever achieve.
Storage provisioning can be repetitive and time-consuming and produce uncertain
results. It is not always obvious how to match capacity to the performance
requirements of a workloads data. Even when a match is achieved, requirements
change, and a storage systems provisioning will require constant adjustment.
Storage tiering puts drives of varying performance levels and cost into a storage pool.
LUNs use the storage capacity they need from the pool, on the devices with the
required performance characteristics. FAST VP collects I/O activity statistics at a 1 GB
granularity (known as a slice). The relative activity level of each slice is used to
determine which slices should be promoted to higher tiers of storage. Relocation is
initiated at the users discretion through either manual initiation or an automated
scheduler. Working at such a granular level removes the need for manual, resource-
intensive LUN migrations while still providing the performance levels required by the
most active dataset.
FAST VP is a licensed feature available on EMC

VNX

series and CLARiiON

CX4
series platforms. The VNX series supports a unified approach to automatic tiering for
both file and block data. CX4 systems running release 30 and later are supported for
the tiering of block data. FAST VP licenses are available a la carte for both platforms,
or as part of a FAST Suite of licenses that offers complementary licenses for
technologies such as FAST Cache, Analyzer, and Quality of Service Manager.
This white paper introduces the EMC FAST VP technology and describes its features,
functions, and management.
Audience
This white paper is intended for EMC customers, partners, and employees who are
considering using the FAST VP product. Some familiarity with EMC midrange storage
systems is assumed. Users should be familiar with the material discussed in the
white papers Introduction to EMC VNX Series Storage Systems and EMC VNX Virtual
Provisioning.


5 EMC FAST VP for Unified Storage SystemsA Detailed Review
Introduction
Data has a lifecycle. As data progresses through its lifecycle, it experiences varying
levels of activity. When data is created, it is typically heavily used. As it ages, it is
accessed less often. This is often referred to as being temporal in nature. FAST VP is a
simple and elegant solution for dynamically matching storage requirements with
changes in the frequency of data access. FAST VP segregates disk drives into the
following three tiers:
Extreme Performance Tier Flash drives
Performance Tier Serial Attach SCSI (SAS) drives for VNX platforms and Fibre
Channel (FC) drives for CX4 platforms
Capacity Tier Near-Line SAS (NL-SAS) drives for VNX platforms and SATA drives
for CX4 platforms
You can use FAST VP to aggressively reduce TCO and/or to increase performance. A
target workload that requires a large number of Performance Tier drives can be
serviced with a mix of tiers, and a much lower drive count. In some cases, an almost
two-thirds reduction in drive count is achieved. In other cases, performance
throughput can double by adding less than 10 percent of a pools total capacity in
Flash drives.
FAST VP has proven highly effective for a number of applications. Tests in OLTP
environments with Oracle
1
or Microsoft SQL Server
2
FAST VP can be used in combination with other performance optimization software,
such as FAST Cache. A common strategy is to use FAST VP to gain TCO benefits while
using FAST Cache to boost overall system performance. There are other scenarios
where it makes sense to use FAST VP for both purposes. This paper discusses
considerations for the best deployment of these technologies.
show that users can lower their
capital expenditure (by 15 percent to 38 percent), reduce power and cooling costs
(by over 40 percent), and still increase performance by using FAST VP instead of a
homogeneous drive deployment.
The VNX series of storage systems delivers even more value over previous systems by
providing a unified approach to auto-tiering for file and block data. FAST VP is
available on the VNX5300

and larger systems. Now, file data served by VNX Data


Movers can also use virtual pools and the same advanced data services as block
data. This provides compelling value for users who wish to optimize the use of high-
performing drives across their environment.


1
Leveraging Fully Automated Storage Tiering (FAST) with Oracle Database Applications EMC white paper
2
EMC Tiered Storage for Microsoft SQL Server 2008Enabled by EMC Unif ied Storage and EMC Fully Automated Storage
Tiering (FAST) EMC white paper

6 EMC FAST VP for Unified Storage SystemsA Detailed Review
Storage tiers
FAST VP can leverage two or all three storage tiers in a single pool. Each tier offers
unique advantages in performance and cost.
Extreme Performance Tier drives: Flash drives (for VNX and CX4 platforms)
Flash drives are having a large impact on the external-disk storage system market.
They are built on solid-state drive (SSD) technology. As such, they have no moving
parts. The absence of moving parts makes these drives highly energy-efficient, and
eliminates rotational latencies. Therefore, migrating data from spinning disks to
Flash drives can boost performance and create significant energy savings.
Tests show that adding a small (single-digit) percentage of Flash capacity to your
storage, while using intelligent tiering products (such as FAST VP and FAST Cache),
can deliver double-digit percentage gains in throughput and response time
performance in some applications. Flash drives can deliver an order of magnitude
better performance than traditional spinning disks when the workload is IOPS-
intensive and response-time sensitive. They are particularly effective when small
random-read I/Os with a high degree of parallelism are part of the I/O profile, as they
are in many transactional database applications. On the other hand, bandwidth-
intensive applications perform only slightly better on Flash drives than on spinning
drives.
Flash drives have a higher per-gigabyte cost than traditional spinning drives. To
receive the best return, you should use Flash drives for data that requires fast
response times and high IOPS. The best way to optimize the use of these high-
performing resources is to allow FAST VP to migrate data to Flash drives at a sub-LUN
level.
Performance Tier drives: SAS (VNX) and Fibre Channel (CX4)
Traditional spinning drives offer high levels of performance, reliability, and capacity.
These drives are based on industry-standardized, enterprise-level, mechanical hard-
drive technology that stores digital data on a series of rapidly rotating magnetic
platters.
The Performance Tier includes 10k and 15k rpm spinning drives, which are available
on all EMC midrange storage systems. They have been the performance medium of
choice for many years. They also have the highest availability of any mechanical
storage device. These drives continue to serve as a valuable storage tier, offering
high all-around performance including consistent response times, high throughput,
and good bandwidth, at a mid-level price point.
The VNX series and CX4 series use different drive attach technologies. VNX systems
use multiple four-lane 6-Gb/s SAS back end while the CX4 series uses one or more 4-
Gb/s Fibre Channel back end. Despite the differences in speed, the drive assemblies
on both series are very similar. When sizing a solution with these drives, you need to
consider the rotational speed of the drive. However, the interconnect to the back end
does not affect rule-of-thumb performance assumptions for the drive.

7 EMC FAST VP for Unified Storage SystemsA Detailed Review
This also applies to the different drive form factors. The VNX series offers 2.5-inch and
3.5-inch form-factor SAS drives. The same IOPS and bandwidth performance
guidelines apply to 2.5-inch 10k drives and 3.5-inch 10k inch drives. The CX4 series
offers only 3.5 inch drives.
FAST VP differentiates tiers by drive type. However, it does not take rotational speed
into consideration. We strongly encourage you to use one rotational speed for each
drive type within a pool. If multiple rotational-speed drives exist in the array, multiple
pools should be implemented as well.
Capacity Tier drives: NL-SAS (VNX) and SATA (CX4)
Using capacity drives can significantly reduce energy use and free up more expensive,
higher-performance capacity in higher storage tiers. Studies have shown that 60
percent to 80 percent of the capacity of many applications has little I/O activity.
Capacity drives can cost about four times less than performance drives on a per-
gigabyte basis, and a small fraction of the cost of Flash drives. They consume up to
96 percent less power per TB than performance drives. This offers a compelling
opportunity for TCO improvement considering both purchase cost and operational
efficiency.
Capacity drives are designed for maximum capacity at a modest performance level.
They have a slower rotational speed than Performance Tier drives. NL-SAS drives for
the VNX series have a 7.2k rotational speed, while SATA drives available on the CX4
series come in 7.2k and 5.4k rpm varieties. The 7.2k rpm drives can deliver
bandwidth performance within 20 percent to 30 percent of performance class drives.
However, the reduced rotational speed is a trade-off for significantly larger capacity.
For example, the current Capacity Tier drive offering is 2 TB, compared to the 600 GB
Performance Tier drives and 200 GB Flash drives. These Capacity Tier drives offer
roughly half the IOPS/drive of Performance Tier drives. Future drive offerings will have
larger capacities, but the relative difference between disks of different tiers is
expected to remain roughly the same.


8 EMC FAST VP for Unified Storage SystemsA Detailed Review
Table 1. Feature tradeoffs for Flash, Performance, and Capacity drives
Flash drives Performance (SAS, FC) Capacity (NL- SAS, SATA)
P
e
r
f
o
r
m
a
n
c
e

High IOPS/GB and low
latency
Sole use response time
<1-5 ms
Multi-access response
time < 10 ms
High bandwidth with
contending workloads
Sole use response time
~5 ms
Multi-access response
time 10-50 ms
Low IOPS/GB
Sole use response time 7-
10 ms
Multi-access response
time up to 100 ms
Leverage storage array SP
cache for sequential and
large block access
S
t
r
e
n
g
t
h
s

Provide extremely fast
access for reads
Can execute multiple
sequential streams better
than SAS/FC
Sequential reads can
leverage read-ahead
Sequential writes can
leverage system
optimizations favoring
disks
Read/write mixes give
predictable performance
Large I/O is serviced fairly
efficiently
Sequential reads can
leverage read-ahead
Sequential writes can
leverage system
optimizations favoring
disks
L
i
m
i
t
a
t
i
o
n
s

Writes slower than reads
Heavy concurrent writes
affect read rates
Single-threaded large
sequential I/O equivalent
to SAS/FC
Uncached writes are
slower than reads
Long response times
when under heavy write
loads
Not as good as SAS/FC at
handling multiple streams
FAST VP operations
FAST VP operates by periodically relocating the most active data up to the highest
available tier (typically the Extreme Performance or Performance Tier). To ensure
sufficient space in the higher tiers FAST VP relocates less active data to lower tiers
(Performance or Capacity Tiers). FAST VP works at a granularity of 1 GB. Each 1 GB
block of data is referred to as a slice. When FAST VP relocates data, it will move the
entire slice to a different storage tier.
Storage pools
Heterogeneous storage pools are the framework that allows FAST VP to fully utilize
each of the storage tiers discussed. Heterogeneous pools are made up of more than
one type of drive. LUNs can then be created at the pool level. These pool LUNs are no
longer bound to a single storage tier; instead, they can be spread across different
storage tiers within the same pool.

9 EMC FAST VP for Unified Storage SystemsA Detailed Review

Figure 1. Heterogeneous storage pool concept
LUNs must reside in a pool to be eligible for FAST VP relocation. Pools support thick
LUNs and thin LUNs. Thick LUNs are high-performing LUNs that use contiguous logical
block addressing on the physical capacity assigned from the pool. Thin LUNs use a
capacity-on-demand model for allocating drive capacity. Thin LUN capacity usage is
tracked at a finer granularity than thick LUNs to maximize capacity optimizations.
FAST VP is supported on both thick LUNs and thin LUNs.
RAID groups are by definition homogeneous and therefore are not eligible for sub-LUN
tiering. LUNs in RAID groups can be migrated to pools using LUN Migration. For a
more in-depth discussion of pools, please see the white paper EMC VNX Virtual
Provisioning - Applied Technology.
FAST VP algorithm
FAST VP uses three strategies to identify and move the correct slices to the correct
tiers: statistics collection, analysis, and relocation.
Statistics collection
A slice of data is considered hotter (more activity) or colder (less activity) than
another slice of data based on the relative activity level of the two slices. Activity
level is determined by counting the number of I/Os for each slice. FAST VP maintains
a cumulative I/O count and weights each I/O by how recently it arrived. This
weighting decays over time. New I/O is given full weight. After approximately 24
hours, the same I/O carries about half-weight. After a week, the same I/O carries very
little weight. Statistics are continuously collected (as a background task) for all pool
LUNs.

10 EMC FAST VP for Unified Storage SystemsA Detailed Review
Analysis
Once per hour, the collected data is analyzed. This analysis produces a rank ordering
of each slice within the pool. The ranking progresses from the hottest slices to the
coldest slices relative to the other slices in the same pool. (For this reason, a hot
slice in one pool may be comparable to a cold slice in another pool.) There is no
system-level threshold for activity level. The most recent analysis before a relocation
determines where slices are relocated.
Relocation
During user-defined relocation windows, 1 GB slices are promoted according to both
the rank ordering performed in the analysis stage and a tiering policy set by the user.
Tiering policies are discussed in detail in the AST VP LUN Management section.
During relocation, FAST VP relocates higher-priority slices to higher tiers; slices are
relocated to lower tiers only if the space they occupy is required for a higher-priority
slice. This way, FAST VP fully utilizes the highest-performing spindles first. Lower-tier
spindles are utilized as capacity demand grows. Relocation can be initiated manually
or by a user-configurable, automated scheduler.
The relocation process targets to create 10percent free capacity in the highest tiers in
the pool. Free capacity in these tiers is used for new slice allocations of high priority
LUNs between relocations.
Managing FAST VP at the storage pool level
FAST VP properties can be viewed and managed at the pool level. Figure 2 shows the
tiering information for a specific pool.

Figure 2. Storage Pool Properties window

11 EMC FAST VP for Unified Storage SystemsA Detailed Review
The Tier Status section of the window shows FAST VP relocation information specific
to the pool selected. Scheduled relocation can be selected at the pool level from the
drop-down menu labeled Auto-Tiering. This can be set to either Scheduled or Manual.
Users can also connect to the array-wide relocation schedule using the button located
in the top right corner. This is discussed in the Automated scheduler section. Data
Relocation Status displays what state the pool is in with regards to FAST VP. The
Ready state indicates that relocation can begin on this pool at any time. The amount
of data bound for a lower tier is shown next to Data to Move Down and the amount of
data bound for a higher tier is listed next to Data to Move Up. Below that is the
estimated time required to migrate all data within the pool to the appropriate tier.
In the Tier Details section, users can see the exact distribution of their data. This
panel shows all tiers of storage residing in the pool. Each tier then displays the
amount of data to be moved down and up, the total capacity available (user
capacity), the consumed capacity, and total capacity available.


12 EMC FAST VP for Unified Storage SystemsA Detailed Review
Physically allocated capacity versus reserved capacity



Figure 3. General and Tiering tabs of the Pool Properties dialog
Figure 3 shows the Tiering and General tabs in the Pool Properties dialog box. The
Consumed Capacity shown under Tiering tab shows the physically allocated capacity
for each tier. It does not show the capacity reserved for thick LUNs.
The Consumed Capacity shown under the General tab shows the physically allocated
capacity (for all of the tiers) and the reserved capacity that has not yet been physically
allocated. For this reason, the consumed capacity under the in the General tab is
greater than the sum of consumed capacities shown under the Tiering tab.
For example, in Figure 3, the General tab shows total Consumed Capacity of 3944.7
GB. However, if you add the Consumed capacities shown under the Tiering tab, the

13 EMC FAST VP for Unified Storage SystemsA Detailed Review
total is 3218.3 GB . The difference indicates there are 1 GB slices of one or more thick
LUNs that have yet been physically allocated (meaning they have not been written to
yet).
The same information is available through the CLI. The capacity figure returned for the
CLI command storagepool list represents the information in the General tab. The
capacity figure returned for the command storagepool list disktype tiers
represents the information in the Tiering tab.
Automated scheduler
The scheduler launched from the Pool Properties dialog boxs Relocation Schedule
button is shown in Figure 2. To this end, relocations can be scheduled to occur
automatically. It is recommended that relocations be scheduled during off-hours to
minimize any potential performance impact the relocations may cause. Figure 3
shows the Manage Auto-Tiering window.

Figure 4. Manage Auto-Tiering window
The data relocation schedule shown in Figure 4 initiates relocations everyday at
10:00 PM for a duration of eight hours (ending at 6:00 AM). You can select the days
on which the relocation schedule should run. In this example, relocations run seven
days a week, which is the default setting. From this status window, you can also
control the data relocation rate. The default rate is set to Medium so as not to
significantly impact host I/O. This rate will relocate data up to 300-400 GB per hour
3
If all of the data from the latest analysis is relocated within the window, it is possible
for the next iteration of analysis and relocation to occur. This happens if the initial
relocation completes and the next analysis occurs within the window. After the new
analysis, slices will be moved according to the latest slice rankings. Any data not
.

3
This rate depends on system type, array utilization, and other tasks competing for array resources. High utilization rates may
reduce this relocation rate.

14 EMC FAST VP for Unified Storage SystemsA Detailed Review
relocated during the window is simply re-evaluated as normal with each subsequent
analysis.
Manual relocation
Manual relocation is initiated by the user through either the Unisphere GUI or the CLI.
It can be initiated at any time. When a manual relocation is initiated, FAST VP
performs analysis on all statistics gathered, independent of its regularly scheduled
hourly analysis, prior to beginning the relocation. This ensures that up-to-date
statistics and settings are properly accounted for prior to relocation. When manually
initiating a relocation, the users specifies both the relocation rate and duration to be
for the relocation.
Although the automatic scheduler is an array-wide setting, manual relocation is
enacted at the pool level only. Common situations when users may want to initiate a
manual relocation on a specific pool include the following:
When reconfiguring the pool (for example, adding a new tier of drives)
When LUN properties have been changed and the new priority structure needs to
be realized immediately
As part of a script for a finer-grained relocation schedule
FAST VP LUN management
Some FAST VP properties are managed at the LUN level. Figure 5 shows the tiering
information for a single LUN.

Figure 5. LUN Properties window

15 EMC FAST VP for Unified Storage SystemsA Detailed Review
The Tier Details section displays the current distribution of 1 GB slices within the LUN.
The Tiering Policy section displays the available options for tiering policy.
Tiering policies
There are four tiering policies available within FAST VP:
Auto-tier (recommended)
Highest available tier
Lowest available tier
No data movement
Auto-tier
Auto-tier is the default setting for all pool LUNs upon their creation. FAST VP relocates
slices of these LUNs based on their activity level. Slices belonging to LUNs with the
auto-tier policy have second priority for capacity in the highest tier in the pool after
LUNs set to the highest tier.
Highest available tier
The highest available tier setting should be selected for those LUNs which, although
not always the most active, require high levels of performance whenever they are
accessed. FAST VP will prioritize slices of a LUN with highest available tier selected
above all other settings.
Slices of LUNs set to highest tier are rank ordered with each other according to
activity. Therefore, in cases where the sum total of LUN capacity set to highest tier is
greater than the capacity of the pools highest tier, the busiest slices occupy that
capacity.
Lowest available tier
Lowest available tier should be selected for LUNs that are not performance- or
response-time-sensitive. FAST VP will maintain slices of these LUNs on the lowest
storage tier available regardless of activity level.
No data movement
No data movement may only be selected after a LUN has been created. FAST VP will
not move slices from their current positions once the no data movement selection has
been made. Statistics are still collected on these slices for use if and when the tiering
policy is changed.
Initial placement
The tiering policy chosen also affects the initial placement of a LUNs slices within the
available tiers. Initial placement with the pool set to auto-tier will result in the data
being distributed across all storage tiers available within the pool. The distribution is

16 EMC FAST VP for Unified Storage SystemsA Detailed Review
based on available capacity in the pool. If 70 percent of a pools free capacity resides
in the lowest tier, then 70 percent of the new slices will be placed in that tier.
LUNs set to highest available tier will have their component slices placed on the
highest tier that has capacity available. LUNs set to lowest available tier will have
their component slices placed on the lowest tier that has capacity available.
LUNs with the tiering policy set to no data movement will use the initial placement
policy of the setting preceding the change to no data movement. For example, a LUN
that was previously set to highest tier but is currently set to no data movement will
still take its initial allocations from the highest tier possible.
Common uses for Highest Available Tiering policy
Response time sensitive applications
When a pool consists of LUNs with stringent response time demands, it is not
uncommon for users to set all LUNs in the pool to highest available tier. That way,
new LUN slices are allocated from the highest tier. Since new data is often the most
heavily used, this provides the best performance for those slices. At the same time, if
all LUNs in the pool are set to highest tier, slices are relocated based on their relative
activity to one another. Therefore, the TCO benefits of FAST VP are still fully realized.
Large scale migrations
The Highest Available Tier policy is useful for large scale migrations into a pool. When
you start the migration process, it is best to fill the highest tiers of the pool first. This
is especially important for live migrations. Using the Auto-tier setting would place
some data in the Capacity tier. At this point, FAST VP has not yet run an analysis on
the new data, so it cannot distinguish between hot and cold data. Therefore, with the
Auto-tier setting, some of the busiest data may be placed in the Capacity tier.
In these cases, the target pool LUNs can be set to highest tier. That way, all data is
initially allocated to the highest tiers in the pool. As the higher tiers fill and capacity
from the Capacity (NL-SAS) tier starts to be allocated, you can stop the migration and
run a manual FAST VP relocation. Assuming an analysis has had sufficient time to run,
relocation will rank order the slices and move data appropriately. In addition, since
the relocation will attempt to free 10 percent of the highest tiers, there is more
capacity for new slice allocations in those tiers.
You continue this iterative process while you migrate more data to the pool, and then
run FAST VP relocations when most of the new data is being allocated to the Capacity
tier. Once all of the data is migrated into the pool, you can make any tiering
preferences they see fit.

17 EMC FAST VP for Unified Storage SystemsA Detailed Review
Using FAST VP for file
In the VNX Operating Environment for File 7, file data is supported on LUNs created in
pools with FAST VP configured on both the VNX Unified and VNX gateways with EMC
Symmetrix

systems.
Management
The process for implementing FAST VP for file begins by provisioning LUNs from a
pool with mixed tiers (or across pools for Symmetrix) that are placed in the protected
File Storage Group. Rescanning the storage systems from the System tab in Unisphere
starts a diskmark that makes the LUNs available to VNX file storage. The rescan
automatically creates a pool for file using the same name as the corresponding pool
for block. Additionally it will create a disk volume in a 1:1 mapping for each LUN that
was added to the File Storage Group. A file system can then be created from the pool
for file on the disk volumes. The FAST VP policy that has been applied to the LUNs
presented to file will operate as it does for any other LUN in the system, dynamically
migrating data between storage tiers in the pool.

Figure 6. FAST VP for file
FAST VP for file is supported in Unisphere and the CLI. All relevant Unisphere
configuration wizards support a FAST VP configuration except for the VNX Provisioning
Wizard. FAST VP properties can be seen within the properties pages of pools for file
(see Figure 7 on page 18), and property pages for volumes and file systems (see
Figure 8 on page 19) but they can only be modified through the block pool or LUN
areas of Unisphere. On the File System Properties pages the FAST VP tiering policy is

18 EMC FAST VP for Unified Storage SystemsA Detailed Review
listed in the Advanced Data Services section along with if thin, compression, or
mirrored is enabled. If an Advanced Data Service is not enabled, it will not appear on
the screen. For more information on thin and compression reference the white papers
EMC VNX Virtual Provisioning and EMC VNX Deduplication and Compression.
Disk type options of the mapped disk volumes are as follows:
LUNs in a storage pool with a single disk type
Extreme Performance (Flash drives)
Performance (10k and 15k rpm SAS drives)
Capacity (7.2k rpm NL-SAS)
LUNs in a storage pool with multiple disk types (used with FAST VP)
Mixed
LUNs that are mirrored (mirrored means remote mirrored through MirrorView or
RecoverPoint)
Mirrored_mixed
Mirrored_performance
Mirrored_capacity
Mirrored_Extreme Performance

Figure 7. File Storage Pool Properties window

19 EMC FAST VP for Unified Storage SystemsA Detailed Review

Figure 8. File System Properties window
Best practices for VNX for File
The entire pool should be allocated to file systems.
The entire pool should use thick LUNs only.
Recommendations for Thick LUNs:
There should be one Thick LUN per physical disk in pool.
Pool LUN count should divisible by five to facilitate striping.
Balance SP ownership.
All LUNs in the Pool should have the same tiering policy. This is necessary to
support slice volumes.
Prior to VNX OE for File V7.0.3x, the use of AVM was not optimal due to the fact
that AVM would concatenate volumes as opposed to striping and slicing. For this
reason, wherever possible you should upgrade to V7.0.3x prior to
implementation. If this is not possible, Manual Volume Management can provide
reasonable performance configurations by creating stripe volumes manually and
building user defined pools.
EMC does not recommend that you use block thin provisioning or compression on
VNX LUNs used for file system allocations. Instead, you should use the File side
thin provisioning and File side deduplication (which includes compression).

20 EMC FAST VP for Unified Storage SystemsA Detailed Review
VNX file configurations, for both VNX and VNX gateways with Symmetrix, will not
expressly forbid mixing LUNs with different data service attributes, although users
will be warned that it is not recommended to mix due to the impact of spreading a
file system across, for example, a thin and a thick LUN or LUNs with different
tiering policies.
Note: VNX file configurations will not allow mixing of mirrored and non-mirrored types
in a pool If you try to do this, the disk mark will fail.
Where the highest levels of performance are required, maintain the use of RAID
Group LUNs for File.
Automatic Volume Manager rules
Automatic Volume Manager (AVM) rules are different when creating a file system with
underlying pool LUNs as opposed to file systems with underlying RAID group LUNs.
The rules for AVM with underlying pool LUNs are as follow:
For VNX, the following rules apply:
Primary VNX Algorithm:
1. Striping is tried first. If disk volumes cannot be striped, then concatenation is
tried.
2. AVM checks for free disk volumes:
If there are no free disk volumes and the slice option is set to No, there is not
enough space available and the request fails.
If there are free disk volumes:
I. AVM sorts them by thin and thick disk volumes.
II. AVM sorts the thin and thick disk volumes into size groups.
3. AVM first checks for thick disk volumes that satisfy the target number of disk
volumes to stripe. Default=5.
4. AVM tries to stripe five disk volumes together, with the same size, same data
service policies, and in an SP-balanced manner. If five disk volumes cannot be
found, AVM tries four, then three, and then two.
5. AVM selects SP-balanced disk volumes before selecting the same data service
policies.
6. If no thick disk volumes are found, AVM then checks for thin disk volumes that
satisfy the target number.
7. If the space needed is not found, AVM uses the VNX for block secondary pool-
based file system algorithm to look for the space.

21 EMC FAST VP for Unified Storage SystemsA Detailed Review
Note: For file system extension, AVM always tries to expand onto the existing volumes
of the file system. However, if there is not enough space to fulfill the size request on
the existing volumes, additional storage is obtained using the above algorithm and
AVM attempts to match the data service policies of the first used disk volume of the
file system. All volumes mentioned above, whether a stripe or a concatenation, are
sliced by default.
Secondary VNX Algorithm (used when space needed is not found in the Primary
Algorithm):
1. Concatenation is used. Striping is not used.
2. Unless requested, slicing will not be used.
3. AVM checks for free disk volumes, and sorts them by Thin and Thick disk volumes.
4. AVM checks for free disk volumes:
If there are no free disk volumes and the slice option is set to no, there is not
enough space available and the request fails.
If there are free disk volumes:
I. AVM first checks for Thick disk volumes that satisfy the size request (equal
to or greater than the file system size).
II. If not found, AVM then checks for Thin disk volumes that satisfy the size
request.
III. If still not found, AVM combines Thick and Thin disk volumes to find ones
that satisfy the size request.
5. If one disk volume satisfies the size request exactly, AVM takes the selected disk
volume and uses the whole disk to build the file system.
6. If a larger disk volume is found that is a better fit than any set of smaller disks,
then AVM uses the larger disk volume.
7. If multiple disk volumes satisfy the size request, AVM sorts the disk volumes from
smallest to largest, and then sorts in alternating SP A and SP B lists. Starting with
the first disk volume, AVM searches through a list for matching data services until
the size request is met. If the size request is not met, AVM searches again but
ignores the data services.

For VNX gateway with Symmetrix, the following rules apply:
1. Unless requested, slicing will not be used.
2. AVM checks for free disk volumes, and sorts them by thin and thick disk volumes
for the purpose of striping together the same type of disk volumes:
If there are no free disk volumes and the slice option is set to no, there is not
enough space available and the request fails

22 EMC FAST VP for Unified Storage SystemsA Detailed Review
If there are free disk volumes:
I. AVM first checks for a set of eight disk volumes.
II. If a set of eight is not found, AVM then looks for a set of four disk volumes.
III. If a set of four is not found, AVM then looks for a set of two disk volumes.
IV. If a set of two is not found, AVM finally looks for one disk volume.
3. When free disk volumes are found:
a. AVM first checks for thick disk volumes that satisfy the size request, which can
be equal to or greater than the file system size. If thick disk volumes are
available, AVM first tries to stripe the thick disk volumes that have the same
disk type. Otherwise, AVM stripes together thick disk volumes that have
different disk types.
b. If thick disks are not found, AVM then checks for thin disk volumes that satisfy
the size request. If thin disk volumes are available, AVM first tries to stripe the
thin disk volumes that have the same disk type, where same means the
single disk type of the pool in which it resides. Otherwise, AVM stripes
together thin disk volumes that have different disk types.
c. If thin disks are not found, AVM combines thick and thin disk volumes to find
ones that satisfy the size request.
4. If neither thick nor thin disk volumes satisfy the size request, AVM then checks for
whether striping of one same disk type will satisfy the size request, ignoring
whether the disk volumes are thick or thin.
5. If still no matches are found, AVM checks whether slicing was requested.
a. If slicing was requested, then AVM checks whether any stripes exist that
satisfy the size request. If yes, then AVM slices an existing stripe.
b. If slicing was not requested, AVM checks whether any free disk volumes can
be concatenated to satisfy the size request. If yes, AVM concatenates disk
volumes, matching data services if possible, and builds the file system.
6. If still no matches are found, there is not enough space available and the request
fails.
Managing Volumes and File Systems with VNX AVM provides further information on
using AVM with mapped pools.

23 EMC FAST VP for Unified Storage SystemsA Detailed Review
General Guidance and recommendations
The following table displays the total number of LUNs that can be set to leverage FAST
VP based on the array model. These limits are the same as the total number of pool
LUNs per system. Therefore, all pool LUNs in any given system can leverage FAST VP.
Table 2. FAST VP LUN limits
Array model VNX5300
CX4- 120
VNX5500


CX4- 240
VNX5700


CX4- 480
VNX7500


CX4- 960
Maximum number of pool LUNs 512 1,024 2,048 2,048
FAST VP and FAST Cache
FAST Cache allows the storage system to provide Flash drive class performance to the
most heavily accessed chunks of data across the entire system. FAST Cache absorbs
I/O bursts from applications, thereby reducing the load on back-end hard disks. This
improves the performance of the storage solution. For more details on this feature,
refer to the EMC CLARiiON, Celerra Unified, and VNX FAST Cache white paper available
on Powerlink.
The following table compares the FAST VP and FAST Cache features.
Table 3. Comparison between the FAST VP and FAST Cache features
FAST Cache FAST VP
Enables Flash drives to be used to
extend the existing caching capacity of
the storage system.
Leverages pools to provide sub-LUN tiering, enabling the
utilization of multiple tiers of storage simultaneously
Has finer granularity 64 KB Less granular compared to FAST Cache 1 GB
Copies data from HDDs to Flash drives
when they get accessed frequently
Moves data between different storage tiers based on a
weighted average of access statistics collected over a
period of time
Continuously adapts to changes in
workload
Uses a relocation process to periodically make storage
tiering adjustments. Default setting is one, 8 hour
relocation per day.
Is designed primarily to improve
performance
While it can improve performance, it is primarily designed
to improve ease of use and reduce TCO
FAST Cache and the FAST VP sub-LUN tiering features can be used together to yield
high performance and improved TCO for the storage system. As an example, in
scenarios where limited Flash drives are available, they can be used to create FAST
Cache and the FAST VP can be used on a two-tier, Performance and Capacity pool.
From a performance point of view, FAST Cache will dynamically provide performance
benefit to any bursty data while FAST VP will move warmer data to Performance drives
and colder data to Capacity drives. From a TCO perspective, FAST Cache with a small
number of Flash drives serves the data that is accessed most frequently, while FAST
VP can optimize disk utilization and efficiency.

24 EMC FAST VP for Unified Storage SystemsA Detailed Review
As a general rule, FAST Cache should be used in cases where storage system
performance needs to be improved immediately for burst-prone data. On the other
hand, FAST VP optimizes storage system TCO as it moves data to the appropriate
storage tier based on sustained data access and demands over time. FAST Cache
focuses on improving performance while FAST VP focuses on improving TCO. Both
features are complementary to each other and help in improving performance and
TCO.
The FAST Cache feature is storage-tier-aware and works with the FAST VP to make sure
that the storage system resources are not wasted by unnecessarily copying data to
FAST Cache if it is already on a Flash drive. If FAST VP moves a chunk of data to the
Extreme Performance Tier (which consists of Flash drives) , FAST Cache will not
promote that chunk of data into FAST Cache, even if FAST Cache criteria is met for
promotion. This ensures that the storage system resources are not wasted in copying
data from one Flash drive to another.
A general recommendation for the initial deployment of Flash drives in a storage
system is to use them for FAST Cache. However, in certain cases FAST Cache does not
offer the most efficient use of Flash drives. FAST Cache tracks I/Os that are smaller
than 128 KB, and requires multiple hits to 64 KB chunks to initiate promotions from
rotating disk to FAST Cache. Therefore, I/O profiles that do not meet this criteria are
better served by Flash in a pool or RAID group.
What drive mix is right for my I/O profile?
As previously mentioned, it is common for a small percentage of overall capacity to
be responsible for most of the I/O activity. This is known as skew. Analysis of an I/O
profile may indicate that 85 percent of the I/Os to a volume only involve 15 percent of
the capacity. The resulting active capacity is called the working set. Software like
FAST VP and FAST Cache keep the working set on the highest-performing drives.
It is common for OLTP environments to yield working sets of 20 percent or less of their
total capacity. These profiles hit the sweet spot for FAST and FAST Cache. The white
papers Leveraging Fully Automated Storage Tiering (FAST) with Oracle Database
Applications and EMC Tiered Storage for Microsoft SQL Server 2008Enabled by EMC
Unified Storage and EMC Fully Automated Storage Tiering (FAST) discuss performance
and TCO benefits for several mixes of drive types.
Other I/O profiles, like Decision Support Systems (DSS), may have much larger
working sets. In these cases, FAST VP may be used to deploy Flash drives because
DSS workloads are not typically FAST Cache-friendly. Capacity Tier drives may be used
to lower TCO. The white paper Leveraging EMC Unified Storage System Dynamic LUNs
for Data Warehouse Deployments on Powerlink offers analysis on the use of storage
pools and FAST VP.
At a minimum, the capacity across the Performance Tier and Extreme Performance Tier
(and/or FAST Cache) should accommodate the working set. However, capacity is not
the only consideration. The spindle count of these tiers needs to be sized to handle
the I/O load of the working set. Basic techniques for sizing disk layouts based on

25 EMC FAST VP for Unified Storage SystemsA Detailed Review
IOPS and bandwidth are available in the EMC VNX Fundamentals for Performance and
Availability white paper on Powerlink.
As discussed above, using FAST Cache as your first line of Flash-drive deployment is a
practical approach when the I/O profile is amenable to its use. In circumstances
where an I/O profile is not FAST-Cache friendly, Flash can be deployed in a pool or
RAID group instead.
Performance Tier drives are versatile in handling a wide spectrum of I/O profiles.
Therefore, we highly recommend that you include Performance Tier drives in each
pool. FAST Cache can be an effective tool for handling a large percentage of activity,
but inevitably, there will be I/Os that have not been promoted or are cache misses. In
these cases, Performance Tier drives offer good performance for those I/Os.
Performance Tier drives also facilitate faster promotion of data into FAST Cache by
quickly providing promoted 64 KB chunks to FAST Cache. This minimizes FAST Cache
warm-up time as some data gets hot and other data goes cold. Lastly, if FAST Cache is
ever in a degraded state due to a faulty drive, the FAST Cache will become read only. If
the I/O profile has a significant component of random writes, these are best served
from Performance Tier drives as opposed Capacity drives.
Capacity drives can be used for everything else. This often equates to comprising
60 percent to 80 percent of the pools capacity. Of course, there are profiles with low
IOPS/GB and or sequential workload that may result in the use of a higher percentage
of Capacity Tier drives.
EMC Professional Services and qualified partners can be engaged to assist with
properly sizing tiers and pools to maximize investment. They have the tools and
expertise to make very specific recommendations for tier composition based on an
existing I/O profile.
Conclusion
Through the use of FAST VP, users can remove complexity and management overhead
from their environments. FAST VP utilizes Flash, performance, and capacity drives (or
any combination thereof) within a single pool. LUNs within the pool can then
leverage the advantages of each drive type at the 1 GB slice granularity. This sub-
LUN-level tiering ensures that the most active dataset resides on the best-performing
drive tier available, while maintaining infrequently used data on lower-cost, high-
capacity drives.
Relocations can occur without user interaction on a predetermined schedule, making
FAST VP a truly automated offering. In the event that relocation is required on-
demand, FAST VP relocation can be invoked through Unisphere on an individual pool.
Both FAST VP and FAST Cache work by placing data segments on the most
appropriate storage tier based on their usage pattern. These two solutions are
complementary because they work on different granularity levels and time tables.

26 EMC FAST VP for Unified Storage SystemsA Detailed Review
Implementing both FAST VP and FAST Cache can significantly improve performance
and reduce cost in the environment.
References
The following white papers are available on Powerlink:
EMC Unified Storage Best Practices for Performance and Availability Common
Platform and Block Applied Best Practices
EMC VNX Virtual Provisioning
EMC CLARiiON Storage System Fundamentals for Performance and Availability
EMC CLARiiON, Celerra Unified, and VNX FAST Cache
EMC Unisphere: Unified Storage Management Solution
An Introduction to EMC CLARiiON and Celerra Unified Platform Storage Device
Technology
EMC Tiered Storage for Microsoft SQL Server 2008Enabled by EMC Unified
Storage and EMC Fully Automated Storage Tiering (FAST)
Leveraging Fully Automated Storage Tiering (FAST) with Oracle Database
Applications
Leveraging EMC Unified Storage System Dynamic LUNs for Data Warehouse
Deployments

Você também pode gostar