Management Version 3
Lab Guide
August 2015
Copyright ©2015 EMC Corporation. All Rights Reserved. Published in the USA. EMC believes the information in
this publication is accurate as of its publication date. The information is subject to change without notice.
THE INFORMATION IN THIS PUBLICATION IS PROVIDED “AS IS.” EMC CORPORATION MAKES NO REPRESENTATIONS
OR WARRANTIES OF ANY KIND WITH RESPECT TO THE INFORMATION IN THIS PUBLICATION, AND SPECIFICALLY
DISCLAIMS IMPLIED WARRANTIES OF MERCHANTABILITY OR FITNESS FOR A PARTICULAR PURPOSE.
Use, copying, and distribution of any EMC software described in this publication requires an applicable software
license. The trademarks, logos, and service marks (collectively "Trademarks") appearing in this publication are
the property of EMC Corporation and other parties. Nothing contained in this publication should be construed as
granting any license or right to use any Trademark without the prior written permission of the party that owns the
Trademark.
EMC, EMC² AccessAnywhere Access Logix, AdvantEdge, AlphaStor, AppSync ApplicationXtender, ArchiveXtender,
Atmos, Authentica, Authentic Problems, Automated Resource Manager, AutoStart, AutoSwap, AVALONidm,
Avamar, Bus-Tech, Captiva, Catalog Solution, C-Clip, Celerra, Celerra Replicator, Centera, CenterStage,
CentraStar, EMC CertTracker. CIO Connect, ClaimPack, ClaimsEditor, Claralert ,cLARiiON, ClientPak, CloudArray,
Codebook Correlation Technology, Common Information Model, Compuset, Compute Anywhere, Configuration
Intelligence, Configuresoft, Connectrix, Constellation Computing, EMC ControlCenter, CopyCross, CopyPoint, CX,
DataBridge , Data Protection Suite. Data Protection Advisor, DBClassify, DD Boost, Dantz, DatabaseXtender, Data
Domain, Direct Matrix Architecture, DiskXtender, DiskXtender 2000, DLS ECO, Document Sciences,
Documentum, DR Anywhere, ECS, elnput, E-Lab, Elastic Cloud Storage, EmailXaminer, EmailXtender , EMC
Centera, EMC ControlCenter, EMC LifeLine, EMCTV, Enginuity, EPFM. eRoom, Event Explorer, FAST, FarPoint,
FirstPass, FLARE, FormWare, Geosynchrony, Global File Virtualization, Graphic Visualization, Greenplum,
HighRoad, HomeBase, Illuminator , InfoArchive, InfoMover, Infoscape, Infra, InputAccel, InputAccel Express,
Invista, Ionix, ISIS,Kazeon, EMC LifeLine, Mainframe Appliance for Storage, Mainframe Data Library, Max
Retriever, MCx, MediaStor , Metro, MetroPoint, MirrorView, Multi-Band Deduplication,Navisphere, Netstorage,
NetWorker, nLayers, EMC OnCourse, OnAlert, OpenScale, Petrocloud, PixTools, Powerlink, PowerPath,
PowerSnap, ProSphere, ProtectEverywhere, ProtectPoint, EMC Proven, EMC Proven Professional, QuickScan,
RAPIDPath, EMC RecoverPoint, Rainfinity, RepliCare, RepliStor, ResourcePak, Retrospect, RSA, the RSA logo,
SafeLine, SAN Advisor, SAN Copy, SAN Manager, ScaleIO Smarts, EMC Snap, SnapImage, SnapSure, SnapView,
SourceOne, SRDF, EMC Storage Administrator, StorageScope, SupportMate, SymmAPI, SymmEnabler, Symmetrix,
Symmetrix DMX, Symmetrix VMAX, TimeFinder, TwinStrata, UltraFlex, UltraPoint, UltraScale, Unisphere, Universal
Data Consistency, Vblock, Velocity, Viewlets, ViPR, Virtual Matrix, Virtual Matrix Architecture, Virtual
Provisioning, Virtualize Everything, Compromise Nothing, Virtuent, VMAX, VMAXe, VNX, VNXe, Voyence, VPLEX,
VSAM-Assist, VSAM I/O PLUS, VSET, VSPEX, Watch4net, WebXtender, xPression, xPresso, Xtrem, XtremCache,
XtremSF, XtremSW, XtremIO, YottaYotta, Zero-Friction Enterprise Storage.
3
4
EXERCISE 3: STORAGE PROVISIONING AND TIERING ........................................... 13
EXERCISE 4: FC SAN TOPOLOGIES ....................................................................... 15
EXERCISE 5: MTBF AND MTTR ............................................................................ 19
EXERCISE 6: INFORMATION AVAILABILITY .......................................................... 21
EXERCISE 7: BACKUP AND ARCHIVE .................................................................... 23
EXERCISE 8: REPLICATION .................................................................................. 27
5
6
Purpose: To reinforce the concepts presented in the lecture portion of
the module, Intelligent Storage Systems (ISS)
Participants are required to provide a solution for the
Tasks:
deliverables based on the given scenario and requirements.
References: Module 4: Intelligent Storage Systems (ISS)
7
An organization plans to deploy a new business application in their data center. The new application
requires 1.5 TB of storage capacity for business and application data. During peak workload period,
the application is expected to generate 5200 IOPS (I/Os per second) with a typical I/O size of 4 KB. The
available disk drive option is a 15,000 RPM drive with 250 GB capacity. The disk drive has an average
seek time of 4.2 milliseconds (ms) and a data transfer rate of 80 MB/s.
Requirements
The application is business critical and its response time must be within an acceptable range.
Deliverables
Calculate the number of disk drives that is required to meet both the capacity and the performance
requirements of the application.
Solution
Step 1: Calculate the time required to perform one I/O. This depends on the disk service time.
o Disk service time = Average seek time + Rotational latency + Data transfer time.
o The average seek time = 4.2 ms (given).
o The rotational latency is half the time taken for a full rotation. Since the rotation speed
is 15,000 RPM, one rotation takes 60 / 15000 seconds. Therefore the time taken for
half a rotation is 0.5 x (60 / 15000) = 2 ms.
o The data transfer rate is 80 MB/s. Therefore, the transfer of one 4 KB I/O takes 4 KB /
80 MB/s = 0.05 ms.
o Therefore, disk service time = 4.2 ms + 2 ms + 0.05 ms = 6.25 ms.
Step 2: Calculate the maximum number of IOPS the disk drive can perform:
o Maximum number of IOPS is 1 / 6.25 ms = 160 IOPS.
o For an acceptable response time, the disk controller utilization must be less than 70%
Therefore, the maximum number of IOPS at 70% utilization is 160 X 0.7 = 112 IOPS.
Step 3: Calculate the number of disk drives for the application:
o Disk drives required to meet performance requirement = 5200 IOPS / 112 IOPS = 47
o Disk drives required to meet capacity requirement = 1.5 TB / 250 GB = 6
o The number of disk drives required for the application is:
Maximum (Capacity, Performance) = Maximum (6, 47) = 47 disk drives
End of Exercise
8
Purpose: To reinforce the concepts presented in the lecture portion of
the module, Intelligent Storage Systems (ISS)
Participants are required to provide a solution for the
Tasks:
deliverables based on the given scenario and requirements.
References: Module 4: Intelligent Storage Systems (ISS)
9
An organization has multiple regional offices across the country. All financial activity is managed and
tracked by a single accounting application. The application is hosted on a dedicated compute system
with storage disks configured with RAID 0. The application generates around 15% random write
operations, and the remaining 85% operations are random reads. The accounting data is currently
stored on a 5‐disk RAID 0 set. Each disk has an advertised formatted capacity of 250 GB, and the total
size of the application’s data is 900 GB. The company performs nightly backups and removes old
information. Therefore, the amount of data is unlikely to change significantly over the next six
months.
Requirements
The organization wants the accounting data to be highly available. For this, the organization plans to
implement fault tolerance in the storage infrastructure.
Challenges
The organization’s IT budget is depleted as it is the end of the financial year. Therefore, it is not
possible for the organization to purchase any new disk drives.
Deliverables
Recommend a RAID level that the organization can use to reconfigure their storage while
fulfilling their needs.
Justify the choice of RAID level based on cost, performance, and availability of the new
solution.
Solution
Step 1: Consider the formatted capacity of the disks. A 250 GB disk would hold about 232.8 GB
of application data. The total size of the application data is 900 GB, which will fit on 4 disks. For
fault tolerance, consider a solution which uses parity‐based RAID. RAID 3 and RAID 5 are the
two solutions that will not require the purchase of any additional disks.
Step 2: Choice of RAID level and its justification:
o RAID 3 is useful only in environments where the data access is in large, sequential
blocks. Therefore, it is excluded.
o RAID 5 would perform reasonably well because the proportion of writes does not
exceed 25%. It is therefore recommended that the disks be reconfigured with RAID 5.
The advantage is that it provides low‐cost data protection. The disadvantage is that if a
disk fails, there will be a high overhead due to the disk rebuild operation.
10
11
12
Purpose: To reinforce the concepts presented in the lecture portion of
the module, Block‐based Storage System
Participants are required to provide a solution for the
Tasks:
deliverables based on the given scenario and requirements.
References: Module 5: Block‐based Storage System
13
An organization plans to deploy an OLTP application and a back‐office application in their
environment. The organization has a block‐based storage system with SSD, FC drives, and SATA drives.
The OLTP application is mission‐critical and will generate a large number of small‐sized, random I/Os
with the number of reads expected to be around 75%. The back‐office application will be deployed for
non‐production operations.
Requirements
The organization is concerned about the throughput and response time of the OLTP application. The
OLTP application requires high performance with low latency. The storage administrator wants full
control of the storage provisioned for this application for precise data placement on physical drives.
The back‐office application can tolerate performance variations. The capacity requirement of the
back‐office application is unpredictable and may vary.
Deliverables
Suggest a storage provisioning and tiering strategy to meet the organization’s requirements.
Solution
Use traditional storage provisioning for the OLTP application. This will enable meeting the
application’s performance and response time requirements, while providing greater control to
the storage administrator. Since the OLTP application is read‐heavy, use server flash caching
on the compute system hosting the application to improve the application’s performance.
Use virtual storage provisioning for the back‐office application. This enables oversubscription,
dynamic capacity expansion, and improved storage capacity utilization.
End of Exercise
14
Purpose: To reinforce the concepts presented in the lecture portion of
the module, Fibre Channel (FC) SAN
Participants are required to provide a solution for the
Tasks:
deliverables based on the given scenario and requirements.
References: Module 9: Fibre Channel (FC) SAN
15
The storage infrastructure of an IT organization includes three block‐based storage systems direct‐
attached to a heterogeneous mix of 45 compute systems. All compute systems are dual‐attached to
the storage systems for high availability. Because each storage system has 32 front‐end ports, each
could support a maximum of 16 compute systems. However, each existing storage system has the
storage drive capacity to support a maximum of 32 compute systems.
Challenges and Requirements
The organization plans to purchase 45 more compute systems to meet its growth requirements. If it
continues using direct‐attached storage, the organization needs to purchase additional storage
systems to connect these new compute systems. The organization realizes that its existing storage
systems are poorly utilized; therefore, it plans to implement FC SAN to overcome the scalability and
utilization challenges. The organization uses high‐performance applications; therefore, it wants to
minimize the number of ISLs or hop count required for compute‐to‐storage traffic.
Deliverables
Given that 72‐port FC switches are available for interconnectivity, propose a fabric topology to
address organization’s challenges and requirements. Justify your choice of the fabric topology. In
addition, determine the minimum number of switches required in the fabric.
Solution
Single‐switch topology cannot be implemented because the available FC switch does not provide the
required number of ports for node connectivity. Full mesh topology is not suitable for an environment
that requires high scalability. Partial mesh although provides more scalability than full mesh, but
several hops or ISLs may be required for the network traffic to reach its destination.
Therefore, the recommended solution is core‐edge topology. The core‐edge topology provides higher
scalability than mesh topology. It also provides a maximum of one‐hop or one‐ISL storage access to all
compute systems in the environment. As ISLs between the edge switches are not needed, the core‐
edge topology conserves the overall switch port utilization which increases connectivity within the
fabric.
Total number of storage system ports = 3 storage systems × 32 ports = 96 ports.
Number of switches required at the core = 96 storage system ports / 72 ports per switch ≈ 2 switches.
The core switches provide 144 ports of which 96 ports will be used for storage system connectivity.
The remaining 48 ports can be used for ISLs and future growth.
Number of switches at the edge = 180 compute system ports / 72 ports per switch ≈ 3 switches.
The edge switches provide 216 ports of which 180 ports will be used for compute system connectivity.
Remaining 36 ports can be used for ISLs and future growth.
So, at minimum, two core switches and three edge switches are required to implement the core‐edge
fabric.
End of Exercise
17
18
Purpose: To reinforce the concepts presented in the lecture portion of
the module, Introduction to Business Continuity
Participants are required to provide a solution for the
Tasks:
deliverables based on the given scenario and requirements.
References: Module 12: Introduction to Business Continuity
19
A system has three components and requires all three to be operational for 24 hours from Monday to
Friday. Failure of component 1 occurs as follows:
Monday = No failure
Tuesday = 5 am to 7 am
Wednesday = No failure
Thursday = 4 pm to 8 pm
Friday = 8 am to 11 am
Deliverables
Calculate the MTBF and MTTR of component 1.
Solution
MTBF is calculated as: total uptime/number of failures
Total downtime = 2 hours on Tuesday + 4 hours on Thursday + 3 hours on Friday = 9 hours
Total uptime = (5 × 24) – 9 = 111 hours
So, MTBF = 111 / 3 = 37 hours
MTTR is calculated as: total downtime/number of failures
Total downtime = 2 hours on Tuesday + 4 hours on Thursday + 3 hours on Friday = 9 hours
So, MTTR = 9 hours / 3 = 3 hours
End of Exercise
20
Purpose: To reinforce the concepts presented in the lecture portion of
the module, Introduction to Business Continuity
Participants are required to provide a solution for the
Tasks:
deliverables based on the given scenario and requirements.
References: Module 12: Introduction to Business Continuity
21
A system has three components and requires all three to be operational from 8 am to 5 pm, Monday
to Friday. Failure of component 2 occurs as follows:
Monday = 8 am to 11 am
Tuesday = No failure
Wednesday = 4 pm to 7 pm
Thursday = 5 pm to 8 pm
Friday = 1 pm to 2 pm
Deliverables
Calculate the availability of component 2.
Solution
Availability is calculated as: system uptime / (system uptime + system downtime)
System downtime = 3 hours on Monday + 1 hour on Wednesday + 1 hour on Friday = 5 hours
System uptime = total operational time – system downtime = 45 hours – 5 hours = 40 hours
Availability (%) = (40 / 45) × 100 = 88.9%
End of Exercise
22
Purpose: To reinforce the concepts presented in the lecture portion of
the module, Backup and Archive
Participants are required to provide a solution for the
Tasks:
deliverables based on the given scenario and requirements.
References: Module 13: Backup and Archive
23
A major multinational bank runs business‐critical applications in a virtualized data center. It has over a
million of customers and multiple remote/branch offices (ROBO) across different geographic
locations. The bank currently uses tape as its primary backup storage media for backing up virtual
machines (VMs) and application data. It uses an agent‐based backup solution for backing up data. It
currently performs a full backup every Sunday and an incremental backup on other days. It also has a
file‐sharing environment in which multiple NAS systems serve all the users. During NAS backup, the
data is backed up from application servers to backup device. Approximately 25% of data in the
production environment is inactive data (fixed content).
Challenges
The backup operations consume resources on the compute systems that are running multiple VMs.
This is significantly impacting the applications deployed on the VMs. During NAS backup, the
application servers are impacted because the data is backed up from these servers to the backup
device.
The organization’s backup environment has a huge amount of redundant data. This has led to an
increase in the infrastructure cost and is also impacting the backup window (backup takes more time
to complete). Recovering data or VMs also takes more time and affects the SLAs.
The branch offices also have limited IT resources for managing backup. Backing up data from branch
offices to a centralized data center was restricted due to the time and cost involved in sending huge
volumes of data over the WAN.
The organization incurs a huge investment and operational expense in managing an offsite backup
infrastructure at remote site for disaster recovery purpose.
Requirements
1. Need faster backup and restore to meet the SLAs
2. Need to eliminate redundant copies of data
3. Need an effective solution to address the backup and recovery challenges of remote and
branch offices
24
Deliverables
Recommend solutions that will meet the organization’s requirements.
Solution
1. Implement disk‐based backup solution to improve the backup and recovery performance for
meeting SLAs
2. Implement deduplication solution to eliminate the redundant copies of data
3. Disk‐based backup solutions along with source‐based deduplication
i. Eliminates the challenges associated with centrally backing up branch office data
ii. Deduplication considerably reduces the required network bandwidth
4. Implement image‐based backup that helps to offload backup operation from VMs to a proxy
server
i. No backup agent is required inside the VM to backup
5. Deploy NDMP‐based backup solution for NAS environment
i. In NDMP‐based backup, data is sent directly from the NAS head to the backup device
without impacting application servers
6. Organization can implement data archiving solutions that archive fixed content from the
production environment
i. Reduce the amount of data to be backed up
7. Organization could choose backup as a service to replicate the backup copy to the cloud
instead of the offsite backup
i. Saves CAPEX and reduces the management overhead to the organization
End of Exercise
25
26
Purpose: To reinforce the concepts presented in the lecture portion of
the module, Replication
Participants are required to provide a solution for the
Tasks:
deliverables based on the given scenario and requirements.
References: Module 14: Replication
27
An organization runs a business‐critical application that stores data in a LUN with RAID 1
configuration. The size of the LUN is 1 TB. The average data that changes in 24 hours is 60 GB. The
organization has two data centers which are 1000 miles apart.
Requirements
• Organization needs a solution to address logical corruption of data
• RPO requirement for the application is 1 hour
• In the event of a logical corruption, up to 8 hours of old data needs to be restored
• The amount of storage used for data protection should be minimal
• The organization also requires a disaster recovery (DR) solution that should not impact the
response time of the application
Deliverables
• Suggest an appropriate storage system‐based local replication solution to meet the RPO
requirement with minimum amount of storage. Estimate the storage required for the solution.
• Recommend an appropriate remote replication solution for DR purpose.
Solution
• Pointer‐based virtual replication technology is the most obvious choice because the changes to
the source are less than 30% and the solution requires less storage space.
• The organization requires to create a pointer‐based virtual replica every hour and to maintain
each replica for 8 hours. Therefore, a total of 8 replicas need to be maintained to meet
organization’s requirement. If a logical corruption occurs, data can be restored from the latest
replica.
o At Hour 01:00 Create Replica 1
o At Hour 02:00 Create Replica 2
o At Hour 03:00 Create Replica 3
28
End of Exercise
29
30