ETHYLENE VINYL ACETATE EVA GLASS EVA SAFETY GLASS EVA LAMINATED GLASS COLORED EVA GLASS COLLECTIONS
EVA Overview
Transcript of EVA Overview
© 2007 Hewlett-Packard Development Company, L.P.The information contained herein is subject to change without notice
Enterprise VirtualArrays
Choong Ming TzeTechnical [email protected] 2007
The EVA familyA technical overview
HP StorageWorks product portfolio
The B-Series SAN switch family
Common Fabric OS 5.x
Fabric ManagerEnhanced capabilities
SAN Director 4/25632-384 4Gb ports
FICON support
HP 400 MP-Router(16FC + 2IP ports)
SAN Switch 4/64(32-64 ports)
SAN Switch 4/32(16-32 ports)
SAN Switch 4/8 & 4/168 and 16 ports
Brocade 4Gb SAN Switchfor HP p-class BladeSystem
Brocade 4Gb SAN Switchfor HP c-Class BladeSystem
HP MPR BladeFor the 4/256
4/48 Port BladeFor the 4/256
The C-series SAN switch familySmall & Medium-Sized
BusinessEnterprise & Service Provider
MDS 9000 FamilySystems
MDS 9020*
MDS 9120and 9140
MDS 9216and 9216i
MDS 9506MDS 9509 MDS 9513
Supervis
or1 and 2
14-Port, 16-Port, 32-Port 1 & 2 Gb FC
12-Port, 24-Port, 48-Port 1, 2 & 4Gb FC
4-Port10Gb FC
IP Storage Services – iSCSI
and FCIP
SSM (Virtualization;
Intelligent fabric
Applications)
MDS 9000 Modules
Mgmt.
OS
Cisco Fabric Manager
Cisco MDS 9000 Family SAN-OS*FabricWare
StorageWorks market coverage
MSA20, 30, 50
MSA500
MSA1000
MSA1500MSA1510i
EVA4000EVA6000
EVA8000XP 10000 XP 12000
Availa
bili
ty
Scalability
plugged into the data center fabric
to maximize scalability and
availability• High connectivity• High scalability• High efficiency• Highest disaster
tolerance solutions
• Universal connectivity and heterogeneity
Enterprise
Departmentalflexible and
scalable entry-level fibre
channel storage
• Scalable modularity
• Heterogeneous• Ease of
administration• Price/scalability
Branch OfficeSimple, affordable, fault tolerant Smart
Array technology
• clustering & shared storage
• minimal infrastructure
• DtS conversion• price/availability
high performance internal / external storage with Smart Array technologies
• price/capacity
Workgroups
MSA60, 70
Transition SlideThe HP StorageWorks Enterprise Virtual Array EVA
The EVA family
EVA4000
EVA6000
EVA8000
Leading in array virtualization and ease of use
*Note: Legacy 36GB FC and 250, 400GB FATA disks are still fully supported
• A revolutionary redesign of the proven EVA3000 and EVA5000 Storage Arrays
• Three family members for a broad range of prices, storage capacities and performance
• 4Gbps FC Controller• iSCSI Connectivity Option • Concurrent support of various FC and FATA Disks in
the same Disk Enclosures− 72, 146, 300GB FC
− 250, 400, 500GB FATA
• Virtual RAID Arrays: Vraid0, Vraid1, Vraid5• Industry standard multi-path failover support
− MPIO
− Pvlink
− DMP etc.
• Native HBAs Support (Sun, IBM, HP)• Local and remote copy support• Broad range of solutions and integrations available
The EVA family specifications
EVA4000
EVA6000
EVA8000
EVA4000 EVA6000 EVA8000Controller HSV200 HSV210
Cache size 4GB 8GB
RAID Levels VRAID0, VRAID1, VRAID5
Supported OSWindows 2000/2003, HP-UX, Linux, IBM AIX, OpenVMS, Tru64, SUN Solaris, VMWare, Netware
Supported Drives
FC: 72, 146GB/15krpm, 146, 300GB/10krpm
FATA: 250, 400, 500GB
Host ports 4 8
Device ports 4 8
Mirror ports 4
Backend loop switches 0 2 4
# of Drives 8 - 56 16 – 112 8 – 240
# of Enclosures 1 - 4 4 - 8 2 – 18
Max Capacity 28TB 56TB 120TB
The EVA4000 architecture
ManagementServer (Windows)
• 1 to 4 Disk enclosures
• 8 to 56 FC Disks
• 2 HSV Controllers
Fabric 1
HSV200 controller 2HSV200 controller 1
• 4Gbps Front-End
Fabric 2
Heterogeneous Servers
The EVA6000 architecture
Fabric 1
• 4-8 Disk enclosures
• 16 to 112 FC Disks
• 2 HSV Controllers
• 2 FC Loop SwitchesFC loop switch FC loop switch
Fabric 2
HSV200 controller 2HSV200 controller 1
ManagementServer (Windows)
• 4Gbps Front-End
The EVA8000 architecture
• 2-18 Disk enclosures12 in the first rack6 in the utility cabinet
• 8 to 240 FC Disks
• 2 HSV Controllers
• 4 FC Loop Switches
FC loop switch FC loop switchFC loop switch FC loop switch
Fabric 1 Fabric 2
HSV210 controller 2HSV210 controller 1• 4Gbps Front-
End
ManagementServer (Windows)
Heterogeneous Servers
EVA Performance (based on 2GB controllers)
Controller limits: 100% cache hits
Workload EVA5000 EVA8000
512 B Reads IOPs 141’000 215’000
256kB Reads MB/s 700 1600
Maximum Data Transfer Rates for 128 KB Sequential Workloads (MB/s)
Workload EVA4000 EVA6000 EVA5000 EVA8000
Reads 340 770 530 1,430
Vraid 1 Writes 160 355 165 530
Vraid 5 Writes 260 515 153 525
Throughputs (IOPs) under Random Workloads (4 KB Transfers @<30ms)
Workload EVA4000 EVA6000 EVA5000 EVA8000
Reads 14,500 27,600 50,000 55,900
Vraid 1 Writes 8,000 15,200 20,600 22,300
Vraid 5 Writes 4,400 8,000 12,200 13,000
Vraid 1 OLTP (60r/40w)
11,300 21,200 30,400 32,600
Vraid 5 OLTP (60r/40w)
7,000 13,900 22,100 23,300
Transition SlideThe Benefits of the EVA Virtualization
Traditional Disk Array Approach
RAID Controller
Traditional Disk Array Approach
RAID5RAID1
RAID Controller
Sp
are
Sp
are
Disk Groups & RAID Level
Dedicated Spare Disk(s)
Traditional Disk Array Approach
RAID5RAID1
LUN 1
LUN 0
LUN 2
RAID Controller
0 1 2S
pare
Sp
are
Presented LUNs
Traditional Disk Array Approach
RAID5RAID1
RAID0RAID5 RAID1
LUN 1
LUN 0
LUN 2
LUN 7
LUN 6LUN 3LUN 4
LUN 5
RAID Controller
0 1 2 3 4 5 6 7
Sp
aS
pa
rereS
pare
Sp
are
Presented LUNs
Dedicated Spare Disk
Disk Groups & RAID Level
RAID levels in separate small Disk Groups, dispersed LUNs, beware of hot-spots
Virtual Array Controller
HP Virtual Array Approach
Block Mapping
Table
Spare Capacity
Disk Group(s)
Disk groups, segments, block mapping tables & sparing
HP Virtual Array Approach
An EVA can have
• 1 to 16 disk groups• 8 to 240 disks per disk
group
Disk groups
Virtual Array Controller
HP Virtual Array Approach
21Presented LUNs
LUN 1 (RAID1)
LUN 2 (RAID5)
LUN/vdisk allocation
20
HP Virtual Array Approach
An EVA can have
• from 1 to 1024 virtual disks/LUNs
• LUN sizes from 1GB to 2TB in steps of 1GB
• any combination of VRAID 0, 1, 5
LUNs/vdisks and their allocation
Virtual Array Controller
HP Virtual Array Approach
1 2 3
LUN 1 LUN 2LUN 3
Capacity upgrade and load leveling
2
HP Virtual Array ApproachCapacity upgrade, disk group growth
HP Virtual Array Approach
Virtual Array Controller
1 3
LUN 1 (RAID0)
LUN 2 (RAID1)
LUN 3 (RAID5)
2
Block Mapping
Table
Presented LUNs
Spare Capacity
Disk Group(s)
All RAID levels within a Disk Groups, optimal striping, no hot-spots
Virtual Array Controller
HP Virtual Array Approach
1 2
LUN 1 LUN 2
Online Volume Growth
25
HP Virtual Array ApproachOnline Volume Growth
The value of the EVA virtualizationLower management and • training costs• Easy to use intuitive
web-interface• Unifies storage into
a common pool• Effortlessly create
virtual RAID volumes (LUNs)
Buy less• Significantly increase
utilization and reduce stranded capacity
Improved application availability
• Enterprise-class availability• Dynamic pool and
Vdisk (LUN) expansion• No storage reconfiguration
down time
Improve performance
– service more customers• Vraid striping across all
disks in disk group• Eliminate I/O hot spots• Automatic load leveling
Transition SlideHP StorageWorks EVA Software Solutions
EVA software
XCS6.0 / Command View EVA 6.0− Windows based authentication (same as RSM)
• Impacting GUI, SSSU and API• Single sign-on
− Support of new firmware features• Mirrorclone• Snapshot Restore• Enhanced async CA• Non-migrate disk firmware update• Progress indicators
− Usability enhancements• Single Page creation of snapshots, snapclones, mirrorclones,
diskgroups, storage initialization• Delete SnapClone while normalizing• CA links status
XCS6.0 / CV EVA 6.0 GotchasAs of 22.11.06
• AppRM (replacement for FRS)−Not supported for CA volumes
−Only SnapClones supported
• MetroCluster EVA−Support expected December 06
• Data Protector ZDB and Instant Recovery−No container support yet
−No MirroClone support yet
• Storage Essentials−Only supported with SE5.1 SP1 expected December 06
Command View SecurityNew with XCS 6.0• Totally new security model
− No longer relying on System Management Homepage/WBEM− Instead CV EVA now uses Windows based authentication
• Use your Windows account to log into CV EVA
− Consistently used across all interfaces (CV GUI, SSSU, API)− Providing two levels of access
• Admin access: full access to all functionality• User acess: read-only acess
− Introducing user id based auditing• If turned on all actions are logged by user• Written to a file (locally or share)
and/or to the Windows Application Event Log
• Side effect:− Command View EVA will not longer rely on the System Management
homepage− Therefore port has been changed to:
https://localhost:2372/command_view_eva
Non-migrate disk drive firmware update
New with XCS 6.0• Pre-XCS 6 possibilities− Massive disk drive code load to update all drives
at a time• Single image applied like an EVA firmware code load
• EVA will be offline for several minutes
− Single ungrouped disk drive code load• Every drive has to be ungrouped, updated and re-
grouped
• Massive time and effort
• New with XCS 6 (above are still possible)− Ability to code load disk drives while they are
grouped
− Pre-requisit: • No VRaid 0 VDisk must exist on that Disk Group
− Process: • EVA will take disk out of operation,
code load it and then reintroduce it
• Any write to that disk will be buffered and applied once disk drive is back
• Reads will be generated out of RAID information
HP Command View EVAPowerfully Simple Management
• Provides a powerfully, simple management experience for all EVA arrays
• Automate and aggregate management tasks
• Offers proactive remote monitoring services for maximum uptime
• Intuitive, easy-to-use GUI • Enables you to quickly
expand a LUN online, configure LUNs or RAID groups, or add physical disks with just a few mouse clicks
• Uses standards-based SMI-S• Allows you to easily
provision online storage and instantly replicate data
HP Command View EVA
HP Enterprise Virtual Arrays
SMIS and APIs
HP Command View EVA suite
Performance Monitoring
LUN masking
CLUI-scripting/agents
Configuration, discovery, events & monitoring, security
Basic replication
Management Applications
HP IS
EE
solu
tion
s -
rem
ote
m
on
itori
ng
CV EVA deployment options
• Choice and flexibility to maximize your investment
• Broad Microsoft Windows OS coverage
• Host-based or direct host attached device management
CV EVA
Customer application
SAN
General Purpose Server
Or
Up to 16 EVAs
CV EVA
NAS OS
HP ProLiant Storage Server (NAS)
Gigabit Ethernet (iSCSI)Or Fibre channel
Up to 16 EVAs
HostHost CV EVA
OV SOM v1.2
Or
SAN
Management Server-dedicated
CV EVA ≥5.0 required for EVA4000/6000/8000
Existing OV SOMInstalls; includes OV SNM
SAN
EVA familyEVA family
Up to 16 EVAs
HostHostCV EVA
OV SOM v1.2
Or
SAN
HP Storage Management Appliance (discontinued)
SMA SW v1.2
SMA SW v1.2CV EVA ≥5.0 required for EVA4000/6000/8000
Existing OV SOMInstalls; includes OV SNM
EVA family
HP command view EVAperfEVA performance analysis
See the EVAPerf Whitepaper on:http://h18006.www1.hp.com/storage/arraywhitepapers.html
• Performance analysis tool for whole EVA product line
• Shipped with Command View EVA• Integrates with Windows PerfMon • Create your own scripts via a command
prompt• Monitor in real-time and view historical EVA
performance metrics to more quickly identify performance bottlenecks
• Easily monitor and display EVA Performance metrics:
− Host connection data
− port status
− host port statistics
− storage cell data
− physical disk data
− virtual disk data
− CA statistics
EVA Replication SoftwareEnhancements with XCS 6.0xx
• Replication Solution Manager 2.1− Tru64 Host Agent
− Single sign-on
• Business Copy 4.0− MirrorClone feature with Delta Resync and Instant
Restore
− Instant Restore from a Snapshot
• Continuous Access 3.0− Enhanced asynchronous performance and distance
support by using buffer-to-disk (journaling)
Replication Solutions Manager 2.1
Select hosts or storage volumes
Status monitoring
Interactive Topology Manager
Familiar browser based navigation
Auto discovery of storage systems and volumes
Oracle Application Integration
Selectable views
Context sensitive actions and wizards
Local and remote Mgmt
Business copy EVA4 options available:• space efficient vSnapShot• pre-allocated vSnapShot • vSnapClone• Mirror Clone
Controlled from Command View, RSM or SSSU.
Ideally suited to create point-in-time copies to:
• Keep applications online while backing up data
• Test applications against real data before deploying
• Restore a volume after a corruption
• Mine data to improve business processes or customer marketing
point-in-time copy capability for the EVA (local copy)
Space efficient snapshotsVirtually capacity free
time
volumeA
A’snapof A
t0$ create snapshot “A”
t2
contentsidentical
t4t3
volume Areceives more
updates
volume“A” A’
(contents
as of t0)
contentsdifferent
updates t1
volume“A”
A’(contentsas of t0)
contentsdifferent
updates t1
updates t3
t1
volume Areceives
updates (copy on write)
Pre-allocated snapshotsSpace reservation
time
volume“A”
t2 t4t1 t3
volume “A”receives more
updatesvolume “A”
receives updates (copy
on write)
volume“A”
contentsas of t0
contentsdifferent
updates t1
volume“A”
contentsas of t0
contentsdifferent
updates t1
updates t3
snapof A
t0$ create snapshot “A”
contentsidentical
New: Pre-allocated 3-phase snapshotsSpace reservation
volume“A”
t2 t4t1 t3
volume “A”receives more
updatesvolume “A”
receives updates (copy
on write)
volume“A”
contentsas of t0
contentsdifferent
updates t1
volume“A”
contentsas of t0
contentsdifferent
updates t1
updates t3
t0$ create snapshot “A”
snapof A
contentsidentical
t-x
create empty
container
t5
convert snap to an empty container
SnapClone of virtual disksFull copy
time
volumeA
t2
B snapof A
$ create snapclone“A”
contentsidentical
t4t1 t3
Cloning finished
A
volumeA
Vol B content as vol A at t0
relation suspended
A B
cloning process starts
volumeA
contentsas of t0
contentsdifferent
Aupdates t1
volume Areceives
updates (copy on write)
t0
3-phase SnapCloneFull copy
volumeA
t2 t4t1 t3
Cloning finished
A
volumeA
content of A as at t0
relation suspended
A B
cloning process starts
volumeA
content of A as at t0
contentsdifferent
Aupdates t1
B snapof A
$ create snapclone“A”
contentsidentical
t-x
create empty
container
volume Areceives updates (copy on write)
t5
convert B to an empty container
t0
Business Copy 4.0 MirrorClones
New with XCS 6.0• A Mirror Clone is a pre-normalized Clone− Full Clone of the source
• requires 100% of the capacity (if same raid level)
− Synchronous mirror between source VDisks and MirrorClone• Once synchronized data is always identical (unless fractured)
− MirrorClone can be in a different Disk Group/have a different Raid level• Tiered storage approach• Can be used to protect against physical failures
• Point-in-time copy is established at moment fracture is made− differences are tracked via bitmap− Delta resynch/restore is accomplished by only
resynchronizing/restoring data that is marked different• Primary advantages
− Data is available at instant of split− Delta resynch takes less time than a full copy
MirrorClone Tasks• Initial creation
− Will establish MirrorClone relationship and start initial copy
• Fracture (only permitted when fully synchronized)− Will establish a point-in-time copy by stopping replication of writes to MirrorClone
− Deltas are tracked in a bitmap (for both source VDisks and MirrorClone)
− Allows MirrorClone to be presented
• Resync (only permitted when fractured)− Will resync the deltas from the source VDisk to the MirrorClone leading to a
synchronized MirrorClone
• Restore (only permitted when fractured)− Will restore the source VDisk back to the point-in-time the MirrorClone was
fractured
− Instant access to restored data
• Detach (only permitted when fractured)− Will break the MirrorClone relation and convert the MirrorClone into a standalone
VDisk
− If exist, Snapshots from the MirrorClone will stay intact and attached to the former MirrorClone
MirrorClone creation
HostE:
Synchronized MirrorClone• Once MirrorClone is synchronized data on both
volumes is kept identical• Writes are applied to both volumes• Reads are satisfied by source only
Reads
Writes
HostE:
User...• Creates empty container with same size as source
VDisk• Raid level and Disk Group can be differentContainer
Reads
Writes
HostE:
User...• Creates MirrorClone using the Container as targetEVA...• Establishes MirrorClone relationship• Start inital synchronization of MirrorClone (Volume
stays fully accessible to host)
Reads
Writes behind copy fence
Writes
HostE:
Initial situation• VDisk, presented to the host, volume mounted
Reads
Writes
MirrorClone source (Production VDisk)
MirrorClone target
MirrorClone fracture and resynch
HostE:
User...• Fractures MirrorClone EVA ...• Stops applying writes to the MirrorClone target • Instead changes are marked in a delta bitmap
Reads
Writes
MirrorClone source (Production Vdisk)
MirrorClone target
User...• Can present fractured MirrorClone for various
purposes (Read and write).EVA ...• Changes to the source and target are recorded in a
delta bitmap
HostE:
Reads
Writes
HostF:
Reads
Writes
User...• Initiates resynchronization of volumes in either direction EVA ...• Copies change block only until source and target are
synchronized
HostE:
Reads
Writes
HostE:
Synchronized MirrorClone• Once MirrorClone is synchronized data on both volumes is
kept identical• Writes are applied to both volumes• Reads are satisfied by source only
Reads
Writes
Combining Snapshots and MirrorClone
Source MC
t0
t1
t2
• MirrorClone and SnapShot can be combined in a way that you take the Snapshot from the MirrorClone target
• Advantages:− A way to get around the Snapshot
copy-before-write performance impact
− „cross disk group“ Snapshots by putting MirrorClone into different diskgroup
• The Snapshots will allocate space in the MirrorClone disk group
• Disadvantages: − No Snapshot restore in the first release
• Workaround could be to detach MirrorClone, then restore and present as original LUN
• Direct restore is planned for end 2006
Continuous access EVARemote copy capability for the EVA
Continuous Access EVA delivers array-based remote data replication – protecting your data and ensuring continued operation from a disaster.
• Ideal suited to:• Keep byte for byte copies
of data at a remote site for instant recovery
• Replicate data from multiple sites to one for consolidated backup
• Shift operations to a recovery site for primary site upgrades and maintenance
• Ensure compliance to government legislation and business objectives
Continuous Access EVARemote copying• What does it do?
− Replicates LUNs between EVAs
− Provides disaster recovery
− Simplifies workload management
− Allows point-in-time database backup
− Provides restore without latency
• How does it work?− Creates up to 256 Copy Sets for all
specified logical units in the array over Fibre Channel and FC extensions
− Synchronous and asynchronous support up to 20’000km (200ms round trip time)
− Works with all EVAs
Copy Set
Copy Set
DestVOL
DestVOL
SourceVOL
SourceVOL
DR groups and managed sets
Vdisk
DR Group
Managed Set
DR Group• Consistent Group of replicated copy sets (Vdisks)
− Up to 256 DR Groups or DR Group members/array
− Up to 32 replicated copy sets / DR Group
− IO ordering across members is guaranteed
− Share a single write history log
− Vdisks within a DR Group behave like a single entity
− Management commands like suspend or failover are handled atomically
− All source members online on same HSV controller
• Therefore a DR Group is the primary level of CA management
− Write Mode ([Synchronous] / Asynchronous)
− Failsafe Mode (Enabled or [Disabled])
− Suspend Mode ([Resume] / Suspend)
− Failover command
Managed Sets• Another level of CA management
− Collection of DR groups for the purpose of common management
− No consistency as between members of a DR Group
− If you perform a management command on a Managed Set this command will be run for all contained DR Group one after the other
Continuous Access EVA 3.0Enhanced Async implemented with XCS 6.0
• Replaced previous CA async • Tunnel resources are 124 x 8k buffers =
1MB on the fly• Enhanced Async uses a write history log
−You set size and location when a copy set is created
−You can force a full copy
−Log is a circular buffer• Log overflows when tail meets head
• Overflow of log forces a full copy of DR group
• Draining the log will require transition to sync CA
Continuous Access 3.0 Enhanced Async
Time
0 8am 12 noon 5pm 12pm
MB/sec 100th Percentile - Sync CA
95th Percentile - Async CA
50th Percentile - Enhanced Async CA
Multiple relationships
EVA8000
EVA5000
EVA3000
EVA5000
EVA4000
EVA6000
EVA8000
EVA6000
• Fan-in of multiple relationships− The ability of one EVA to act as
the destination for different LUNs from more than one source EVA
• Fan-out of multiple relationships
− The ability for different LUNs on one EVA to replicate to different destination EVA
• Bidirectional− one array with copy sets acting
as the source and destination across the same intersite links or fabric
EVA CA SAN configuration2 fabric configuration
Server1 Server2
Management Server
Management Server
EVA1 EVA2
A B A B
All EVA ports are used for host IO and some also for
CA IO
Shared SAN for host and CA traffic.
EVA CA SAN configurationsPhysically separated 6 fabric configuration
Server1 Server2
Management Server
Management Server
EVA1 EVA2
A B A B
4 ports per EVA used for host IO 4 ports per EVA used
for CA IO
CA traffic is only going through the CA SAN
No host IO cross-site possible > CLX EVA
CA configurations: Dedicated CA fabricsPhysically separated & zoned 4 fabric configuration
Server1 Server2
EVA1 EVA2
A B A B
4 ports per EVA used for Hosts 4 ports per EVA used
for CA
CA traffic is only going through the CA SAN if the EVA ports are properly zoned off in the host SANHost IO cross-site possible -> Stretched Cluster
Management Server
Management Server
CA configuration: Dedicated CA zone
Server1 Server2
Management Server
Management Server
EVA1 EVA2
A B A B
4 ports per EVA used for Hosts 4 ports per EVA used
for CA
CA traffic is only going through the CA zone if the EVA ports are properly zoned off in the host zones
Host IO cross-site possible -> Stretched Cluster
Zoned 2 fabric configuration
Transition SlideEVA Solutions
Zero downtime backupRecovering in minutes not hours
Client network
HP-UX Solaris NTW2k
Data ProtectorServer
SAN
P-Vol S-Vol
• Description− Data Protector provides no impact backup,by
performing backup on the copy of the production data; with the option to copy it or move it to tape.
− NEW with Data Protector 6.0: Incremental ZDB for files
• Usage− Data that requires:
• Non-disruptive protection
• Application-aware
• Zero impact backup
− SAN protection
• Benefits− Fully automates the protection process
− All options can be easily configured using simple selections
− The Data Protector GUI permits complete control of the mirror specification
− Administrators can choose the schedule of the backup to tape
Oracle database integration
• What does it do?− Maps the Oracle DB to
Vdisks, DR Groups etc.
− Replicates all Vdisks of specified Oracle Databases
− Allows creating local or remote replicas
− Easy control via RSM GUI
− Can quiesce and resume Oracle
− Provides a topology MAP
– Supported with BC 2.3 / RSM2.0
– Chargeable RSM Option: Application Integration LTU T4390A
Instant recovery for XP and EVA
• Description− Allows Instant Recovery by retrieving the data
directly from the replicas on disk. − This technology moves Zero Downtime Backup a
step further, allowing to keep multiple replicas on disk available and rotating
• Usage− Critical data that has to be recovered within
minutes, instead of hours• Benefits
− Fully automated protection process, including creation and rotation of replicas
− Disk operations permit non-disruptive, application-aware protection as frequently as once an hour
− Administrators can choose disk protection, tape protection, or scheduled combinations; to meet their protection requirements
• Prerequisits− Data Protector or− AppRM (1H07)
Recovering in minutes not hours
P-Vol
BC1BC2
BC3
t0
t-2t-
1
SAN
Client network
HP-UX Solaris NTW2k
Data ProtectorServer
VLS Gateway
HP VLS 300 EVA Gateway
• Seamless integration– Emulates popular tape drives and
libraries
– Same easy to use GUI as VLS6000
– Allows deployment of existing EVA systems for backup use
• Easily scale capacity and performance
• Utilizes existing infrastructure– Switches
– Arrays
SAN
Node 1Node 2 Node n
EVA 1 EVA 2 EVA n
SAN attached Servers
Application Recover Manager• Application support
− First release: Exchange 2003 ,SQL 2000/2005, NTFS File system
− Future: take over all ZDB/IR integrations DP has (other Apps, other OSs)
• Array support− EVA arrays support including copy back restore
− Disk array independence through VSS/VDS (unmap/swap restore only)
• Features− Round-robin replicas
− Built-in scheduler
− User Management
− Sophisticate logging and monitoring
• Distributed architecture − Central management using ‘DP like’ GUI & CLI
− Clustered Cell Server
− Remote client deployment
Application Recovery Manager (AppRM)• A new solution has been created that encapsulates and delivers
Data Protector’s VSS functionality− Announced in May 06, released in Nov 06
− Replacing “Fast Recovery Solutions for Exchange”
• AppRM is− Disk-based (VSS) replication and restore only, no tape backup possible,
but can be used as pre-exec to 3rd party backup application
• AppRM is based on Data Protector 6.0 code− DP will offer same feature set as AppRM, whereas AppRM offers only a
subset of DP functionality (ZDB and IR)
• Target customers: − NON – DP accounts, with the desire of a VSS instant recovery solution,
but no need for a “full” backup software product
− Potential up-sell opportunity to migrate existing backup product to Data Protector
Application Recover Manager • AppRM follows Data Protector licensing scheme
−Capacity based−More expensive than FRS, especially for only a few larger
systems−but also more functionality−TB licenses are based on the source capacity independent
of the number of copies
• T4395A HP StorageWorks AppRM Cell Manager Win LTU• T4396A HP StorageWorks AppRM Online Backup Win LTU • T4399A HP StorageWorks AppRM Inst. Recovery EVA 1 TB • T4400A HP StorageWorks AppRM Inst. Recovery EVA 10 TB
• First version AppRM 6.0−Aligned with DP version numbering
MetroCluster EVA for HP-UX
DataCenter 1 DataCenter 2
HP ContinuousAccess EVA
ServiceGuardfor HP-UX
Metro Cluster EVA
Up toseveral 100km
• What does it do?– Provides manual or automated site-failover
for Server and Storage resources
• Supported environments:– HP-UX 11i V1 & 11i V2
– Serviceguard ≥11.15
• Requirements:– EVA Disk Arrays
– Metrocluster
– Continuous Access EVA
– Max 200ms network round-trip delay
– Command View EVA & SMI-S
Cluster extension EVA for Windows
MSCS on Windows
DataCenter 1 DataCenter 2
HP ContinuousAccess EVA
Cluster Extension XP
Up to 500km
• What does it do?– Provides manual or automated site-failover for
Server and Storage resources
• Supported environments:– Microsoft Windows 2003 Enterprise Edition (32-
bit & 64-bit)
– Microsoft Windows 2003 Data Center Server(64-bit)
– NAS4000 & 9000
– HP Proliant Storage Server
– Microsoft Cluster Service 5.2
– Up to 500km
• Requirements:– EVA Disk Arrays
– Cluster Extension EVA
– Continuous Access EVA
– Max 20ms network round-trip delay
– Command View EVA & SMI-S
Cluster extension EVA for Linux
Serviceguard/Li
DataCenter 1 DataCenter 2
HP ContinuousAccess EVA
Cluster Extension EVA
Up to 500km
• What does it do?– Provides manual or automated site-failover
for Server and Storage resources
• Supported environments:– Serviceguard for Linux as the cluster service
– SG 11.16.02 with RH EL 4
– SG 11.16.01 for SuSe SLES 9
• Requirements:– EVA Disk Arrays
– Cluster Extension EVA
– Continuous Access EVA
– Max 20ms network round-trip delay
– Command View EVA & SMI-S
Windows 2003 stretched cluster with CA
Continuous Access EVA
App A
App BQuorum Quoru
m
Quorum
App A
App B
• CA Failover• Restart Servers(Rescan)DRG A
DRG B
Cluster extension EVA CLX– Manual Move of App A
DRG A
DRG BContinuous Access
EVA
App A
App BQuorum
Quorum
App A
• Move App A
Quorum or Witness
Server
Cluster extension EVA – Storage Failure
DRG A
DRG BContinuous Access
EVA
App A
App BQuorum Quoru
m
Quorum
App A
App B
April 8, 202395
Majority Node Set Quorum– File Share Witness• What is it?
− A patch for Windows 2003 SP1 clusters provided by Microsoft (KB921181)
• What does it do?− Allows the use of a simple file share to provide a vote for an MNS quorum-based 2-
node cluster• In addition to introducing the file share witness concept, this patch also introduces a
configurable cluster heartbeat (for details see MS Knowledge Brief)
• What are the benefits?− The „arbitrator“ node is no longer a full cluster member.
• A simple fileshare can be used to provide this vote.• No single subnet requirement for network connection to the „arbitrator“.
− One arbitrator can serve multiple clusters. However, you have to set up a separated share for each cluster.
− The „abitrator“ exposing the share can be • a standalone server• a different OS architecture (e.g. a 32-bit Windows server providing a vote for a IA64 cluster)
Majority Node Set Quorum– File Share Witness
# cluster nodes
# node failures
2 0
3 1
4 1
5 2
6 2
7 3
8 3
App A
App B
App A
App B
\\arbitrator\share
Get vote
0 (1 with MNS fileshare witness)
App A
HP SAN certification and support
• HP StorageWorks SAN Design Guide– Architecture guidance
– Massive configuration support
– Implementation best practices
– Incorporation of new technologies
– Include now IP Storage implementation like iSCSI, NAS/SAN Fusion, FC-IP
• Provides the benefit of HP engineering when building a scalable, highly available enterprise storage network
• Documents HP Services SAN integration, planning and support services
http://www.hp.com/go/sandesignHP SAN architecture
Rules
The EVA global service portfolioHP StorageWorks EVA base product warranty
Foundation Service Solution• 2 years parts, 2 years of labor and 2 years of
hardware onsite 24x7, 4-hour response for EVA controller pair and drive shelves (enclosures) as defined by product SKU and the Hard Disk Drives purchased with the array
• 2 years, 24x7, 2-hour response phone-in and updates for Virtual Controller Software (VCS)
• Array Installation and Startup (includes drive shelves and hard disks purchased with the EVA)
HP care pack services for storage• HP H/W Support
−4-hour 24 x 7
−Years 3, 4, and 5
• HP S/W Support−24 x 7 technical support
−Software product updates
• Premium Hardware & Software Services−Support Plus 24
−Proactive 24
−HP Critical Service
Why should you choose the new EVA?Inherited from the current EVA• Easiest management and setup • Virtualization allows better use of resources and automatic
striping to prevent hot spots• Dynamic LUN expansion• Full set of local and remote copy options• Known solid HP support
Added with the new EVA• Easier implementation and coexistence due to support of
industry standard multipathing and native HBAs• Higher performance • Higher capacities – The EVA8000 supports architecturally >
200TB
Transition SlideHp StorageWorks™ – the Right Choice
Hp logo black on white