Unity Replication
Unity Replication
Copyright ©2016 EMC Corporation. All Rights Reserved. Published in the USA. EMC believes the information in this publication is accurate as o f its publication date. The information is subject
to change without notice.
THE INFORMATION IN THIS PUBLICATION IS PROVIDED “AS IS.” EMC CORPORATION MAKES NO REPRESENTATIONS OR WARRANTIES OF ANY KIND WITH RESPECT TO THE INFORMATION IN
THIS PUBLICATION, AND SPECIFICALLY DISCLAIMS IMPLIED WARRANTIES OF MERCHANTABILITY OR FITNESS FOR A PARTICULAR PURPOSE.
Use, copying, and distribution of any EMC software described in this publication requires an applicable software license. The trademarks, logos, and service marks (collectively "Trademarks")
appearing in this publication are the property of EMC Corporation and other parties. Nothing contained in this publication sh ould be construed as granting any license or right to use any
Trademark without the prior written permission of the party that owns the Trademark.
EMC, EMC² AccessAnywhere Access Logix, AdvantEdge, AlphaStor, AppSync ApplicationXtender, ArchiveXtender, Atmos, Authentica, Authentic Problems, Automated Resource Manager,
AutoStart, AutoSwap, AVALONidm, Avamar, Bus-Tech, Captiva, Catalog Solution, C-Clip, Celerra, Celerra Replicator, Centera, CenterStage, CentraStar, EMC CertTracker. CIO Connect,
ClaimPack, ClaimsEditor, Claralert ,cLARiiON, ClientPak, CloudArray, Codebook Correlation Technology, Common Information Model, Compuset, Compute Anywhere, Configuration
Intelligence, Configuresoft, Connectrix, Constellation Computing, EMC ControlCenter, CopyCross, CopyPoint, CX, DataBridge , Data Protection Suite. Data Protection Advisor, DBClassify, DD
Boost, Dantz, DatabaseXtender, Data Domain, Direct Matrix Architecture, DiskXtender, DiskXtender 2000, DLS ECO, Document Sciences, Documentum, DR Anywhere, ECS, elnput, E-Lab,
Elastic Cloud Storage, EmailXaminer, EmailXtender , EMC Centera, EMC ControlCenter, EMC LifeLine, EMCTV, Enginuity, EPFM. eRoom, Event Explorer, FAST, FarPoint, FirstPass, FLARE,
FormWare, Geosynchrony, Global File Virtualization, Graphic Visualization, Greenplum, HighRoad, HomeBase, Illuminator , InfoArchive, InfoMover, Infoscape, Infra, InputAccel, InputAccel
Express, Invista, Ionix, ISIS,Kazeon, EMC LifeLine, Mainframe Appliance for Storage, Mainframe Data Library, Max Retriever, MCx, MediaStor , Metro, MetroPoint, MirrorView, Multi-Band
Deduplication,Navisphere, Netstorage, NetWorker, nLayers, EMC OnCourse, OnAlert, OpenScale, Petrocloud, PixTools, Powerlink, PowerPath, PowerSnap, ProSphere, ProtectEverywhere,
ProtectPoint, EMC Proven, EMC Proven Professional, QuickScan, RAPIDPath, EMC RecoverPoint, Rainfinity, RepliCare, RepliStor, ResourcePak, Retrospect, RSA, the RSA logo, SafeLine, SAN
Advisor, SAN Copy, SAN Manager, ScaleIO Smarts, EMC Snap, SnapImage, SnapSure, SnapView, SourceOne, SRDF, EMC Storage Administrator, StorageScope, SupportMate, SymmAPI,
SymmEnabler, Symmetrix, Symmetrix DMX, Symmetrix VMAX, TimeFinder, TwinStrata, UltraFlex, UltraPoint, UltraScale, Unisphere, Universal Data Consistency, Vblock, Velocity, Viewlets, ViPR,
Virtual Matrix, Virtual Matrix Architecture, Virtual Provisioning, Virtualize Everything, Compromise Nothing, Virtuent, VMAX, VMAXe, VNX, VNXe, Voyence, VPLEX, VSAM-Assist, VSAM I/O
PLUS, VSET, VSPEX, Watch4net, WebXtender, xPression, xPresso, Xtrem, XtremCache, XtremSF, XtremSW, XtremIO, YottaYotta, Zero-Friction Enterprise Storage.
Remote Replication is one method that enables data centers to avoid disruptions in
operations. In a disaster recovery scenario, if the source site becomes unavailable, the
replicated data will still be available for access from the remote site. Remote Replication
uses a Recovery Point Objective (RPO) which is an amount of data, measured in units of
time to perform automatic data synchronization between the source and remote systems.
The RPO for asynchronous replication is configurable. The RPO for synchronous replication
is set to zero. The RPO value represents the acceptable amount of data that may be lost in
a disaster situation. The remote data will be consistent to the configured RPO value. The
minimum and maximum RPO values are 5 minutes and 1440 minutes (24 hours).
Remote Replication is also beneficial for keeping data available during planned downtime
scenarios. If a production site has to be brought down for maintenance or testing the
replica data can be made available for access from the remote site. In a planned downtime
situation, the remote data is synchronized to the source before being made available and
there is no data loss.
The main focus of this training is with remote replication since it has more elements to
configure, create and manage.
Asynchronous Replication architecture utilizes Unified Snapshots. The system creates two
snapshots for the source storage resource and two corresponding snapshots on the
destination storage resource. These system created snapshots cannot be modified. Based
on the replication RPO value the source snapshots are updated in an alternating fashion to
capture the source data state differences, known as deltas. The data delta for the RPO
timeframe is replicated to the destination. After the data is replicated the corresponding
destination snapshot is updated. The two corresponding snapshots capture a common data
state, known as a common base. The common base can be used to restart a stopped or
interrupted replication session.
The first step of the initial process for asynchronous replication is to create a storage
resource of the exact same capacity on the destination system. The storage resource is
created automatically by the system and contains no data.
In the next step, corresponding snapshot pairs are created automatically on the source and
destination systems. They capture point-in-time data states of their storage resource.
The first snapshot on the source system is used to perform an initial copy of its point-in-
time data state to the destination storage resource. This initial copy can take a significant
amount of time if the source storage resource contains a large amount of existing data.
Once the initial copy is complete, the first snapshot on the destination system is updated.
The data states captured on the first snapshots are now identical and form a common base.
Because the source storage resource is constantly changing, its data state is no longer
consistent with the first snapshot point-in-time. In the synchronization process, the second
snapshot on the source system is updated, capturing the current data state of the source.
A data difference, or delta is calculated from the two source system snapshots and a
differential copy is made from the second snapshot to the destination storage resource.
After the differential copy is complete, the second snapshot on the destination system is
updated to form a common base with its corresponding source system snapshot.
The cycle of differential copies continues for the session by alternating between the first
Copyright 2016 EMC Corporation. All rights reserved. Unity Replication ‹#›
The architecture for Unity Synchronous Replication is shown here. The graphic illustrates a
remote synchronous replication session for a LUN. The architecture is the same for
replicating any other block-based storage resource synchronously.
The same fundamental remote replication connectivity and communication between the
source and destination systems seen earlier for asynchronous remote replication are also
required for synchronous replication. A data connection to carry the replicated data is
required and is formed using fibre channel connections between the replicating systems. A
communication channel is also required to manage the replication session. For synchronous
replication, part of the management is provided using Replication Interfaces that are IP
based interfaces for SPA and SPB using specific Sync Replication Management Ports. The
management communication between the replicating systems is established on a
Replication Connection. It defines the management interfaces and credentials for the source
and destination systems.
Synchronous Replication architecture utilizes Write Intent Logs (WIL) on each of the
systems involved in the replication. These are internal LUNs created automatically by each
system. There is a WIL for SPA and one for SPB on each system. They hold fracture logs
that are designed to track changes to the source LUN should the destination LUN become
unreachable. When the destination becomes reachable again it will automatically recover
synchronization to the source using the fracture log, thus avoiding the need for a full
synchronization.
The first step of the initial process for synchronous replication is to create a storage
resource of the exact same capacity on the destination system. The storage resource is
created automatically by the system and contains no data.
In the next step, SPA and SPB Write Intent Logs are automatically created on the source
and destination systems.
An initial synchronization of the source data is then performed. It copies all of the existing
data from the source to the destination. The source resource is available to production
during the initial synchronization but the destination will be unusable until the
synchronization completes.
Should the destination become unreachable, the replication session will be out of
synchronization. The source Write Intent Log for the SP owning the resource will track the
changes. When the destination becomes available the systems will automatically recover
synchronization using the WIL.
An Active session state indicates normal operations and the source and destination are In
Sync.
A Paused session state indicates the replication has been stopped and will have a Sync
State of Consistent indicating the WIL will be used to perform synchronization of the
destination.
A Failed Over session will have one of two Sync States. It can show Inconsistent meaning
the Sync State was not In Sync or Consistent prior to the Failover. If the Sync State was In
Sync prior to Failover, it will be Out of Sync after session Failover.
A Lost Sync Communications session state indicates the destination is unreachable. It can
have any of the following Sync States: Out of Sync, Consistent or Inconsistent.
A Sync State of Syncing indicates a transition from Out of Sync, Consistent or Inconsistent
due to the session changing to an Active state from one of its other states; for example if
the system has been recovered from the Lost Sync Communications state.
For Block, the replication is created on a LUN or a group of LUNs in the case of a
Consistency Group. For File, the replication is configured on a NAS server and file systems.
For VMware the storage resource is either going to be a LUN-based VMFS datastore or a file
system-based NFS datastore. When creating each of these storage resources, the Unity
system provides a wizard for their creation. Each wizard provides an option to automatically
create the replication on the resource. Each resource replication creation is nearly identical
to the other resources.
For resources already created, replications can be created manually from their Properties
page. As with the wizard, the replication creation from the resource Properties page is
nearly identical to the other resources. The following few slides will show replication
creation within the Block storage LUN creation wizard and the File storage file system
creation wizard. It will also show replications created manually from the LUN and file
system Properties pages.
Video demonstrations will also be provided for the resource replication creation.
The CLI console command can be used to verify the Fibre Channel port that the system has
specified as the Synchronous FC Ports on the SPs. The slide shows an example of running
the UEMCLI command “/remote/sys show –detail” command. In the abbreviated
example output, Fibre Channel Port 4 is specified by the system as the Synchronous FC port
for SPA and SPB.
Once the Synchronous FC Ports on the source and destination systems have been verified,
Fibre Channel connectivity can be established between the corresponding ports on the SPs
of each system. Direct connect or zoned fabric connectivity is supported.
Although the Synchronous FC ports will also support host connectivity, it is recommended
that they be dedicated to Synchronous Replication.
The system’s user interface provides a Replications management page that includes a
section for the Interfaces. With the Interfaces section selected, the + icon is used to create
new Replication Interfaces. On the creation screen, an available Ethernet Port from the
system must be selected. An IP address and subnet mask must be provided for both SPs.
Gateway addressing is optional and a VLAN ID configuration is also provided if needed.
As noted before; there is a dependency between a NAS Server and a file system. The NAS
server must be replicated prior to any associated file system.
In the example NAS Server replication shown, the NAS Server has an associated file system
and a separate replication session will be created for it. The table details the destination
resources that will be used for the file system. It can be selected and edited to further
define its destination resources.
In the example, sessions for the NAS Server and its associated file system will be created.
When it is complete the created sessions can be viewed from the Replications page by
selecting the Sessions section.
The Replication Interfaces need to be configured on both the source and destination
systems.
In the example, the session settings for the replication and destination are displayed.
When it is complete the created sessions can be viewed from the Replications page by
selecting the Sessions section.
Similar management of a session can also be done from the Properties page of the
replicated object. The Replication tab displays information about the session, provides
certain editable fields and buttons to delete or perform various replication operations.
The example is from an asynchronous file system replication that is in a normal state. From
the source it is possible to perform session Pause, Sync or Failover with Sync operations.
From the destination it is only possible to perform a session Failover operation.
The example illustrates the order of failover for a NAS Server and an associated file system.
Failover must be done first to the NAS Server, then to its associated file system. The same
is true for the Resume operation after Failover. The Resume operation is initiated first on
the NAS Server then the associated file system. Failback is done in the same order; first to
the NAS Server then to the associated file system.
The example illustrates the process of the operation. It starts with issuing the Failover with
Sync operation from Site A which is the primary production site. The operation will remove
access to the replicated object on Site A. A synchronization from the Site A object to the
Site B object happens next and when it completes the replication session is paused. The
operation then makes the Site B object available for access.
The example illustrates the process of the operation. The primary production site becomes
unavailable and all its operations cease. Data is not available and replication between the
sites can no longer proceed. A Failover operation is issued from Site B which is the
secondary production site. The operation Pauses the existing replication session so that the
session will not start again should Site A become available. The operation then makes the
Site B object available for access and production can resume.
The example illustrates the process of a Resume operation for a session that is failed over.
The Site A replicated object must be available before the replication session can be
resumed. The Resume operation is issued from Site B. The operation will restart the
Paused session in the reverse direction. This will update the Site A object with any changes
that may have been made to the Site B object during the failover. This results in the
session resuming in the reverse direction and returned to a normal state. One may use this
method of restarting replication rather than a Failback operation if production had been
serviced from the Site B object for a significant amount of time and thus has accumulated a
significant amount of change from the Site A object. Replication in the reverse direction will
synchronize the Site A object to the data state of the Site B object. To return production to
the Site A object would require a session Failover operation followed by another Resume
operation.
The example illustrates the process of a Failback operation. The Site A replicated object
must be available before the Failback operation can be initiated on a session. The Failback
operation is issued from Site B. The operation will remove access to the Site B object and
synchronize the Site A object to the data state of the Site B object. The operation then
allows access to the Site A object for production. Replication is restarted using the Site A
object as a source and the Site B object as a destination. This single operation returns the
object’s replication state as it was prior to the failover. One would use this operation to fail
back from failovers lasting for only short time periods. It is important to note that if the Site
B object had accumulated a significant amount of change due to long periods of failover,
the resync time can take a significant amount of time.