Modes of Remote Replication
Modes of Remote Replication
Modes of Remote Replication
Chapter 12
Remote Replication
Modes of Remote Replication
The two basic modes of remote replication are synchronous and asynchro-
nous. In synchronous remote replication, writes must be committed to the
source and remote replica (or target), prior to acknowledging “write
complete” to the host (see Figure 12-1).
Additional writes on the source cannot occur until each preceding write has
been completed and acknowledged. This ensures that data is identical on the
source and replica at all times. Further, writes are transmitted to the remote
site exactly in the order in which they are received at the source.
The degree of impact on response time depends primarily on the distance between
sites, band- width, and quality of service (QOS) of the network connectivity
infrastructure.
In asynchronous replication, data at the remote site will be behind the source by at
least the size of the buffer.
If the same location is written multiple times in the buffer prior to transmission
to the remote site, only the final version of the data is transmitted. This feature
conserves link bandwidth.
The LVM on the remote host receives the writes and commits them to the remote
volume group.
Prior to the start of replication, identical volume groups, logical volumes, and file
systems are created at the source and target sites. Initial synchro- nization of data
between the source and replica is performed.
After the initial synchronization, production work can be started on the source
volumes and replication of data can be performed over an existing standard IP
network (see Figure 12-5).
Chapter 12 ■ Remote Replication 5
IP
The remote host receives the logs and applies them to the remote database.
6 Section III ■ Backup, Archive, and Replication
Log Log
IP
Data Data
All DBMSs switch log files at preconfigured time intervals or when a log file is full.
The current log file is closed at the time of log switching, and a new logfile is opened.
When a log switch occurs, the closed log file is transmitted by the source host to the
remote host. The remote host receives the log and updates the standby database.
This process ensures that the standby database is consistent up to the last committed
log. RPO at the remote site is finite and depends on the size of the log and the
frequency of log switching.
Available network bandwidth, latency, rate of updates to the source database, and the
frequency of log switching should be considered when determining the optimal size of
the log file.
Similar to LVM-based remote replication, the existing standard IP network can be used for
replicating log files. Host-based log shipping requires low network bandwidth because
it transmits only the log files at regular intervals.
Chapter 12 ■ Remote Replication 7
This relieves the burden on the host CPUs, which can be better used for applications
running on the host.
A source and its replica device reside on different storage arrays. Data can be
transmitted from the source storage array to the target storage array over a shared or a
dedicated network.
Additional writes on that source cannot occur until each preceding write has been
completed and acknowledged.
In the case of synchronous remote replication, to optimize the replication process and
to minimize the impact on application response time, the write is placed on cache of
the two arrays.
The intelligent storage arrays destage thesewrites to the appropriate disks later.
If the network links fail, replication is suspended; however, production work can
continue uninterrupted on the source storage array.
The array operating environment keeps track of the writes that are not transmitted to
the remote storage array.
When the network links are restored, the accumulated data is transmitted to the
remote storage array. During the time of network link out- age, if there is a failure at
the source site, some data will be lost, and the RPO at the target will not be zero.
8 Section III ■ Backup, Archive, and Replication
1 2
4 3
So u rce
1 Write from the production host is received by the source storage array.
3 Acknowledgment is sent to the source storage array by the remote storage array.
1 3
2 4
So u rce
4 After the writes are received by the target array, it sends an acknowledgment to the source array.
Figure 12-9 shows the sequence of operations in a disk-buffered remote replica- tion. At
the beginning of the cycle, the network links between the two arrays are suspended, and
there is no transmission of data.
While production applicationruns on the source device, a consistent PIT local replica
of the source device is created. The network links are enabled, and data on the local
replica in the source array transmits to its remote replica in the target array.
After synchroni- zation of this pair, the network link is suspended, and the next local
replica of the source is created. Optionally, a local PIT replica of the remote device on
thetarget array can be created.
The frequency of this cycle of operations depends on the available link bandwidth and
the data change rate on the source device. Because disk-buffered technology uses local
replication, changes made to the source and its replica are possible to track.
Therefore, all the resynchronization operations between the source and target can be
done incrementally.
In disk-buffered remote replication, the RPO at the remote site is in the order of
hours.
For example, a local replica of the source device is created at 10:00 a.m., and this
data transmits to the remote replica, which takes 1 hour to complete.
Changes made to the source device after 10:00 a.m. are tracked. Another local replica
of the source device is created at 11:00 a.m. by applying track changes between the
source and local replica (10:00 a.m. copy)
10 Section III ■ Backup, Archive, and Replication
During the next cycle of transmission (11:00 a.m. data), the source data has
moved to 12:00 p.m.
The local replica in the remote array has the 10:00 a.m. data until the 11:00
a.m. data is successfully transmitted to the remote replica.
Source Device
Local Replica
2 4
3
Production
Host
Remote Replica
Source Storage Array Target Storage Array
Local Replica
3 Data from the local replica in the source array is transmitted to its remote replica in the target array.
4 Optionally, a local PIT replica of the remote device on the target array is created.
Figure 12-10 describes CDP remote replication. In this method, the replica is
synchronized with the source, and then the replication process starts.
After the replication starts, all the writes from the host to the source are split
into two copies.
One of the copies is sent to the local CDP appliance at the source site, and the
other copy is sent to the production volume.
After receiving the write, the appliance at the source site sends it to the
appliance at the remote site.
Then, the write is applied to the journal volume at the remote site. For an
asynchronous operation, writes at the source CDP appliance are accumulated,
and redundantblocks are eliminated.
Then, the writes are sequenced and stored with their corresponding
timestamp. The data is then compressed, and a checksum is generated.
After the data is received, the remote appliance verifies the checksum to ensure
the integrity of the data.
The data is then uncompressed and written to the remote journal volume.
As a next step, data from the journal volume is sent to the replica at
predefined intervals.
12 Section III ■ Backup, Archive, and Replication
Host
Write Splitter
Local Remote
CDP Appliance CDP Appliance
SAN SAN/WAN SAN
Production
Volume
Source
Storage Array
CDP Replica
Journal
Remote
Storage Array