Jump to content

RecoverPoint

From Wikipedia, the free encyclopedia

This is an old revision of this page, as edited by W Nowicki (talk | contribs) at 22:13, 31 October 2016 (reduce some product details: not HOWTO). The present address (URL) is a permanent link to this revision, which may differ significantly from the current revision.

RecoverPoint is a continuous data protection product offered by EMC Corporation which supports asynchronous and synchronous data replication of block-based storage. RecoverPoint was originally created by Kashya, which was bought by EMC in 2006.

Capabilities

  • Block-based journaling.
  • Synchronous and asynchronous replication.
  • Any-Point-In-Time - Every write is tracked and stored as a different snapshot. Alternatively, groups of writes can be aggregated according to configuration in order to reduce storage space and network traffic.
  • Heterogeneous (multi-vendor) storage arrays via Fibre Channel.
  • WAN-based compression.
  • Tracking multiple volumes as a single consistency group]

Replication

RecoverPoint continuous data protection (CDP) tracks changes to data at a block level and journals these changes.[1] The journal then allows rolling data to a previous "Point-In-Time" in order to view the drive contents as they were before a certain data corruption. CDP can journal each write individually, hence enabling any-point-in-time snapshots, or it can be configured to combine consecutive writes in order to reduce journal space and improve bandwidth. CDP works only over a storage area network - the RecoverPoint appliances needs to be configured for the replica and the journal Logical Unit Numbers (LUNs).

RecoverPoint continuous remote replication (CRR) enables a replica in a remote site. For such a setup, RecoverPoint appliances clusters are required in both the local and remote sites. These 2 clusters communicate over either FC or IP. RecoverPoint applies compression and de-duplication in order to reduce wide area network traffic. As of RecoverPoint 3.4, only one remote site. CRR can be combined with CDP in order to provide concurrent local and remote (CLR) replication.

The consistency group (CG) term is used for grouping several LUNs together in order to ensure write-order consistency over several volumes. This is used for example with a database that stores its data and journal on different logical drives. These logical drives must be kept in-sync on the replica if data-consistency needs to be preserved. Other examples are multi-volume file systems such as ZFS or Windows' Dynamic Disks. RecoverPoint 3.4 supports up to 128 CGs and 2048 LUNs.[2]

Write splitting

Similar to other continuous data protection products, and unlike backup products, RecoverPoint needs to obtain a copy of every write in order to track data changes. EMC advertises RecoverPoint as heterogenous due to its support of multi-vendor server, network and storage environments.[3]

Host-based write splitting is done using a device driver that is installed on the server accessing the storage volumes. The usage of a host-based splitter allows replication of selected non-EMC storage. Available fabric-based splitters are for Brocade Communications Systems SAN switches and for Cisco Systems SANTap. This requires the investment in additional switch blades. This configuration allows splitting from all operating systems regardless of their version, and is agnostic to the storage array vendor.

Storage array splitters are only supported on a subset of EMC storage products. This method allows write splitting from all operating systems, and does not require special SAN switching hardware. The RecoverPoint/SE is a slimmed-down version that only supports this type of splitter.

Architecture

Each site requires installation of a cluster that is composed of 2-8 RecoverPoint appliances. The multiple appliances work together as an high availability cluster. Each appliance is connected via Fibre Channel (FC) to the SAN, and must access both the server (SCSI initiator) and the storage (SCSI target). Each appliance must also be connected to an IP network for management. Replication takes place over either FC or standard IP. One or more host-, fabric- or array- splitters would split traffic to both the storage and the appliances.

Integration with other products

Besides integration with EMC products such as AppSync, ViPR, Replication Manager, Control Center and Unisphere, and the CLARiiON,VNX, Symmetrix and VPLEX storage arrays, RecoverPoint integrates with the following products:

Integration with VMware vSphere, VMware Site Recovery Manager and Microsoft Hyper-V allows protection to be specified per VM instead of per volumes that are available to the hypervisor.

Integration with Microsoft Shadow Copy, Exchange and SQL Server and Oracle Database Server allows RecoverPoint to temporarily stop writes by the host in order to take consistent application-specific snapshots.

The usage of APIs/CLIs allows customers to integrate RecoverPoint with custom internal software.[1]

Notes

References

  • EMC RecoverPoint [1]