RecoverPoint: Difference between revisions
repaired dab links, wikified |
Remove original research |
||
Line 1: | Line 1: | ||
'''RecoverPoint''' is a [[continuous data protection]] solution offered by [[EMC Corporation]] which supports [[Asynchronous I/O|asynchronous]] and [[Synchronization|synchronous]] data [[Replication (computer science)|replication]] of block-based storage. |
|||
{{ad|date=June 2011}} |
|||
'''RecoverPoint''' is a [[continuous data protection]] solution offered by [[EMC Corporation]]. Capabilities include bi-directional[[Asynchronous I/O|asynchronous]] and [[Synchronization|synchronous]] data [[Replication (computer science)|replication]] across heterogeneous [[Disk array|storage arrays]]. Block-based storage is supported as of today, both [[Fibre Channel]] and [[iSCSI]]. All replication takes place over standard [[Internet Protocol|IP]] for asynchronous replication and Fibre Channel for synchronous replication. RecoverPoint handles Fibre Channel to IP conversions for asynchronous replication. |
|||
== Capabilities == |
|||
The core behind this product's ingenuity{{weasel word|date=June 2011}} is its block-based [[Journaling file system|journaling]] and WAN-based [[Data compression|compression]] capabilities. Every [[Input/output|write]] (or groups of writes) are aggregated together into a single snapshot for distribution. Snapshots are important in comparison to [[Mirror (computing)|mirroring]] technologies due to the fact [[Garbage In, Garbage Out]]. If logical filesystem [[Data corruption|corruption]] manifests itself on the source[[Volume (computing)|volume]], it will be replicated to the destination volume. This is not the case with snapshot-based technology, as each snapshot is a point in time of the production volume. Delineation between one snapshot and another is hard defined{{clarify|date=June 2011}}. Therefore with any snapshot there will be a "before and after" view of the production volume. The before being prior to logical corruption, the after being post corruption. Mirroring is a moving target, its only a matter of time before the logical corruption becomes viral and infects the remote copy. Therefore journal-based replication is the future and the "now" of local and site-to-site replication.{{or|date=June 2011}} |
|||
* Block-based [[Journaling file system|journaling]]. |
|||
* Any-Point-In-Time - Every [[Input/output|write]] is tracked and stored as a different [[Snapshot (computer storage)|snapshot]]. Alternatively, groups of writes can be aggregated according to configuration in order to reduce storage space and network traffic. |
|||
* Heterogeneous (multi-vendor) [[Disk array|storage arrays]] via [[Fibre Channel]]. |
|||
* WAN-based [[Data compression|compression]]. |
|||
* Tracking multiple [[Volume (computing)|volume]] as a single [[consistency group]]. |
|||
== Replication == |
== Replication == |
||
RecoverPoint continuous data protection (CDP) tracks changes to data at a block level and journals these changes <ref>http://www.emc.com/collateral/software/white-papers/h4175-recoverpoint-clr-operational-dr-wp.pdf</ref>. The journal then allows rolling data to a previous "Point-In-Time" in order to view the drive contents as they were before a certain data corruption. CDP can journal each write individually, hence enabling "Any-Point-In-Time" snapshots, or it can be configured to combine consecutive writes in order to reduce journal space and improve bandwidth. CDP works only over [[Storage area network|SAN]] - the RecoverPoint [[Computer appliance|appliances]] needs to be [[DNS zone|zoned]] and masked with both the master, the replica and the journal [[Logical Unit Number|LUNs]]. |
RecoverPoint continuous data protection (CDP) tracks changes to data at a block level and journals these changes <ref>http://www.emc.com/collateral/software/white-papers/h4175-recoverpoint-clr-operational-dr-wp.pdf</ref>. The journal then allows rolling data to a previous "Point-In-Time" in order to view the drive contents as they were before a certain data corruption. CDP can journal each write individually, hence enabling "Any-Point-In-Time" snapshots, or it can be configured to combine consecutive writes in order to reduce journal space and improve bandwidth. CDP works only over [[Storage area network|SAN]] - the RecoverPoint [[Computer appliance|appliances]] needs to be [[DNS zone|zoned]] and masked with both the master, the [[Mirror (computing)|replica]] and the journal [[Logical Unit Number|LUNs]]. |
||
RecoverPoint |
RecoverPoint continuous remote replication (CRR) enables a replica in a remote site. For such a setup, RecoverPoint appliances clusters are required in both the local and remote sites. These 2 clusters communicate over IP. RecoverPoint applies compression and de-duplication in order to reduce WAN traffic. As of RecoverPoint 3.4, only one remote site. CRR can be combined with CDP in order to provide [[concurrent local and remote]] (CLR) replication. |
||
The consistency group (CG) term is used for grouping several LUNs together in order to ensures write-order consistency over several volumes. This is used for example with a database that stores its data and journal on different logical drives. These logical drives must be kept in-sync on the replica if data-consistency needs to be preserved. Other examples are multi-volume file systems such as [[ZFS]] or Windows' [[Logical Disk Manager|Dynamic Disks]]. RecoverPoint 3.4 supports up to 128 CGs and 2048 LUNs <ref>https://globalsp.ts.fujitsu.com/dmsp/docs/ss_recoverpoint.pdf</ref>. Each LUN can contain up to 2 TB, and the total supported capacity can be up to 150 TB. |
The consistency group (CG) term is used for grouping several LUNs together in order to ensures write-order consistency over several volumes. This is used for example with a database that stores its data and journal on different logical drives. These logical drives must be kept in-sync on the replica if data-consistency needs to be preserved. Other examples are multi-volume file systems such as [[ZFS]] or Windows' [[Logical Disk Manager|Dynamic Disks]]. RecoverPoint 3.4 supports up to 128 CGs and 2048 LUNs <ref>https://globalsp.ts.fujitsu.com/dmsp/docs/ss_recoverpoint.pdf</ref>. Each LUN can contain up to 2 TB, and the total supported capacity can be up to 150 TB. |
||
Line 24: | Line 29: | ||
Each site requires installation of a cluster that is composed of 2-8 RecoverPoint appliances. The multi appliances backup one another in a [[high availability]] [[Computer cluster|cluster]]. Each appliance is connected via Fibre Channel to the SAN, and must be zoned together with both the server (SCSI initiator) and the storage (SCSI target). Each appliance must also be connected to an IP network for management. |
Each site requires installation of a cluster that is composed of 2-8 RecoverPoint appliances. The multi appliances backup one another in a [[high availability]] [[Computer cluster|cluster]]. Each appliance is connected via Fibre Channel to the SAN, and must be zoned together with both the server (SCSI initiator) and the storage (SCSI target). Each appliance must also be connected to an IP network for management. |
||
All replication takes place over standard [[Internet Protocol|IP]] for asynchronous replication and Fibre Channel for synchronous replication. |
|||
One or more host-, fabric- or array- splitters would split traffic to both the storage and the appliances. |
One or more host-, fabric- or array- splitters would split traffic to both the storage and the appliances. |
Revision as of 10:27, 17 June 2011
RecoverPoint is a continuous data protection solution offered by EMC Corporation which supports asynchronous and synchronous data replication of block-based storage.
Capabilities
- Block-based journaling.
- Any-Point-In-Time - Every write is tracked and stored as a different snapshot. Alternatively, groups of writes can be aggregated according to configuration in order to reduce storage space and network traffic.
- Heterogeneous (multi-vendor) storage arrays via Fibre Channel.
- WAN-based compression.
- Tracking multiple volume as a single consistency group.
Replication
RecoverPoint continuous data protection (CDP) tracks changes to data at a block level and journals these changes [1]. The journal then allows rolling data to a previous "Point-In-Time" in order to view the drive contents as they were before a certain data corruption. CDP can journal each write individually, hence enabling "Any-Point-In-Time" snapshots, or it can be configured to combine consecutive writes in order to reduce journal space and improve bandwidth. CDP works only over SAN - the RecoverPoint appliances needs to be zoned and masked with both the master, the replica and the journal LUNs.
RecoverPoint continuous remote replication (CRR) enables a replica in a remote site. For such a setup, RecoverPoint appliances clusters are required in both the local and remote sites. These 2 clusters communicate over IP. RecoverPoint applies compression and de-duplication in order to reduce WAN traffic. As of RecoverPoint 3.4, only one remote site. CRR can be combined with CDP in order to provide concurrent local and remote (CLR) replication.
The consistency group (CG) term is used for grouping several LUNs together in order to ensures write-order consistency over several volumes. This is used for example with a database that stores its data and journal on different logical drives. These logical drives must be kept in-sync on the replica if data-consistency needs to be preserved. Other examples are multi-volume file systems such as ZFS or Windows' Dynamic Disks. RecoverPoint 3.4 supports up to 128 CGs and 2048 LUNs [2]. Each LUN can contain up to 2 TB, and the total supported capacity can be up to 150 TB.
Write Splitting
Similar to other continuous data protection products, and unlike backup products, RecoverPoint needs to obtain a copy of every write in order to track data changes. RecoverPoint supports three methods or write splitting: host-based, fabric-based and in the storage array. EMC advertises RecoverPoint as hetrogenous due to its support of multi-vendor server, network and storage environments [3].
Host-based write splitting is done using a device driver that is installed on the server accessing the storage volumes. The usage of a host-based splitter allows replication of non-EMC storages. However, splitters are not available for all operating systems and versions.
Available fabric-based splitters are for Brocade SAN switches and for Cisco SANTap. This requires the investment in additional switch blades which cost money. This configuration allows splitting from all operating systems regardless of their version, and is agonistic to the storage array vendor.
Storage array splitters are only supported on a subset of EMC storages. This method allows write splitting from all operating systems, and does not require special SAN switching hardware. The RecoverPoint/SE is a slimmed-down version that only supports this type of splitters.
Architecture
Each site requires installation of a cluster that is composed of 2-8 RecoverPoint appliances. The multi appliances backup one another in a high availability cluster. Each appliance is connected via Fibre Channel to the SAN, and must be zoned together with both the server (SCSI initiator) and the storage (SCSI target). Each appliance must also be connected to an IP network for management.
All replication takes place over standard IP for asynchronous replication and Fibre Channel for synchronous replication.
One or more host-, fabric- or array- splitters would split traffic to both the storage and the appliances.
When configuring a consistency group, there is a need to select source LUNs on which the data will be monitored, target LUNs in the same size, and journal LUNs. The management GUI will indicate when the target LUNs are identical the the source LUNs, and will enable selecting an older timestamp in order to roll back the target LUNs to an historical state.
Integration with other products
Beyond integration with EMC products such as the CLARiiON storage array, Replication Manager and Control Center, RecoverPoint integrates with the following products:
Integration with VMWare vCenter and Microsoft Hyper-V allows protection to be specified per VM instead of per volumes that are available to the hypervisor.
Integration with Microsoft Shadow Copy, Exchange and SQL Server and Oracle Database Server allows RecoverPoint to temporarily stop writes by the host in order to take consistent application-specific snapshots.
The usage of APIs/CLIs allows customers to integrate RecoverPoint with custom internal software.[4]
Notes
- ^ http://www.emc.com/collateral/software/white-papers/h4175-recoverpoint-clr-operational-dr-wp.pdf
- ^ https://globalsp.ts.fujitsu.com/dmsp/docs/ss_recoverpoint.pdf
- ^ http://www.emc.com/collateral/software/data-sheet/h2769-emc-recoverpoint-family.pdf
- ^ http://www.emc.com/collateral/software/white-papers/h4175-recoverpoint-clr-operational-dr-wp.pdf
References
- EMC RecoverPoint, a Single Solution for CDP and Disaster Recovery
- EMC plans to use Kashya technologies in RecoverPoint
- Continuous Data Protection With EMC RecoverPoint
- EMC Updates RecoverPoint SAN CDP/Replication Engine
External links
- EMC RecoverPoint [1]