Welcome to Wikibon's Data Protection Portal. This Portal is a resource for IT professionals interested in understanding how to apply backup and recovery technologies to create a business capability that protects information in the most cost effective and appropriate manner. We invite you to be actively involved in the Data Protection Portal. Please join Wikibon, browse the portal and participate. You can write a Wikitip, Ask a Question or attend a Peer Incite Research Meeting.
The Wikibon Data Protection Information Portal contains data protection industry research, articles, expert opinion, case studies, and data storage company profiles.
Check out these Peer Incite Podcasts related to Data Protection:
1. Dave Vellante summarizes Eric Peterson of SaskEnergy's discussion about the organization's implementation of Tivoli Storage Manager (TSM) and an incremental forever backup strategy (8:36)
2. Data De-duplication: Greasing the Rails of the Backup Window. Wikibon summarizes source vs target-based De-dupe and where each makes sense for users. (6:42)
|>>Join our Group||>>Become a Fan||>>Follow @Wikibon||>>Read the Blog|
If you're considering data deduplication/disk-based backup, the fundamental starting point should be RPO (recovery point objective) and RTO (recovery time objective). If RPO's are very tight and you want to send dedupe data over the network - then you'll want to consider server-based solutions like Avamar, which are more expensive.
If on the other hand your applications have more relaxed RPO requirements then you'll have more time to de-dupe and send de-dupe data over the net. This approach will save a bundle.
Another consideration is size and scale. If you're small (less than 10 tb) then there are a variety of vendor/product choices. If greater than 10TB, the choices narrow significantly.
The other consideration is doing in-line or post process dedupe. Data domain and Diligent, etc do inline, Falconstor and others do post process.
From a business perspective, post process allows you to do the backup faster, then you perform the dedupe. But you've added another process in the stream meaning the whole procedure takes longer and is more complex. Meaning RTO will be extended.
In line slows down the backup process but overall it's faster and easier to manage. So your RPO goes up but your RTO goes down.
You might also want to consider outsourcing disk-based backup as a remote service (e.g. IBM and others) which brings other considerations.
Featured Case Study
A data center within P&H Mining Equipment (P&H) had a problem recovering from a tape library that was too small. Recovery of files and emails was taking up to 72 hours, and IT was increasingly concerned that it would not be possible to recover from a major disaster. Rather than increase the capacity of the tape library from 500 to 1,000 tapes, P&H choose an innovative solution of a ten terabyte data de-duplication system from Data Domain.
Featured How-to Note
Implementing a failproof backup and recovery capability will protect an organization from data loss and downtime as a result of any of the following: hardware or software failure, power failure, natural disaster, or human error. There are two fundamental considerations when implementing a failproof backup and recovery capability: how quickly the organization needs to recover the data and how much data it can afford to lose. The challenge is finding the balance between data protection/recovery and the amount of investment required. This research note will provide guidelines to help make this determination.