Welcome to Wikibon's Data Protection Portal. This Portal is a resource for IT professionals interested in understanding how to apply backup and recovery technologies to create a business capability that protects information in the most cost effective and appropriate manner. We invite you to be actively involved in the Data Protection Portal. Please join Wikibon, browse the portal and participate. You can write a Wikitip, Ask a Question or attend a Peer Incite Research Meeting.
The Wikibon Data Protection Information Portal contains data protection industry research, articles, expert opinion, case studies, and data storage company profiles.
Check out these Peer Incite Podcasts related to Data Protection:
1. Dave Vellante summarizes Eric Peterson of SaskEnergy's discussion about the organization's implementation of Tivoli Storage Manager (TSM) and an incremental forever backup strategy (8:36)
2. Data De-duplication: Greasing the Rails of the Backup Window. Wikibon summarizes source vs target-based De-dupe and where each makes sense for users. (6:42)
|>>Join our Group||>>Become a Fan||>>Follow @Wikibon||>>Read the Blog|
Wikibon estimates that the effects of applying compression to primary storage are additive when combined with traditional data de-duplication solutions (e.g. Data Domain, Falconstor, Diligent, etc). This is based on discussions with practitioners and an analysis of each technology.
As an example, let's assume:
In a perfect scenario, the capacity reduction ratio for the technologies in combination would yield a final 20:1 data reduction ratio for the backup stream. In a worst case scenario, the data de-duplication ratio would yield 10:1, meaning compression has no additive effect.
Wikibon believes the 'typical' rule-of-thumb is that combined, these technologies will yield a roughly 15:1 data reduction ratio, assuming these base reduction ratios and appropriate data candidates. In practice, compression on primary storage is likely to yield 30-60% improvements in capacity and as such, in real world environments the combined or 'blended' ratio would be lower but still substantially higher than data de-duplication as a standalone solution backing up non-compressed data.
Wikibon believes that this rule of thumb should work for either in-line data de-duplication (e.g. from Data Domain, Diligent, FalconStor etc., and targeted at back-up and restore), or background de-duplication (e.g. the NetApp A-SIS feature which is suitable for finding duplicate 4K blocks in on-line storage).
Featured Case Study
A data center within P&H Mining Equipment (P&H) had a problem recovering from a tape library that was too small. Recovery of files and emails was taking up to 72 hours, and IT was increasingly concerned that it would not be possible to recover from a major disaster. Rather than increase the capacity of the tape library from 500 to 1,000 tapes, P&H choose an innovative solution of a ten terabyte data de-duplication system from Data Domain.
Featured How-to Note
Implementing a failproof backup and recovery capability will protect an organization from data loss and downtime as a result of any of the following: hardware or software failure, power failure, natural disaster, or human error. There are two fundamental considerations when implementing a failproof backup and recovery capability: how quickly the organization needs to recover the data and how much data it can afford to lose. The challenge is finding the balance between data protection/recovery and the amount of investment required. This research note will provide guidelines to help make this determination.