eWEEK content and product recommendations are editorially independent. We may make money when you click on links to our partners. Learn More.
1What Impact Will Deduplication Have on Backup Performance?
High performance is essential to large enterprises that need to move exponentially growing, massive data volumes to the safety of a backup environment within a finite backup window. Understanding the performance distinctions between each category of deduplication technology-particularly as they change over time-is essential for choosing the most appropriate one for the specific environment.
2Will Deduplication Degrade Restore Performance?
3How Will Capacity and Performance Scale as the Environment Grows?
Calculate how much data you will be able to store on a single system with deduplication with your specific deduplication ratios, policies, data types and growth rate. Understand the implications of exceeding that capacity. For example, if exceeding capacity requires you to spread backups across additional systems, consider the costs of additional administrative complexity, capital expense and disruption to your environment.
4How Efficient Is the Deduplication for Large Databases?
Ensure that your deduplication IT has been optimized to handle sub-8Kb data comparisons while maintaining performance levels. Large, mission-critical databases, such as Oracle, SAP, SQL Server and DB2 typically have data change in segments of 8KB or less. However, many deduplication features cannot deliver perform comparisons of data in less than 16KB segments without dramatically slowing the backup process.
5How Efficient is the Deduplication in Progressive Incremental Backup Environments?
Some dedupe packages are inefficient in deduplicating TSM progressive incremental backups, and backups from applications that fragment their data, such as NetWorker and HP Data Protector. Ask the vendor whether the deduplication technology is able to use the metadata from these backup applications to identify the areas of data that are likely to contain duplicate data so that they can perform a byte-level comparison of that data for optimal capacity reduction-while maintaining high performance.
6What Are Realistic Expectations for Capacity Reduction?
Rather than pushing for higher generic deduplication ratios, a more effective strategy for large enterprises is to choose a solution that guarantees the ability to move data to safety within backup windows while also providing efficient deduplication. Concurrent processing and deterministic ingest rate, deduplication and replication are key enablers to an enterprise environment.
7Can Administrators Monitor Backup, Dedupe, Replication and Restore Enterprisewide?
8Can Deduplication Help Reduce Replication Bandwidth Requirements for Large Enterprise Data Volumes?
9Can IT "Tune" Deduplication to Meet Its Needs?
Enterprise data-protection environments may have data types that have special deduplication requirements. Look for solutions that enable IT to choose the datasets that they want to deduplicate by backup policy and data type, and those that automatically detect the type of data being backed up and perform. Opt for a technology that enables IT to choose the method of deduplication that is most efficient for each data type.
10How Much Experience Does the Vendor Have With Large Enterprise Backup Environments?
Enterprise data centers with massive data volumes and complex policies need a data-protection vendor with demonstrated expertise with enterprise-class backup applications, such as NetBackup, NetBackup OST and Tivoli Storage Manager. They should be prepared to provide backup assessments and guidance on how to optimize the overall backup infrastructure for maximum backup, replication and data deduplication performance in these environments.