Buyers’ guide: Data deduplication

Exponential data growth is emerging as arguably the biggest issue IT managers face. But, handily, data deduplication is one available tool to help tackle the problem. Tim Lohman reports.

Assessing a Dedupe Solution

So, it’s pretty clear that dedupe can be a pretty powerful technology, but given the breadth of offerings in the marketplace — either as part of a wider suite of information management tools or as standalone solution — consideration needs to be given to a few things before deciding on a solution.

The first step in this process, Forrester analyst, Tim Sheedy advises, is to take a step back and ask whether the dedupe solution you are acquiring is to address a single point problem or whether there is a broader organisational requirement for dedupe.

“Is it a single requirement to get rid of excess customer records or do you need to do broader data quality, cleansing or enhancement, as that’s when you’ll begin looking at an external third party provider as opposed to buying a piece of software and managing it yourself,” Sheedy says.

Assessing whether you need to do the deduping in-line or after-the-fact will also drive you to a different set of vendors, as will assessing how well a potential vendor sits within your broader architecture. As mentioned above, you’ll also want to consider whether a broader suite of information tools, rather than a stand alone data dedupe solution is appropriate.

“The problem with the whole information management landscape over the past few years is that you have a single piece of software for data quality, another for dedupe, another for data loading and ETL [extract, transform, and load] so that you end up managing 15 or 20 vendors and you have to integrate it all yourself or pay a lot of money to an IBM, CSC or Accenture to come in and do it for you,” Sheedy says.

“The advantage of the big companies is that they integrate all this themselves so you know your BI and ETL and other applications will work together.”

Another crucial consideration is to determine whether post- or pre-processing dedupe is right for you. In the post-processing camp, TechnologyOne’s Bauer says his organisation decided on its particular solution for reasons of speed and effectiveness.

“We found [post-processing] dedupe generates better dedupe than the in-line as it has more time and available resources to process the deduping and assess the blocks against the dedupe pattern it has already created,” he says.

“The advantage of block-level dedupe is that, while we initially used it against CIFS [Common Internet File System] data, but we can also use it against block level storage LUNs [Logical Unit Number] that have virtual machines mounted against them or raw disc served up out of the SAN (storage Area Network), we can dedupe that as well.”

The results, Bauer says, were a 50 per cent-plus reduction on file server data, and in some instances up to 90 per cent reduction on source code repository data.

In the pre-processing camp, Brad Jansons, IT Manager at Australian Motoring Services (AMS) argues that in-line, in its case the Hitachi Data Protection Suite from Commvault, was the way to reduce storage costs, store more data more efficiently, and decrease back up and restore times.

“In-line has the advantage over post-processing as with the latter you have to have the space available there in the first place — to store the excess data — which partly defeats the purpose of dedupe,” he says.

“As the dedupe is running to the local SAN there’s no issue with performance and the amount of data we are deduping means that in-line-related performance issues aren’t an issue. We are also not a 24/7 business so backups are done outside of hours.”

With its in-line solution, AMS has seen up to 50 per cent reduction in disc volumes and has been able to move to disc, instead of tape, for its backup and recovery. Moving to dedupe also saved the organisation from investing in another tray of high performance disc for its SAN.

Lastly, in devising a business case, Gartner advises that it is also worth considering whether the bandwidth savings created by the use of dedupe make disaster recovery, centralised backup of remote offices, and application deployment methodologies operationally and economically viable.

Dedupe Tricks and Tips When it comes to dedupe, one of the more important tips is to understand that not all data can be deduplicated.

Compressed data, such as pictures, voice, and video files, and encrypted content are examples which dedupe doesn’t necessarily lend itself to. However, dedupe can be an effective data-reduction method for data sets that contain common bit strings. Clearly it’s important to choose a dedupe solution that is right for your business and needs.

But another tip is to keep an eye out for the creation of hot spots in your storage environment. As TechnologyOne’s Bauer explains, dedupe effectively takes a large amount of data and reduces it down to a smaller amount stored on a reduced number of discs. As a result, the increased number of data requests against a smaller number of discs could lead to the I/O capability of those discs being overloaded.

“You will have an end-user impact and they will find it is slower to access data from that set of discs,” Bauer says. “Instead of having 100 users accessing files across 100 discs, you will have 100 users accessing files on one disc.”

The software developer managed the issue by making sure that deduped data was separated out across a number of higher performance discs.

“If we knew there would be a higher I/O load then we made sure it would be on higher performance disc,” Bauer says. “What it means is that you can have a small number of high performance discs for high-load data, and cheaper lower performance disc for the low-load data. All up you end with less discs and less cost.”

Data deduplication products

Hitachi Data Protection Suite: Includes Data deduplication which provides users the choice of when and where to use deduplication — remote sites, local file and databases, or in reducing an organisation’s data footprint when transferred to tape. The wider suite incorporates backup and recovery, point-in-time replication, archiving, and storage resource management under a single graphical user interface. tinyurl.com/qt6wlz

Double-Take Atom Deduplication: A file-level deduplication feature of the Double-Take Backup, Atom Deduplication scans the Double-Take Backup repository in real time to locate duplicate files among all the data. Replicates and stores only byte-level changes to the protected data, minimising any duplication caused by the back up solution itself. tinyurl.com/368k2zv

EMC Avamar: Enables data reduction and secure backup for VMware environments, remote offices, LAN/NAS servers, and desktop/laptop systems, and reduces backup time, growth of secondary storage, and network utilisation. EMC claims to reduce daily back up data up to 500x, back up times up to 10x, and total storage up to 50x. tinyurl.com/bkgokp

Quantum DXi series: Hardware-based DXi-series claims an average 125 per cent increase in backup performance, 87 per cent fewer failed backup jobs, and typical disk capacity reductions of 90 per cent or more and 95 per cent in virtual environments. tinyurl.com/2vmoyog

NetApp Deduplication: Part of NetApp’s ONTAP architecture, NetApp deduplication can be used across primary data, backup data, and archival data. Users can schedule deduplication to occur during off-peak times, select which datasets to deduplicate, and perform a full byte-for-byte validation before removing any duplicate data. tinyurl.com/mjprw8

Symantec Backup Exec 2010: Integrated data deduplication allowing dedupe at the source or remote server, at the media server, deduplication appliance-level or on data travelling from remote offices to headquarters. tinyurl.com/6hx22w

Join the newsletter!

Or

Sign up to gain exclusive access to email subscriptions, event invitations, competitions, giveaways, and much more.

Membership is free, and your security and privacy remain protected. View our privacy policy before signing up.

Error: Please check your email address.

Tags storagedata deduplicationForrester ResearchTechnologyOneCurtin UniversityAustralian Motoring Services

More about Accenture AustraliaAndrew Corporation (Australia)CSC AustraliaCurtin UniversityCurtin University of TechnologyEMC CorporationetworkGartnerHitachi AustraliaHitachi DataIBM AustraliaIBM AustraliaIFSLANLogicalMicrosoftNASNetAppNetAppQuantumSymantecVMware Australia

Show Comments
[]