News
1/17/2011
12:11 PM
George Crump
George Crump
Commentary
50%
50%

Backup Deduplication 2.0 - Power Savings

In our last entry we opened a discussion of what is needed as we move into the next era of backup deduplication and focused on integration to backup software. Another area that is becoming increasingly important is to be able to lower the power requirements that disk backup deduplication hardware requires. Power is a pressing issue in the data center and disk backup systems need to address those concerns

In our last entry we opened a discussion of what is needed as we move into the next era of backup deduplication and focused on integration to backup software. Another area that is becoming increasingly important is to be able to lower the power requirements that disk backup deduplication hardware requires. Power is a pressing issue in the data center and disk backup systems need to address those concerns.When it comes to power consumption most IT professionals will think of spin down drives, drives that can either slow down or power off depending on when they were last accessed. Deduplication vendors are going to want you to think differently about that. Instead, focus on how many fewer drives deduplication will use than a standard disk backup approach. While that is a fair line of reasoning, at the end of the day spinning drives, no matter how optimized, use more power than disk backup's biggest competitor, tape.

The answer to one disk backups biggest weakness is to figure out how to integrate power managed drives into disk deduplication systems. The use of these drives can be troublesome with disk solutions that use deduplication. Deduplication makes heavy use of indexing to identify redundant data, it performs frequent data integrity checks and often use garbage collection techniques to remove old data that no longer has active pointers. All of this constant access makes it difficult to spin down a drive for any significant amount of time.

There are ways to get some power management in backup deduplication systems. For example you can add deduplication technology to a spin down system as we discussed in "Power Managed Dedupe". The deduplication software can be optimized to narrow its garbage collection windows and error checking so the system could be in a spin down mode for the bulk of the non-backup window. Further multiples of these systems could be used over time with backup re-directed to different units at different times, alternating by quarter for example. The downside to this approach of course would be some increase in redundancy of the backup data set but it would increase power optimization. Over time though deduplication systems are going to have to learn to self-isolate old data.

The backup software applications that can do their own deduplication may be able to perform this for you as well. By setting up different drive groups in a power managed array or even using different arrays you could send deduplicated backup data to backup pools, which would give the system more time to power the drives down.

Clustered or scale out based disk backup systems are going to have to take all of this a step further, since each node is a potential power consumer. They are going to have to be able to move data to older nodes and then power or at least idle those nodes down. Steps could be taken to not only power the drives down but lower fan speed and processor speed which could lead to a very efficient scale out story. That is either going to require sophisticated node communication or an internal sub-dividing of the nodes to segregate off the infrequently accessed data set.

Another option for power efficiency is to use backup virtualization as we talked about in our recent article "Backup Virtualization Brings Flexibility to Disk Backup". Leveraging this technology backups could be sent to a very small high speed disk cache, then quickly spool that data to a disk deduplication system for medium term storage and then finally spill to tape as the data becomes old and infrequently accessed. This gives you the use of each backup device for what it is already best at instead of waiting for technology to fill in the gaps.

Track us on Twitter: http://twitter.com/storageswiss

Subscribe to our RSS feed.

George Crump is lead analyst of Storage Switzerland, an IT analyst firm focused on the storage and virtualization segments. Find Storage Switzerland's disclosure statement here.

Comment  | 
Print  | 
More Insights
Register for Dark Reading Newsletters
White Papers
Cartoon
Current Issue
Dark Reading Tech Digest, Dec. 19, 2014
Software-defined networking can be a net plus for security. The key: Work with the network team to implement gradually, test as you go, and take the opportunity to overhaul your security strategy.
Flash Poll
10 Recommendations for Outsourcing Security
10 Recommendations for Outsourcing Security
Enterprises today have a wide range of third-party options to help improve their defenses, including MSSPs, auditing and penetration testing, and DDoS protection. But are there situations in which a service provider might actually increase risk?
Video
Slideshows
Twitter Feed
Dark Reading - Bug Report
Bug Report
Enterprise Vulnerabilities
From DHS/US-CERT's National Vulnerability Database
CVE-2014-2208
Published: 2014-12-28
CRLF injection vulnerability in the LightProcess protocol implementation in hphp/util/light-process.cpp in Facebook HipHop Virtual Machine (HHVM) before 2.4.2 allows remote attackers to execute arbitrary commands by entering a \n (newline) character before the end of a string.

CVE-2014-2209
Published: 2014-12-28
Facebook HipHop Virtual Machine (HHVM) before 3.1.0 does not drop supplemental group memberships within hphp/util/capability.cpp and hphp/util/light-process.cpp, which allows remote attackers to bypass intended access restrictions by leveraging group permissions for a file or directory.

CVE-2014-5386
Published: 2014-12-28
The mcrypt_create_iv function in hphp/runtime/ext/mcrypt/ext_mcrypt.cpp in Facebook HipHop Virtual Machine (HHVM) before 3.3.0 does not seed the random number generator, which makes it easier for remote attackers to defeat cryptographic protection mechanisms by leveraging the use of a single initial...

CVE-2014-6228
Published: 2014-12-28
Integer overflow in the string_chunk_split function in hphp/runtime/base/zend-string.cpp in Facebook HipHop Virtual Machine (HHVM) before 3.3.0 allows remote attackers to cause a denial of service (application crash) or possibly have unspecified other impact via crafted arguments to the chunk_split ...

CVE-2014-6229
Published: 2014-12-28
The HashContext class in hphp/runtime/ext/ext_hash.cpp in Facebook HipHop Virtual Machine (HHVM) before 3.3.0 incorrectly expects that a certain key string uses '\0' for termination, which allows remote attackers to obtain sensitive information by leveraging read access beyond the end of the string,...

Best of the Web
Dark Reading Radio
Archived Dark Reading Radio
Join us Wednesday, Dec. 17 at 1 p.m. Eastern Time to hear what employers are really looking for in a chief information security officer -- it may not be what you think.