News
6/19/2012
11:45 AM
George Crump
George Crump
Commentary
Connect Directly
RSS
E-Mail
50%
50%

Deduplication Performance: More Than Processing Power

Storage performance problems can't be solved by just throwing more processing power at them.

Deduplication, the process of identifying redundant data segments between separate files, is moving its way through the storage infrastructure. The next deduplication targets are primary storage systems and all-flash arrays. As we saw in a recent poll Storage Switzerland conducted, the number one concern is performance impact. You want deduplication, but you don't want it to impact your applications' performance.

Overcoming performance problems is more than just throwing more processing power at the problem. You need intelligent design of the deduplication logic itself. Certainly more processing power helps. As most storage systems upgrade to the latest Intel processors, they will have a new lease on life when it comes to providing all the storage services that you expect from your storage system, like snapshots, replication, and cloning.

Deduplication is a little different though, because there is a database-like lookup that must occur with most deduplication technologies. Anytime there is a lookup, the processor's speed becomes less important because it has to wait on the device that holds the table that performs the lookup. The device's speed becomes more important than processing power.

[ Learn about The Storage Problem Technology Can't Solve. ]

What causes the deduplication lookup? Most deduplication technologies create a table to store information about unique and similar data. As data is sent to a device with deduplication enabled, that data is segmented and then given a hash code. Think of it as a unique serial number. The data is then stored and the serial number or hash code is stored in the table mentioned above. As more data is sent to the device, it is also segmented and given a code. If that code matches a previous code then it is identical data. The table is updated, but the redundant data is not stored. This is where your capacity savings comes from. It is also where your potential performance bottleneck comes from.

As more and more data is stored on the system, that table grows in size. The more unique the data is, the more that table will grow, and, of course, the more total data there is, the more that table will grow. This is important when we talk about deduplication in primary storage. Primary storage does not have near the level of redundancy that backup data does. As a result, the number of entries in the hash table can be dramatically larger than it would be in a backup environment.

Why is this important? It impacts the type of storage device where you can store the lookup table. Most vendors, aware that the speed of the device where the deduplication table is stored is an issue, try to store the entire table in RAM. There is no waiting, but there is a cost and capacity issue

What happens when the number of entries, or the total dataset under management, becomes so large that the deduplication lookup table gets too big to store in RAM? You have to page to a hard disk or SSD, so, essentially, RAM becomes a cache. The problem is that normal cache logic won't work in this situation. This is not a first-in, first-out scenario. You need to verify uniqueness across the entire dataset, not just the most recently stored data.

As we will cover in our upcoming webinar "What is Breaking Deduplication?" if RAM is not used properly, then that means a repeated trip to the hard disk or SSD AND it means that the powerful processor in the storage system has to wait. Its extra processing power--and the money you spent on it--goes to waste.

Follow Storage Switzerland on Twitter

George Crump is lead analyst of Storage Switzerland, an IT analyst firm focused on the storage and virtualization segments. Storage Switzerland's disclosure statement.

Big data places heavy demands on storage infrastructure. In the new, all-digital Big Storage issue of InformationWeek Government, find out how federal agencies must adapt their architectures and policies to optimize it all. Also, we explain why tape storage continues to survive and thrive.

Comment  | 
Print  | 
More Insights
Register for Dark Reading Newsletters
Partner Perspectives
What's This?
In a digital world inundated with advanced security threats, Intel Security seeks to transform how we live and work to keep our information secure. Through hardware and software development, Intel Security delivers robust solutions that integrate security into every layer of every digital device. In combining the security expertise of McAfee with the innovation, performance, and trust of Intel, this vision becomes a reality.

As we rely on technology to enhance our everyday and business life, we must too consider the security of the intellectual property and confidential data that is housed on these devices. As we increase the number of devices we use, we increase the number of gateways and opportunity for security threats. Intel Security takes the “security connected” approach to ensure that every device is secure, and that all security solutions are seamlessly integrated.
Featured Writers
White Papers
Cartoon
Current Issue
Dark Reading's October Tech Digest
Fast data analysis can stymie attacks and strengthen enterprise security. Does your team have the data smarts?
Flash Poll
10 Recommendations for Outsourcing Security
10 Recommendations for Outsourcing Security
Enterprises today have a wide range of third-party options to help improve their defenses, including MSSPs, auditing and penetration testing, and DDoS protection. But are there situations in which a service provider might actually increase risk?
Video
Slideshows
Twitter Feed
Dark Reading - Bug Report
Bug Report
Enterprise Vulnerabilities
From DHS/US-CERT's National Vulnerability Database
CVE-2013-4594
Published: 2014-10-25
The Payment for Webform module 7.x-1.x before 7.x-1.5 for Drupal does not restrict access by anonymous users, which allows remote anonymous users to use the payment of other anonymous users when submitting a form that requires payment.

CVE-2014-0476
Published: 2014-10-25
The slapper function in chkrootkit before 0.50 does not properly quote file paths, which allows local users to execute arbitrary code via a Trojan horse executable. NOTE: this is only a vulnerability when /tmp is not mounted with the noexec option.

CVE-2014-1927
Published: 2014-10-25
The shell_quote function in python-gnupg 0.3.5 does not properly quote strings, which allows context-dependent attackers to execute arbitrary code via shell metacharacters in unspecified vectors, as demonstrated using "$(" command-substitution sequences, a different vulnerability than CVE-2014-1928....

CVE-2014-1928
Published: 2014-10-25
The shell_quote function in python-gnupg 0.3.5 does not properly escape characters, which allows context-dependent attackers to execute arbitrary code via shell metacharacters in unspecified vectors, as demonstrated using "\" (backslash) characters to form multi-command sequences, a different vulner...

CVE-2014-1929
Published: 2014-10-25
python-gnupg 0.3.5 and 0.3.6 allows context-dependent attackers to have an unspecified impact via vectors related to "option injection through positional arguments." NOTE: this vulnerability exists because of an incomplete fix for CVE-2013-7323.

Best of the Web
Dark Reading Radio
Archived Dark Reading Radio
Follow Dark Reading editors into the field as they talk with noted experts from the security world.