The challenges with this method are that it creates two storage areas to manage, an area that is waiting to be examined for duplicates and an area for the examined data. It also delays time to create a DR copy of data. A common use for deduplicated systems is to leverage their ability to only store unique data segments and replicate only those new segments to the remote location. With the post-process method, you have to wait until the deduplication step is complete, until data can be replicated. The post-process step can be very time consuming and delay the update of the DR site by 6-10 hours.
As a result, companies that started with data deduplication as a core (Data Domain,Permabit, Diligent) part of their technology have a distinct advantage. The other companies will have to make the post-process data deduplication much more seamless than it is today, exit from deduplication altogether, or re-write their code bases to support in-line data deduplication.
George Crump is founder of Storage Switzerland, an analyst firm focused on the virtualization and storage marketplaces. It provides strategic consulting and analysis to storage users, suppliers, and integrators. An industry veteran of more than 25 years, Crump has held engineering and sales positions at various IT industry manufacturers and integrators. Prior to Storage Switzerland, he was CTO at one of the nation's largest integrators.