News
6/17/2010
08:34 AM
George Crump
George Crump
Commentary
50%
50%

Keep Everything Forever, Part II - Indexing

In our last entry we reintroduced the idea of a keep everything forever storage retention strategy. We also touched on some of the basic capabilities like cost effective storage options and data movement options that can make a forever retention strategy realistic. In this entry we will look at what is one of the most important requirements the ability to find what you have in the archive.

In our last entry we reintroduced the idea of a keep everything forever storage retention strategy. We also touched on some of the basic capabilities like cost effective storage options and data movement options that can make a forever retention strategy realistic. In this entry we will look at what is one of the most important requirements the ability to find what you have in the archive.The fact that you have a keep it all retention strategy is going to be of little value to you if you can't find a discrete piece of information when you need it. Typically you are keeping all this data in case at some point down the road you need to produce it. Most often this will be in response to some sort of legal action or needing to prove adherence to a regulation. While you don't need rapid restore speed in these cases you do need to be able to deliver it in a timely manner. You can longer throw people at the problem and manually dig through information. In some cases you have to have a system in place that can not only find what you do have but also prove that you don't have something.

This comes down to implementing a solution that can index all the information that you have on storage and give you the ability to search on key words or phrases. I believe this index has to be storage vendor agnostic and potentially even device type agnostic. For example the ability that some backup applications have to build index information on the data that passes through it is good but the only thing being indexed is just that information that passes through the application. What about the data that does not get backed up by that application? Also what if you change your mind and switch applications? Are you really prepared to lock into a particular backup application for decades? And if you were do you think it is realistic to expect to send all your data through that application forever?

Beyond being storage and vendor agnostic the indexing system has to be able to scale to handle the ever growing amount of data that it will be responsible for. Scale here comes in two ways. The first is the ability to scale to meet ingestion rates. In other words how much information can the indexing system process in a given window? The other is how large can the meta-data that the indexing system creates scale to? If you are planning on keeping information forever that meta-data catalog's performance becomes critical.

The final piece also relates to the meta-data that the indexing system creates, how space efficient is the meta-data? You don't want to have to double your capacity requirements to hold all this meta-data, so the indexing system has to be able to capture the information it needs but do so very space efficiently.

Finding information in a keep it forever strategy is important, but probably most important is if you can actually afford to keep it forever. Our final entry in this series will cover trying to determine if a keep it forever strategy is cost effective.

Track us on Twitter: http://twitter.com/storageswiss

Subscribe to our RSS feed.

George Crump is lead analyst of Storage Switzerland, an IT analyst firm focused on the storage and virtualization segments. Find Storage Switzerland's disclosure statement here.

Comment  | 
Print  | 
More Insights
Register for Dark Reading Newsletters
White Papers
Cartoon
Current Issue
Flash Poll
10 Recommendations for Outsourcing Security
10 Recommendations for Outsourcing Security
Enterprises today have a wide range of third-party options to help improve their defenses, including MSSPs, auditing and penetration testing, and DDoS protection. But are there situations in which a service provider might actually increase risk?
Video
Slideshows
Twitter Feed
Dark Reading - Bug Report
Bug Report
Enterprise Vulnerabilities
From DHS/US-CERT's National Vulnerability Database
CVE-2015-0732
Published: 2015-07-28
Cross-site scripting (XSS) vulnerability in Cisco AsyncOS on the Web Security Appliance (WSA) 9.0.0-193; Email Security Appliance (ESA) 8.5.6-113, 9.1.0-032, 9.1.1-000, and 9.6.0-000; and Content Security Management Appliance (SMA) 9.1.0-033 allows remote attackers to inject arbitrary web script or ...

CVE-2015-2974
Published: 2015-07-28
LEMON-S PHP Gazou BBS plus before 2.36 allows remote attackers to upload arbitrary HTML documents via vectors involving a crafted image file.

CVE-2015-4287
Published: 2015-07-28
Cisco Firepower Extensible Operating System 1.1(1.86) on Firepower 9000 devices allows remote attackers to bypass intended access restrictions and obtain sensitive device information by visiting an unspecified web page, aka Bug ID CSCuu82230.

CVE-2015-4288
Published: 2015-07-28
The LDAP implementation on the Cisco Web Security Appliance (WSA) 8.5.0-000, Email Security Appliance (ESA) 8.5.7-042, and Content Security Management Appliance (SMA) 8.3.6-048 does not verify X.509 certificates from SSL servers, which allows man-in-the-middle attackers to spoof servers and obtain s...

CVE-2015-4692
Published: 2015-07-27
The kvm_apic_has_events function in arch/x86/kvm/lapic.h in the Linux kernel through 4.1.3 allows local users to cause a denial of service (NULL pointer dereference and system crash) or possibly have unspecified other impact by leveraging /dev/kvm access for an ioctl call.

Dark Reading Radio
Archived Dark Reading Radio
What’s the future of the venerable firewall? We’ve invited two security industry leaders to make their case: Join us and bring your questions and opinions!