News
7/14/2008
01:58 PM
George Crump
George Crump
Commentary
50%
50%

Block-Level Tiered Storage

Tiered storage no longer has the hype surrounding it that it did a few years ago. The concept was simple -- move data from expensive Fibre drives to inexpensive SATA drives. SATA drive technology was just coming into its own and the price and modest capacity made it a good fit for the concept. As a result, every storage manufacturer on the planet was proposing a tiered storage strategy. There were seminars, Webinars, white papers (guilty as charged, I wrote more than a few of them), yet only a f

Tiered storage no longer has the hype surrounding it that it did a few years ago. The concept was simple -- move data from expensive Fibre drives to inexpensive SATA drives. SATA drive technology was just coming into its own and the price and modest capacity made it a good fit for the concept. As a result, every storage manufacturer on the planet was proposing a tiered storage strategy. There were seminars, Webinars, white papers (guilty as charged, I wrote more than a few of them), yet only a fraction of accounts ever implemented the strategy.Why? Well, it was movement. Movement of data from Point A to Point B in the data center is always a challenge. It is a challenge for backups, it is a challenge for replication, and it is a challenge for ILM or archive strategies. Over the last few weeks I've written about several types of data movement strategies, but none are as simple to implement as block-level tiered storage.

While my term for it isn't as catchy as 3PAR's Nearline for Online or Compellent's data progression, the concept is to have the storage system monitor data at a block level and be able to set policies based on the block itself, typically on the block's age. Basically, if a block of data hasn't been accessed in the last 90 days, move it from expensive Fibre Channel disk to SATA disk. There are two key advantages of this method compared with traditional data movement techniques.

First, block-level tiered storage can operate on almost any type of data. Traditional archiving is focused on file level data, and while there are a few database archive utilities available, they're not in widespread use. Second, the use can be expanded beyond just a simple move from Fibre to ATA. It could be a move from a mirrored Fibre 15k RAID group to a standard RAID 5 group made up with 10k Fibre drives and then eventually to ATA.

Block-level tiered storage works across all platforms connected to the array, with no additional agents to install on those servers, no crawl or file system walks like discussed in previous entries, no worries about upgrades to the OS breaking your data movement application. Essentially, you activate the capability to define your policies and then it just works.

While not a replacement for an archive strategy, this is possibly one of the more practical methods to implement tiered storage. Archiving is going to require understanding the file as a whole, setting retention strategies around that file, possible control modification based on the type of file, and also having the ability to search the content of that file. That said, block-level tiered storage is a cost-effective way to manage the active or near-active data set.

George Crump is founder of Storage Switzerland, an analyst firm focused on the virtualization and storage marketplaces. It provides strategic consulting and analysis to storage users, suppliers, and integrators. An industry veteran of more than 25 years, Crump has held engineering and sales positions at various IT industry manufacturers and integrators. Prior to Storage Switzerland, he was CTO at one of the nation's largest integrators.

Comment  | 
Print  | 
More Insights
Register for Dark Reading Newsletters
White Papers
Cartoon
Current Issue
Dark Reading December Tech Digest
Experts weigh in on the pros and cons of end-user security training.
Flash Poll
10 Recommendations for Outsourcing Security
10 Recommendations for Outsourcing Security
Enterprises today have a wide range of third-party options to help improve their defenses, including MSSPs, auditing and penetration testing, and DDoS protection. But are there situations in which a service provider might actually increase risk?
Video
Slideshows
Twitter Feed
Dark Reading - Bug Report
Bug Report
Enterprise Vulnerabilities
From DHS/US-CERT's National Vulnerability Database
CVE-2014-4807
Published: 2014-11-22
Sterling Order Management in IBM Sterling Selling and Fulfillment Suite 9.3.0 before FP8 allows remote authenticated users to cause a denial of service (CPU consumption) via a '\0' character.

CVE-2014-6183
Published: 2014-11-22
IBM Security Network Protection 5.1 before 5.1.0.0 FP13, 5.1.1 before 5.1.1.0 FP8, 5.1.2 before 5.1.2.0 FP9, 5.1.2.1 before FP5, 5.2 before 5.2.0.0 FP5, and 5.3 before 5.3.0.0 FP1 on XGS devices allows remote authenticated users to execute arbitrary commands via unspecified vectors.

CVE-2014-8626
Published: 2014-11-22
Stack-based buffer overflow in the date_from_ISO8601 function in ext/xmlrpc/libxmlrpc/xmlrpc.c in PHP before 5.2.7 allows remote attackers to cause a denial of service (application crash) or possibly execute arbitrary code by including a timezone field in a date, leading to improper XML-RPC encoding...

CVE-2014-8710
Published: 2014-11-22
The decompress_sigcomp_message function in epan/sigcomp-udvm.c in the SigComp UDVM dissector in Wireshark 1.10.x before 1.10.11 allows remote attackers to cause a denial of service (buffer over-read and application crash) via a crafted packet.

CVE-2014-8711
Published: 2014-11-22
Multiple integer overflows in epan/dissectors/packet-amqp.c in the AMQP dissector in Wireshark 1.10.x before 1.10.11 and 1.12.x before 1.12.2 allow remote attackers to cause a denial of service (application crash) via a crafted amqp_0_10 PDU in a packet.

Best of the Web
Dark Reading Radio
Archived Dark Reading Radio
Now that the holiday season is about to begin both online and in stores, will this be yet another season of nonstop gifting to cybercriminals?