Dark Reading is part of the Informa Tech Division of Informa PLC

This site is operated by a business or businesses owned by Informa PLC and all copyright resides with them.Informa PLC's registered office is 5 Howick Place, London SW1P 1WG. Registered in England and Wales. Number 8860726.

News

9/15/2008
12:16 PM
George Crump
George Crump
Commentary
50%
50%

Survived Ike? Time Will Tell...

Ike tore through Houston, home one of our office and our lab. Once again, businesses have to learn that surviving the initial hit is only the beginning. The "P" part of a Disaster Recovery Plan is very critical to the long-term survival of the business.

Ike tore through Houston, home one of our office and our lab. Once again, businesses have to learn that surviving the initial hit is only the beginning. The "P" part of a Disaster Recovery Plan is very critical to the long-term survival of the business.About half our team and our lab happens to be in Houston. Everyone on our team there is safe and, while we avoided the hard hit that Galveston and the south sides of Houston received from Hurricane Ike, our lab is without power, as are many Houston-area businesses, and we are being told that it will be that way for weeks.

We, like hopefully most businesses, have our data stored on disaster-proof hard drives as well as replicated out to our Dallas office (where I safely rode out the storm). For businesses in the greater Houston area, that "P" part of Disaster Recovery Planning is now being tested. So much work is spent on getting the data out, and even doing testing, but often DR tests are a "pass/fail" sort of occurrence. It looks like Ike is going to be a longer-term situation.

We have all heard the stats of businesses that go out of business after a major disaster. Many times, while the disaster shuts the doors, it's the prolonged time away that puts the lock on the doors. So after the obvious first step of getting the data out of the facility, one that I and countless others have written about, comes living with the DR sites capabilities.

Having the data safe is only half the battle. Doing something with that data, and for a prolonged period of time, is the other half. Here are some challenges of users that I am talking to.

• First, a common mistake is not getting the data far enough away or not having your DR location far enough away. Right now, much of Houston is without power, and except for facilities that have generators, it's possible that both your primary and your DR site may be down.

• Second, many businesses replicated to a cheaper, possibly less reliable and lower performing, storage system. There is serious concern about running mission-critical apps on these systems for the next few weeks. When you are designing your disaster recovery environment, make sure you have storage in the remote location suitable to sustain the performance and reliability that your applications demand.

• Third, some customers "over-virtualized" their DR site. One of server virtualization's big appeals is to be able to put less hardware in the DR site by virtualizing much of the server environment in the DR location. Is the virtual infrastructure able to sustain operations for a few weeks or will that environment be over-burdened? This morning, customers are scrambling to order additional hardware to spread out this load. As part of your DR testing, make sure you don't overburden the DR servers and make sure you know who to call to quickly bring up another virtualization host and distribute that load.

• Fourth, there are many instances of no backup at the DR site! The DR site just became production for many businesses and when it did that, it now needs protection. There are many sites running right now without the safety net of a backup. As part of your planning, make sure you keep the backup capability current with the needs of the DR site. Also make sure you have the ability to replicate data out of your DR site to another site, just in case you're unlucky enough to have a problem at your DR site, too.

One of the critical elements in disaster planning is to plan for a long-term activation of your DR site. Always factor in at least four weeks of remote operation and make sure your DR site is designed to handle that workload.

• The final step is planning for failback. One of the most challenging steps, and typically the least planned or tested, is the failback, the process of returning production to its original location after a disaster. We'll discuss that next entry.

Track us on Twitter: http://twitter.com/storageswiss.

Subscribe to our RSS feed.

George Crump is founder of Storage Switzerland, an analyst firm focused on the virtualization and storage marketplaces. It provides strategic consulting and analysis to storage users, suppliers, and integrators. An industry veteran of more than 25 years, Crump has held engineering and sales positions at various IT industry manufacturers and integrators. Prior to Storage Switzerland, he was CTO at one of the nation's largest integrators.

Comment  | 
Print  | 
More Insights
Comments
Newest First  |  Oldest First  |  Threaded View
COVID-19: Latest Security News & Commentary
Dark Reading Staff 5/28/2020
Stay-at-Home Orders Coincide With Massive DNS Surge
Robert Lemos, Contributing Writer,  5/27/2020
Register for Dark Reading Newsletters
White Papers
Video
Cartoon Contest
Write a Caption, Win a Starbucks Card! Click Here
Latest Comment: This comment is waiting for review by our moderators.
Current Issue
How Cybersecurity Incident Response Programs Work (and Why Some Don't)
This Tech Digest takes a look at the vital role cybersecurity incident response (IR) plays in managing cyber-risk within organizations. Download the Tech Digest today to find out how well-planned IR programs can detect intrusions, contain breaches, and help an organization restore normal operations.
Flash Poll
Twitter Feed
Dark Reading - Bug Report
Bug Report
Enterprise Vulnerabilities
From DHS/US-CERT's National Vulnerability Database
CVE-2020-11844
PUBLISHED: 2020-05-29
There is an Incorrect Authorization vulnerability in Micro Focus Service Management Automation (SMA) product affecting version 2018.05 to 2020.02. The vulnerability could be exploited to provide unauthorized access to the Container Deployment Foundation.
CVE-2020-6937
PUBLISHED: 2020-05-29
A Denial of Service vulnerability in MuleSoft Mule CE/EE 3.8.x, 3.9.x, and 4.x released before April 7, 2020, could allow remote attackers to submit data which can lead to resource exhaustion.
CVE-2020-7648
PUBLISHED: 2020-05-29
All versions of snyk-broker before 4.72.2 are vulnerable to Arbitrary File Read. It allows arbitrary file reads for users who have access to Snyk's internal network by appending the URL with a fragment identifier and a whitelisted path e.g. `#package.json`
CVE-2020-7650
PUBLISHED: 2020-05-29
All versions of snyk-broker after 4.72.0 including and before 4.73.1 are vulnerable to Arbitrary File Read. It allows arbitrary file reads to users with access to Snyk's internal network of any files ending in the following extensions: yaml, yml or json.
CVE-2020-7654
PUBLISHED: 2020-05-29
All versions of snyk-broker before 4.73.1 are vulnerable to Information Exposure. It logs private keys if logging level is set to DEBUG.