Risk
4/20/2012
01:29 PM
George Crump
George Crump
Commentary
50%
50%

How To Protect Big Data Analytics

Big data analytics often means big challenges when it comes to data protection. Here are some things to keep in mind when you're working in these environments.

Big Data Talent War: 10 Analytics Job Trends
Big Data Talent War: 10 Analytics Job Trends
(click image for larger view and for slideshow)
Data protection is often the forgotten part of any trend in the data center, and the launch of big data initiatives is no exception to this trend. Data protection is too often an afterthought. What is particularly challenging with big data, especially big data analytics, as I discussed in a recent column, is that it is the perfect storm for a data protection disaster.

Big data analytics has all the things you don’t want to see when you're trying to protect data. First, it can have a very unique sample set--for example, a device that monitors a soil sample every 30 seconds, a camera that takes thousands of images every minute, or a cell phone call center that logs millions of text messages. All that data is unique to that moment; if it is lost it is impossible to recreate.

That uniqueness also means that the data is probably not deduplicatable. As I discussed in a recent article, you may need to either turn off deduplication, or at least factor in a very low effective rate, in such environments. This means that the capacity of the backup appliance may have to be close to what the real data set is than in other backup situations where you may be counting on a high level of dedupe effectiveness.

[ Bigger data sets mean bigger compliance challenges. Read more at Big Data's Dark Side: Compliance Issues. ]

The large number of files that can be resident in big data analytic environments is also a challenge. In order for the backup application and the appliance to churn through this large number of files, the bandwidth to the backup server and/or the backup appliance needs to be large, and the receiving devices must be able to ingest data at the rate that the data can be delivered. They also need significant CPU processing power to churn through billions of files.

There's also a database component to big data that needs to be considered. Analytic information is often processed into either an Oracle or Hadoop environment of some sort, so live protection of that environment may be required. This means a smaller number of larger files need to be backed up.

This is a worst-case mix workload of high performance: billions of small files with a small number of large files, which may break many backup appliances. Finding one that can ingest this mixed workload of data at full speed, that has a deduplication configuration that won't impact performance, and that can scale to massive capacities may be the biggest challenge in the big data backup market. You may have to consider tape, and if so, the disk backup vendor needs to know how to work with it.

The other form of big data, big data archive, should be less of an issue if it's designed correctly. If the design uses tape as part of the archive, then backup can be built in as part of the workflow. Designing the storage infrastructure for big data archive environments will be the subject of an upcoming column.

Follow Storage Switzerland on Twitter

George Crump is lead analyst of Storage Switzerland, an IT analyst firm focused on the storage and virtualization segments. Storage Switzerland's disclosure statement.

The Enterprise 2.0 Conference brings together industry thought leaders to explore the latest innovations in enterprise social software, analytics, and big data tools and technologies. Learn how your business can harness these tools to improve internal business processes and create operational efficiencies. It happens in Boston, June 18-21. Register today!

Comment  | 
Print  | 
More Insights
Comments
Oldest First  |  Newest First  |  Threaded View
JTAYLOR9009
50%
50%
JTAYLOR9009,
User Rank: Apprentice
4/20/2012 | 7:36:11 PM
re: How To Protect Big Data Analytics
Interesting article. RDBMS & NoSQL both have their places in the world. I like to think of them like tools; sometime you need a sledgehammer and sometimes you need a chisel. Still equally important.
MarkNeilson
50%
50%
MarkNeilson,
User Rank: Apprentice
2/23/2014 | 8:38:51 AM
re: How To Protect Big Data Analytics
Analytics is always essential to take care of the de-duplication problem. The analytics clearly explains the overall issues and the increased demand of data management in the upcoming years. If the data is bigger there will be compliance issues. So, it is quite important to take care of such data related issues along with its quality and management. You can also take the help of Data Cleansing software that will solve all your issues.
Register for Dark Reading Newsletters
White Papers
Cartoon
Current Issue
Flash Poll
Video
Slideshows
Twitter Feed
Dark Reading - Bug Report
Bug Report
Enterprise Vulnerabilities
From DHS/US-CERT's National Vulnerability Database
CVE-2011-4403
Published: 2015-04-24
Multiple cross-site request forgery (CSRF) vulnerabilities in Zen Cart 1.3.9h allow remote attackers to hijack the authentication of administrators for requests that (1) delete a product via a delete_product_confirm action to product.php or (2) disable a product via a setflag action to categories.ph...

CVE-2012-2930
Published: 2015-04-24
Multiple cross-site request forgery (CSRF) vulnerabilities in TinyWebGallery (TWG) before 1.8.8 allow remote attackers to hijack the authentication of administrators for requests that (1) add a user via an adduser action to admin/index.php or (2) conduct static PHP code injection attacks in .htusers...

CVE-2012-2932
Published: 2015-04-24
Multiple cross-site scripting (XSS) vulnerabilities in TinyWebGallery (TWG) before 1.8.8 allow remote attackers to inject arbitrary web script or HTML via the (1) selitems[] parameter in a copy, (2) chmod, or (3) arch action to admin/index.php or (4) searchitem parameter in a search action to admin/...

CVE-2012-5451
Published: 2015-04-24
Multiple stack-based buffer overflows in HttpUtils.dll in TVMOBiLi before 2.1.0.3974 allow remote attackers to cause a denial of service (tvMobiliService service crash) via a long string in a (1) GET or (2) HEAD request to TCP port 30888.

CVE-2015-0297
Published: 2015-04-24
Red Hat JBoss Operations Network 3.3.1 does not properly restrict access to certain APIs, which allows remote attackers to execute arbitrary Java methos via the (1) ServerInvokerServlet or (2) SchedulerService or (3) cause a denial of service (disk consumption) via the ContentManager.

Dark Reading Radio
Archived Dark Reading Radio
Join security and risk expert John Pironti and Dark Reading Editor-in-Chief Tim Wilson for a live online discussion of the sea-changing shift in security strategy and the many ways it is affecting IT and business.