Analytics

5/30/2018
02:30 PM
Craig Hinkley
Craig Hinkley
Commentary
Connect Directly
Twitter
LinkedIn
RSS
E-Mail vvv
50%
50%

Machine Learning, Artificial Intelligence & the Future of Cybersecurity

The ability to learn gives security-focused AI and ML apps unrivaled speed and accuracy over their more basic, automated predecessors. But they are not a silver bullet. Yet.

Machine learning (ML) and artificial intelligence (AI) are not what most people imagine them to be. Far removed from R2-D2 or WALL-E, today's bots, sophisticated algorithms, and hyperscale computing can "learn" from past experiences to influence future outcomes.

This ability to learn gives cybersecurity-focused Al and ML applications unrivaled speed and accuracy over their more basic, automated predecessors. This might sound like the long-awaited silver bullet, but AI and ML are unlikely, at least in the near future, to deliver the much-heralded "self-healing network." The technology does, however, bring to the table a previously unavailable smart layer that forms a critical first-response defense from hackers.

The Double-Edged Sword
AI and ML would be complete game changers for cybersecurity teams if not for the fact that hackers have also embraced the technologies. This means that, although AI and ML form an increasing part of the cybersecurity solution, they more frequently contribute to the cybersecurity problem.

So, when thinking about AI and ML, it's important not to take an insular approach. Don't just focus on what your company needs in isolation. Consider what your competitors might be adopting in regard to scanning technology for locating security defects in code or vulnerabilities in production — and how you can best keep up. Think about what hackers could be deploying — and how you can counter it. Working in this way will help identify the new policies, procedures, processes, and countermeasures that must be put in place to keep your organization safe and to get the full benefit from any investment in AI and ML.

Cybersecurity Job Prospects
When the IT world first started talking about AI and ML, there was a deep-rooted concern that "the robots" would take over human jobs. In the cybersecurity sector, nothing could be further from the truth. No enterprise actually wants to give up human control of their security systems and, in fact, most organizations will need more security experts and data scientists to operate or "teach" the software.

Let's take a minute to understand why. Without human monitoring and continuous input, the current generation of AI and ML software cannot reliably learn and adapt; neither can it highlight when the data sets it relies on are becoming corrupted, question whether its conclusions are correct, or guarantee compliance. Indeed, most AI and ML projects fail when either the software hasn't been programmed to ask the right questions in order to learn, or, when trying to learn, the software is presented with flawed data. More will fail in the future if they cannot demonstrate compliance with global legislation and industry-specific regulations. 

Longer term, use of AI and ML to combat cybersecurity threats might bring about closer coordination between cybersecurity professionals and data scientists. It's not unfeasible that cybersecurity teams might recruit data scientists or that companies will begin to look for cybersecurity experts with specific data science expertise. Eventually both roles and disciplines could even merge.

So, far from discouraging graduates to study cybersecurity, AI, and data science, the growth in both technologies should encourage students to take these courses and acquire some specialization in the field. Looking broadly across the IT security sector, the current skills and knowledge gap is unlikely to go away — and, in fact, as companies struggle to understand AI on a practical level, the number of open job vacancies could increase.

Who Is in Charge?
It's important that we as humans don't lose the capacity to oversee and manage AI and ML technology — in particular, that we don't abdicate responsibility for the outcomes produced by AI and ML software. The law has some catching up to do in this regard, but we already are seeing a lot more written about AI and ML transparency, trustworthiness, and interoperability — particularly for those using AI or ML within regulated markets such as banking and insurance.

It is a brave new world out there. So, stay abreast of new AI- and ML-based cybersecurity technologies, products, and services. Some of these are going to be real industry turning points, and you don't want to be the last person finding out about them. As AI and ML begin to play even more direct and obvious roles in IT infrastructures, it's vital for cybersecurity folks to keep their knowledge current and relevant. Try to get to at least one conference a year on the topic, jump on a webinar once a quarter, read some quality independent research each month so you have a real feel for what's happening out there.

This is the next frontier, and it's time to boldly go.

Related Content:

Craig Hinkley joined WhiteHat Security as CEO in early 2015, bringing more than 20 years of executive leadership in the technology sector to this role. Craig is driving a customer-centric focus throughout the company and has broadened WhiteHat's global brand and visibility ... View Full Bio
Comment  | 
Print  | 
More Insights
Comments
Newest First  |  Oldest First  |  Threaded View
artificial intelligence
50%
50%
artificial intelligence,
User Rank: Apprentice
7/20/2018 | 1:57:05 AM
Artificial intelligence
  • Very useful blog with clear standardisation and great quality.
12 Free, Ready-to-Use Security Tools
Steve Zurier, Freelance Writer,  10/12/2018
Most IT Security Pros Want to Change Jobs
Dark Reading Staff 10/12/2018
Most Malware Arrives Via Email
Dark Reading Staff 10/11/2018
Register for Dark Reading Newsletters
White Papers
Video
Cartoon Contest
Write a Caption, Win a Starbucks Card! Click Here
Latest Comment: This comment is waiting for review by our moderators.
Current Issue
Flash Poll
The Risk Management Struggle
The Risk Management Struggle
The majority of organizations are struggling to implement a risk-based approach to security even though risk reduction has become the primary metric for measuring the effectiveness of enterprise security strategies. Read the report and get more details today!
Twitter Feed
Dark Reading - Bug Report
Bug Report
Enterprise Vulnerabilities
From DHS/US-CERT's National Vulnerability Database
CVE-2018-17534
PUBLISHED: 2018-10-15
Teltonika RUT9XX routers with firmware before 00.04.233 provide a root terminal on a serial interface without proper access control. This allows attackers with physical access to execute arbitrary commands with root privileges.
CVE-2018-17980
PUBLISHED: 2018-10-15
NoMachine before 5.3.27 and 6.x before 6.3.6 allows attackers to gain privileges via a Trojan horse wintab32.dll file located in the same directory as a .nxs file, as demonstrated by a scenario where the .nxs file and the DLL are in the current working directory, and the Trojan horse code is execute...
CVE-2018-18259
PUBLISHED: 2018-10-15
Stored XSS has been discovered in version 1.0.12 of the LUYA CMS software via /admin/api-cms-nav/create-page.
CVE-2018-18260
PUBLISHED: 2018-10-15
In the 2.4 version of Camaleon CMS, Stored XSS has been discovered. The profile image in the User settings section can be run in the update / upload area via /admin/media/upload?actions=false.
CVE-2018-17532
PUBLISHED: 2018-10-15
Teltonika RUT9XX routers with firmware before 00.04.233 are prone to multiple unauthenticated OS command injection vulnerabilities in autologin.cgi and hotspotlogin.cgi due to insufficient user input sanitization. This allows remote attackers to execute arbitrary commands with root privileges.