Dark Reading is part of the Informa Tech Division of Informa PLC

This site is operated by a business or businesses owned by Informa PLC and all copyright resides with them.Informa PLC's registered office is 5 Howick Place, London SW1P 1WG. Registered in England and Wales. Number 8860726.

Threat Intelligence

2/13/2019
11:20 AM
Dark Reading
Dark Reading
Products and Releases
50%
50%

New Report: Toward AI Security: Global Aspirations for a More Resilient Future

Research from the Center for Long-Term Cybersecurity provides comparative analysis of different nations' strategic plans for artificial intelligence.

BERKELEY, CALIFORNIA—The Center for Long-Term Cybersecurity (CLTC), a research and collaboration hub at the University of California, Berkeley, has issued a new report that presents a novel framework for navigating the complex landscape of artificial intelligence (AI) security. The report, “Toward AI Security: Global Aspirations for a More Resilient Future,” authored by CLTC Research Fellow Jessica Cussins Newman, provides a comparative analysis of emerging AI strategies and policies from ten countries: Canada, China, France, India, Japan, Singapore, South Korea, the United Arab Emirates, the United Kingdom, and the United States.

“Artificial intelligence may be the most important global issue of the 21st century, and how we navigate the security implications of AI could dramatically shape the future,” Cussins Newman wrote in an introduction to the report. “This report uses the lens of global AI security to investigate the robustness and resiliency of AI systems, as well as the social, political, and economic systems with which AI interacts.”

Recent years have seen a significant increase in government attention to AI, as at least 27 national governments have articulated plans or initiatives for encouraging and managing the development of AI technologies. “Toward AI Security” uses a structured framework—referred to as the AI Security Map—to organize dimensions of security in which AI presents threats and opportunities, including the digital/physical, political, economic, and social domains. The map helps structure key topics that are relevant to AI security, and serves as a tool of comparing different nations’ strategic plans.

“Nations thus far have adopted highly divergent approaches in their AI policies, and there is significant variation in how they are preparing for security threats and opportunities,” Cussins Newman wrote. “For example, only half the strategies surveyed discuss the need for reliable AI systems that are robust against cyberattacks, and only two mention challenges associated with the rise of disinformation and manipulation online.” Other notable findings detailed in the report include:

·    Some governments—including those of France, India, and South Korea—are leading the way in acknowledging and preparing for the breadth of disruption likely to result from AI in the future.

·    Only two priorities are shared by all ten of the countries surveyed: promoting AI research and development, and updating training and education resources.

·    Countries have many additional opportunities to coordinate AI security strategies. For example, most countries are trying to address transparency and accountability of AI as well as privacy, data rights, and ethics. Most countries also prioritize private-public partnerships and call for improving digital infrastructure and government expertise in AI.

·    The United States and China share many priorities for advancing AI, including international collaboration; transparency and accountability; updating training and educational resources; private-public partnerships and collaboration; creating reliable AI systems; and promoting the responsible and ethical use of AI in the military.

·    Critical gaps in leadership remain around key issues. For example, only two (or fewer) national strategies address inequality, human rights, disinformation and manipulation, and checks against surveillance, control, and abuse of power.

Based on the analysis of the gaps and opportunities in national AI strategies and policies, the report provides a set of five recommendations for policymakers, including 1) facilitating early global coordination around common interests; 2) using government spending to establish best practices; 3) investigating what may be left “on the table”; 4) holding the technology industry accountable; and 5) integrating multi-disciplinary and community input.

“The steps nations take now will shape AI trajectories well into the future,” Cussins Newman wrote, “and those governments working to develop thoughtful strategies that incorporate global and multistakeholder coordination will have an advantage in establishing the international AI agenda and creating a more resilient future.”

For more information—and to download the report—visit https://cltc.berkeley.edu/TowardAISecurity.

Media Contact:
Matthew Nagamine
[email protected]
1-510-664-7506

About the UC Berkeley Center for Long-Term Cybersecurity

The Center for Long-Term Cybersecurity (CLTC) is a research and collaboration hub housed at UC Berkeley’s I School. Founded with a generous starting grant from the Hewlett Foundation in 2015, the center seeks to create effective dialogue among industry, academia, policy makers, and practitioners around a future-oriented conceptualization of cybersecurity — what it could imply and mean for human beings, machines, and the societies that will depend on both. The CLTC serves as an important resource for both students and faculty interested in cybersecurity and is committed to bringing cybersecurity practitioners and scholars to campus for an ongoing dialogue about cybersecurity. Learn more at https://cltc.berkeley.edu.

Comment  | 
Print  | 
More Insights
Comments
Newest First  |  Oldest First  |  Threaded View
COVID-19: Latest Security News & Commentary
Dark Reading Staff 5/28/2020
The Problem with Artificial Intelligence in Security
Dr. Leila Powell, Lead Security Data Scientist, Panaseer,  5/26/2020
10 iOS Security Tips to Lock Down Your iPhone
Kelly Sheridan, Staff Editor, Dark Reading,  5/22/2020
Register for Dark Reading Newsletters
White Papers
Video
Cartoon Contest
Current Issue
How Cybersecurity Incident Response Programs Work (and Why Some Don't)
This Tech Digest takes a look at the vital role cybersecurity incident response (IR) plays in managing cyber-risk within organizations. Download the Tech Digest today to find out how well-planned IR programs can detect intrusions, contain breaches, and help an organization restore normal operations.
Flash Poll
Twitter Feed
Dark Reading - Bug Report
Bug Report
Enterprise Vulnerabilities
From DHS/US-CERT's National Vulnerability Database
CVE-2020-13693
PUBLISHED: 2020-05-29
An unauthenticated privilege-escalation issue exists in the bbPress plugin before 2.6.5 for WordPress when New User Registration is enabled.
CVE-2020-13173
PUBLISHED: 2020-05-28
Initialization of the pcoip_credential_provider in Teradici PCoIP Standard Agent for Windows and PCoIP Graphics Agent for Windows versions 19.11.1 and earlier creates an insecure named pipe, which allows an attacker to intercept sensitive information or possibly elevate privileges via pre-installing...
CVE-2019-6342
PUBLISHED: 2020-05-28
An access bypass vulnerability exists when the experimental Workspaces module in Drupal 8 core is enabled. This can be mitigated by disabling the Workspaces module. It does not affect any release other than Drupal 8.7.4.
CVE-2020-11082
PUBLISHED: 2020-05-28
In Kaminari before 1.2.1, there is a vulnerability that would allow an attacker to inject arbitrary code into pages with pagination links. This has been fixed in 1.2.1.
CVE-2020-5357
PUBLISHED: 2020-05-28
Dell Dock Firmware Update Utilities for Dell Client Consumer and Commercial docking stations contain an Arbitrary File Overwrite vulnerability. The vulnerability is limited to the Dell Dock Firmware Update Utilities during the time window while being executed by an administrator. During this time wi...