AI is quickly becoming a crucial part of innovation in sectors like healthcare, energy, cybersecurity, and defence. However, securing robust and safe AI infrastructure is a growing concern for governments and federal agencies
While AI can be considered useful for innovation and economic growth, it also brings some challenges regarding national security risks.
AI Infrastructure: Protecting national security as AI evolves
To address emerging threats, the U.S Department of Commerce recently announced the formation of the Testing Risks of AI for National Security (TRAINS) Taskforce.
This new initiative, led by the U.S. AI Safety Institute at the National Institute of Standards and Technology (NIST), brings together experts from various federal agencies to safeguard the nation’s security as AI evolves.
The TRAINS Taskforce is designed to blend research and testing efforts on advanced AI models across key national security domains.
The representation comes from the following federal agencies;
- The Department of Energy and ten of its National Laboratories
- Argonne National Laboratory
- Pacific Northwest National Laboratory
- Lawrence Livermore National Laboratory
- Sandia National Laboratory
- Oak Ridge National Laboratory
- Brookhaven National Laboratory
- Savannah River National Laboratory
- Lawrence Berkeley National Laboratory
- Idaho National Laboratory
- Los Alamos National Laboratory
- The Department of Defense, including the Chief Digital and Artificial Intelligence Office (CDAO) and the National Security Agency
- The Department of Homeland Security, including the Cybersecurity and Infrastructure Security Agency (CISA)
- The National Institutes of Health (NIH) at the Department of Health and Human Services
With AI playing an increasingly important role in military operations, infrastructure management, and even public health systems, its potential misuse by competitors sets significant risks for national security.
The TRAINS Taskforce and the roles of each federal agency
The Testing Risks of AI for National Security (TRAINS) Taskforce looks to identify, measure, and manage security risks through joint evaluations, risk assessments, and red-teaming exercises.
One of the main goals of the TRAINS Taskforce is to develop new methods for evaluating AI systems’ safety, security, and ethical implications. By collaborating across different federal agencies, each with its area of expertise, the task force can take a holistic approach to the problem.
This collaborative effort hopes to identify AI safety not just from a technological standpoint but also in terms of its broader societal and strategic impact.
The DoD and the NSA bring expertise in military applications and cybersecurity, while the DOE contributes its deep knowledge of energy infrastructure and nuclear safety.
The NIH can also provide insights into AI’s impact on healthcare and public health. This combined expertise allows the task force to conduct thorough testing of AI systems in real-world national security scenarios.
Making AI trustworthy: Robust AI infrastructure
The goal is to ensure that AI systems can be trusted in critical situations, reducing the likelihood that they could be exploited by adversaries or malfunction in ways that could cause harm.
The task force’s formation aligns with the wider national security strategy set out in the recent National Security Memorandum on AI, which builds on the key steps the President and Vice President have taken to drive the safe, secure, and trustworthy development of AI.
American leadership in controlling and utilising AI
The TRAINS Taskforce plays a key role in maintaining American leadership in AI innovation while protecting national security interests.
Overall the task force’s plans are important in terms of global competition. As other nations, particularly China and Russia, invest heavily in AI, the U.S. is making sure that its AI innovations are safe, secure, and line up with national security goals.
Not only does the TRAINS task force hope to migrate risks, but it also looks to prevent adversaries from using AI to undermine U.S. security.
By developing and testing AI systems that meet the highest safety and security standards, the task force ensures that American AI technology remains safe and can be used as a force for good.
Through their research, testing and collaboration, this new task force will create a safer, more secure AI landscape, keeping the public and national security safe.