Developing AI, Autonomy Applications Warfighters Can Trust by David Vergun, DOD News April 1, 2024 An important goal of the Defense Advanced Research Projects Agency (DARPA) is developing artificial intelligence that is trustworthy for the Defense Department ... particularly for making life-or-death recommendations to warfighters, said Matt Turek, deputy director of DARPA's Information Innovation Office. Top - An XQ-58A Valkyrie launches for a test mission at Eglin Air Force Base, Florida on August 22, 2023. Artificial intelligence algorithms, developed and trained by the Air Force Research Laboratory’s Autonomous Air Combat Operations, were integrated into the XQ-58A. The algorithms used neural networks to fly the vehicle against simulated opponents using simulated mission systems and mock weapons. Bottom - An unmanned aerial system hovers in the sky above the Eglin Air Force Base, Florida on October 26, 2023.
(Image created by USA Patriotism! from U.S. Air Force photos by 2nd Lt. Rebecca Abordo and Samuel King Jr.) |
AI, machine learning and autonomy are being used by about 70% of DARPA's programs in some form or another, Turek said today at a Center for Strategic and International Studies event.
Another reason AI development is such a priority is to prevent an unexpected breakthrough in technology, or "strategic surprise," by adversaries who might also be developing advanced capabilities, he said, adding that DARPA also aims to create its own strategic surprise.
To accomplish those goals, DARPA is looking for transformative capabilities and ideas from industry and academia, Turek said.
One of the many ways the agency gets these capabilities and ideas is to hold various types of challenges where teams from the private sector can win prizes worth millions of dollars, he said.
An example of that, he said, is DARPA's Artificial Intelligence Cyber Challenge, which uses generative AI technologies ... like large language models ... to automatically find and fix vulnerabilities in open-source software, particularly software that underlies critical infrastructure. Large language models involve processing and manipulating human language to perform such tasks as secure computer coding, decision-making, speech recognition and making predictions.
Turek said a unique feature of this challenge is the partnership between DARPA and state-of-the-art large language model providers that are participating in the challenges, including Google, Microsoft, OpenAI and Anthropic.
Most likely, large language model improvements will also benefit the commercial sector, as well as DOD, Turek said.
An example of the use of autonomy and of AI that DARPA has been testing with the Air Force involves its F-16 fighter jets, he said. Turek said DARPA has four areas of AI research involving industry and academia partners: Proficient artificial intelligence Confidence in the information domain, which includes tools that detect things like manipulated media Secure and resilient systems Defensive and offensive cyber tools.
Turek noted that there's a lot of synergy across those four areas. DARPA | U.S. Department of Defense Our Valiant Troops | I Am The One | Uncommon Valor | Veterans | Citizens Like Us | Spouses Serve Too |
|