PERSONAL Sign in with your SPIE account to access your personal subscriptions or to use specific features such as save to my library, sign up for alerts, save searches, etc.
With the acceleration of Artificial Intelligence (AI) innovation in recent years, organizations across the public sector have been racing to adopt AI into their products and processes. With respect to National Security, AI adoption offers the potential to automate the operation of valuable assets and provide critical warfighter support across the observe–orient–decide–act (OODA) cycle. As we progress towards integration of AI-enabled mission critical systems, the robustness of these systems must be verified, especially against adversarial exploitation. We present a process for red-teaming AI systems used in real-world decision-making. We consider attacks on target AI systems in the context of the larger concept of operations (CONOPS) and use threat modeling to narrow down likely attack vectors.
Joshua D. Harguess andChris M. Ward
"Securing the attack surface of AI-enabled systems (Conference Presentation)", Proc. SPIE 12538, Artificial Intelligence and Machine Learning for Multi-Domain Operations Applications V, 1253805 (12 June 2023); https://doi.org/10.1117/12.2672808
ACCESS THE FULL ARTICLE
INSTITUTIONAL Select your institution to access the SPIE Digital Library.
PERSONAL Sign in with your SPIE account to access your personal subscriptions or to use specific features such as save to my library, sign up for alerts, save searches, etc.
The alert did not successfully save. Please try again later.
Joshua D. Harguess, Chris M. Ward, "Securing the attack surface of AI-enabled systems (Conference Presentation)," Proc. SPIE 12538, Artificial Intelligence and Machine Learning for Multi-Domain Operations Applications V, 1253805 (12 June 2023); https://doi.org/10.1117/12.2672808