In today’s rapidly evolving digital landscape, where software forms the backbone of our interconnected world, safeguarding critical infrastructure and software systems has become paramount. While this code facilitates modern life and drives productivity, it simultaneously creates an expanding attack surface for malicious actors.
As cyber threats continue to escalate, organizations and governments worldwide are increasingly vulnerable to malicious actors. However, the cybersecurity dilemma can be addressed through recent technological advancements. Over the past decade, we’ve witnessed the emergence of promising AI-enabled capabilities. When harnessed responsibly, this technology holds significant potential to address critical societal challenges, notably cybersecurity.
Recognizing this urgent need, the Defense Advanced Research Projects Agency (DARPA) has launched the AI Cyber Challenge (AIxCC) – a groundbreaking competition designed to harness the power of artificial intelligence (AI) and cybersecurity to enhance the security of the nation’s most critical software.
AIxCC: Where Innovation Meets Cybersecurity
At the prestigious Black Hat USA 2023 event, DARPA issued a clarion call to the brightest minds in computer science, AI, and software development to participate in the AI Cyber Challenge. This two-year competition seeks to foster innovation at the convergence of AI and cybersecurity, ushering in a new era of cybersecurity tools.
AIxCC tasks participants with the creation of innovative AI systems to fortify this critical code, offering a total of $18.5 million in prizes to the teams that deliver the most robust solutions. To nurture entrepreneurial innovation, DARPA is ready to provide funding of up to $1 million to seven small businesses, enabling their participation in the competition’s initial phase.
Why the Urgency?
In an age where software underpins every facet of our lives, from financial systems to public utilities, the potential vulnerabilities cannot be overstated. As technology drives productivity and modern life itself, it also expands the attack surface for cyber adversaries. Critical infrastructure, in particular, stands out as an attractive target for malicious cyber actors, primarily due to the scarcity of tools capable of securing these systems at scale.
Recent years have underscored the profound threats posed to society by these malicious actors. These challenges have laid bare the vast expanse that cyber defenders must shield from cyber threats. However, amidst these vulnerabilities, recent technological advancements offer a glimmer of hope.
AIxCС: Forging a Path to Cybersecurity
Currently, identifying and fixing vulnerabilities in software relies on experts who have specialized knowledge. They investigate and address these issues manually, which can be time-consuming and sometimes leads to mistakes. However, over the years, there have been developments in tools and methods for automatic vulnerability discovery and remediation (AVD&R). For instance, the use of Artificial Intelligence (AI) and Machine Learning (ML), especially Large Language Models (LLMs), shows promise in taking AVD&R to the next level. These LLMs can use neural networks and deep learning to reduce false alarms and provide more accurate tools, minimizing the need for human intervention. By combining AI and symbolic reasoning, they can learn new patterns of vulnerabilities, far surpassing current capabilities.
Moreover, they can automatically generate code fixes at scale. For example, CodePilot has demonstrated the ability to write code with minimal human input, and ChatGPT can identify and repair certain vulnerabilities effectively. These advancements are laying the foundation for innovative approaches in AVD&R, making the process more efficient. They can also improve collaboration between humans and computers in addressing software vulnerabilities, reducing the challenges currently faced in using existing tools.
AIxCC represents a pioneering collaboration led by DARPA, bringing together top AI companies. The goal is to harness AI-driven systems to tackle one of society’s most pressing challenges: cybersecurity. Over the past decade, promising AI-enabled capabilities have emerged, offering substantial potential when deployed responsibly. These technologies can play a pivotal role in addressing critical cybersecurity issues by automatically defending vital software at scale, significantly bolstering national and global cybersecurity efforts.
Collaborations
AIxCC represents a collaboration of epic proportions, uniting renowned AI companies with DARPA to provide competitors access to cutting-edge technology. Industry leaders such as Anthropic, Google, Microsoft, and OpenAI are joining forces with DARPA to empower contestants in the development of state-of-the-art cybersecurity systems.
Moreover, AIxCC has forged a close partnership with the Open Source Security Foundation (OpenSSF), a project under the Linux Foundation’s umbrella. OpenSSF will act as a guiding force for teams, aiding them in crafting AI systems capable of tackling crucial cybersecurity challenges, including the protection of critical infrastructure and securing software supply chains. Notably, most of the software, and consequently the code requiring protection, is open-source. Often crafted by community-driven volunteers, open-source software constitutes the backbone of code running on critical infrastructure across the United States, encompassing sectors like electricity and telecommunications.
AI Cyber Challenge Schedule
The AIxCC competition offers two participation tracks: the Funded Track and the Open Track. In the Funded Track, competitors will be chosen from proposals submitted to a Small Business Innovation Research solicitation, with up to seven small businesses receiving funding. On the Open Track, competitors will register directly with DARPA through the competition website and proceed without DARPA funding.
Both tracks will feature a qualifying event during the semifinal phase. The top-scoring teams (up to 20) from this phase will advance to the semifinal competition. Among these, the top performers (up to five) will receive monetary prizes and progress to the final phase. The three highest-scoring competitors in the final competition will secure additional monetary prizes.
AIxCC boasts collaboration with leading AI companies, including Anthropic, Google, Microsoft, and OpenAI, which will provide participants access to their cutting-edge technology and expertise. Additionally, the Open Source Security Foundation (OpenSSF), a Linux Foundation project, will serve as a challenge advisor. Its role is to guide teams in developing AI systems capable of addressing critical cybersecurity concerns, such as safeguarding critical infrastructure and software supply chains.
Crucially, AIxCC competitions will be hosted at DEF CON, with additional events at Black Hat USA. Both events are globally recognized cybersecurity conferences that draw tens of thousands of experts, practitioners, and observers from across the globe to Las Vegas every August. The competition will span two phases: the semifinal phase and the final phase, both taking place in Las Vegas in 2024 and 2025.
The U.S. Government’s AIxCC Finalists: Securing the Foundations of Critical Infrastructure
In a high-stakes bid to bolster cybersecurity for critical infrastructure, the U.S. Defense Advanced Research Projects Agency (DARPA) has selected seven finalist teams for the concluding phase of its AI Cyber Challenge (AIxCC). These teams, chosen from a field of 39 competitors at the DEF CON hacker conference in Las Vegas, have each received $2 million to further develop artificial intelligence-powered systems aimed at securing open-source software—a digital backbone that undergirds vital sectors like finance, water systems, and healthcare. The final showdown is set for next year’s DEF CON, where the most effective system will emerge from a year of refinement and real-world testing.
Open-source software offers speed and flexibility, making it a favored choice for developers and operators of essential services. But its public availability also makes it vulnerable. Malicious actors can easily comb through code to identify and exploit weaknesses, potentially triggering cascading failures in the event of a breach. DARPA’s initiative, launched in partnership with the Advanced Research Projects Agency for Health (ARPA-H), aims to reverse this trend by leveraging AI to harden software and detect vulnerabilities before adversaries can. As part of the contest’s rules, all finalist teams are required to open-source their cybersecurity tools, encouraging widespread adoption across the developer and security communities.
The urgency of this initiative is underscored by recent incidents. One competing group, Team Atlanta, uncovered a previously unknown bug in SQLite, a widely used database language. Other threats have emerged from deeper, more insidious campaigns: earlier this year, an entity operating under the pseudonym “Jia Tan” attempted to quietly embed a backdoor into XZ Utils, a key Linux tool used globally. Security analysts suspect the act may have been orchestrated by nation-state hackers laying groundwork for future attacks. DARPA and cybersecurity experts alike acknowledge that many of the vulnerabilities facing the U.S. are not theoretical. Chinese groups such as Volt Typhoon have already infiltrated American critical infrastructure, reportedly preparing to disrupt vital systems should geopolitical tensions escalate—especially in a Taiwan-related conflict.
The AIxCC competition goes beyond theory by inserting known and novel vulnerabilities into real open-source packages, allowing DARPA to objectively measure how well each AI system performs in detecting and neutralizing them. The use of code sanitizers, or digital tools capable of identifying specific categories of bugs, provides a structured framework for comparison. “The open-source community is incredibly impactful, but it’s not resourced at the level needed for how widespread its use has become,” said Andrew Carney, DARPA’s AIxCC program manager. The stakes are particularly high in healthcare, where small hospitals and clinics, often operating with limited technical expertise, are frequent cyberattack targets. ARPA-H director Renee Wegrzyn emphasized that scalable, AI-driven security tools could be transformative for protecting patient data and health outcomes.
The broader tech ecosystem has joined in this public-private effort. Companies like OpenAI and Anthropic provided LLM infrastructure to participants, enabling advanced automation and natural-language code analysis. Google’s Heather Adkins, VP of Security Engineering, highlighted the impracticality of removing open-source components from commercial systems altogether, noting that more than 96% of commercial codebases include such tools. The government’s push to safeguard this reality has gained momentum: the Office of the National Cyber Director released a major report last week on open-source security, and a new Department of Homeland Security (DHS) office was established to assess and protect open-source software used in critical infrastructure.
The Potential of AIxCC
In the quest for success, AIxCC has the potential to yield the next generation of cutting-edge cybersecurity tools. If AIxCC proves successful, it will not only usher in the next era of cybersecurity tools but also underscore the profound societal benefits that AI can offer by protecting our critical digital infrastructure. The competition signifies a significant step forward in safeguarding the foundations of our digital world and defending our nation’s most critical software. Beyond that, it serves as an exemplar of how AI can be harnessed for the greater good by fortifying society’s critical foundations.
As generative AI’s capabilities grow, so too does its potential for both harm and defense. DARPA Director Stefanie Tompkins expressed cautious optimism, noting that while AI’s misuse remains a concern, the AIxCC illustrates how such tools can also be harnessed for good—securing the software upon which millions rely. By mandating open access to the AI systems developed in this challenge, DARPA and ARPA-H are laying the groundwork for a cybersecurity future where collective defense is not just a goal, but a necessity.