AI access must favor defenders, not hackers: New framework calls for asymmetric cybersecurity strategy
The study warns that without targeted access controls, foundational AI models could be hijacked by attackers for network exploitation, malware development, and other cyber intrusions. In fact, experiments cited in the report show that scaffolding techniques, where external tools enhance AI models, have already enabled previously benign systems to perform advanced, multi-stage network attacks.

A groundbreaking study titled “Asymmetry by Design: Aligning AI Access with Cybersecurity Goals” published on arXiv, offers a detailed roadmap to counterbalance the growing risks of AI misuse in cyberspace. Authored by a team backed by the UK AI Safety Institute, the report argues for a structured system of “differential access” to advanced AI cyber capabilities - an approach designed to tilt the technological balance in favor of defenders rather than malicious attackers.
As artificial intelligence becomes increasingly embedded in both offensive and defensive cybersecurity tools, the report raises alarms about the dangerous trajectory of open and unregulated access to highly capable AI systems. Without decisive intervention, the authors warn that AI's benefits could disproportionately empower threat actors, leaving under-resourced defenders, such as schools, hospitals, and critical infrastructure providers, dangerously exposed.
What is differential access and why is it needed?
The core premise of the study is that while AI-enabled cybersecurity (referred to as AIxCyber) has transformed the ability to detect and respond to threats, this same capability poses a serious double-edged sword. Malicious actors are already leveraging AI to deploy ransomware, breach networks, and automate cyberattacks at scale. In contrast, most defenders lack the funding, technical expertise, and integration capacity to keep up with this arms race.
To address this imbalance, the authors introduce three access models:
- Promote Access: Allow low-risk AIxCyber tools to be widely disseminated for broader adoption and innovation.
- Manage Access: Provide controlled, conditional access to medium-risk capabilities for qualified defenders.
- Deny by Default: Withhold high-risk AI tools except to a vetted group of strategic actors.
Each model forms a point on a continuum, increasing in restrictiveness based on the assessed danger and misuse potential of a given AI system. Crucially, the framework centers “defender-first access” as a universal principle, ensuring that even in the most restrictive scenarios, cybersecurity professionals are not left behind.
The study warns that without targeted access controls, foundational AI models could be hijacked by attackers for network exploitation, malware development, and other cyber intrusions. In fact, experiments cited in the report show that scaffolding techniques, where external tools enhance AI models, have already enabled previously benign systems to perform advanced, multi-stage network attacks.
How should developers choose who gets access?
Beyond general access models, the study offers a six-step decision-making framework for developers and regulators to determine which organizations should receive advanced AI capabilities. These steps include:
- Assessing the model’s cyber capabilities, including downstream risks from fine-tuning or tool integration.
- Setting access goals, such as prioritizing critical infrastructure defense or encouraging rapid patching.
- Classifying defender levels, ranging from “Keystone Defenders” like major utilities and national CERTs to “Force Multipliers” such as academia and open-source communities.
- Choosing an access strategy (Promote, Manage, Deny).
- Weighing strategic and organizational considerations, such as policy alignment or national security priorities.
- Applying technical filters, like circuit breakers, refusal training, or tiered capability throttling.
This structure helps mitigate concerns about arbitrary access control and ensures that decisions are based on risk, maturity, and defensive relevance. The authors argue that differential access must be grounded in specific use-case goals, like countering ransomware in schools or securing industrial control systems in energy grids.
What are the risks and limitations of this strategy?
While the proposed framework offers clarity, the authors are careful to note its limitations. Chief among them is the cultural resistance within the cybersecurity and open-source communities, where openness is often seen as a catalyst for innovation and peer review. Differential access could, if mismanaged, result in power centralization, inequitable tool distribution, and loss of trust.
There are also technical and operational challenges: managing user identity, evaluating model misuse potential, and overseeing compliance in decentralized research environments. For instance, managing access for decentralized actors like white-hat hackers or research collectives would be significantly more complex than for institutional defenders like national security agencies.
Moreover, the threat landscape is rapidly evolving. Developers must not only anticipate potential misuse today but also forecast future risks stemming from downstream development, such as fine-tuned models escaping initial safeguards. This calls for dynamic capability evaluations and ongoing infrastructure investment in authentication, observability, and access governance.
Despite these concerns, the authors argue that the cost of inaction could be far greater. Without structured intervention, cyber defense could become an outdated concept, outpaced by AI-powered adversaries exploiting vulnerabilities faster than defenders can respond.
- FIRST PUBLISHED IN:
- Devdiscourse