Security on the path to AGI

OpenAI Blog News

Summary

OpenAI outlines comprehensive security measures on the path to AGI, including AI-powered cyber defense, continuous adversarial red teaming with SpecterOps, and security frameworks for emerging AI agents like Operator. The company emphasizes proactive threat detection, industry collaboration, and security integration into infrastructure and models.

At OpenAI, we proactively adapt, including by building comprehensive security measures directly into our infrastructure and models.
Original Article Export to Word Export to PDF
View Cached Full Text

Cached at: 04/20/26, 02:51 PM

# Security on the path to AGI Source: [https://openai.com/index/security-on-the-path-to-agi/](https://openai.com/index/security-on-the-path-to-agi/) Security threats evolve constantly and as we get closer to AGI, we expect our adversaries to become more tenacious, numerous and persistent\. At OpenAI, we proactively adapt in multiple ways, including by building comprehensive security measures directly into our[infrastructure⁠](https://openai.com/index/securing-research-infrastructure-for-advanced-ai/)and models\. **AI\-powered cyber defense**To protect our users, systems and intellectual property, we’re leveraging our own AI technology to scale our cyber defenses\. We developed advanced methods to detect cyber threats and respond rapidly\. As a supplement alongside conventional threat detection and incident response strategies, our AI\-driven security agents help enhance threat detection capabilities, enable rapid response to evolving adversarial tactics, and equip security teams with precise, actionable intelligence necessary to counter sophisticated cyberattacks\. **Continuous adversarial red teaming**We have partnered with[SpecterOps⁠\(opens in a new window\)](https://specterops.io/), renowned experts in security research and adversarial operations, to rigorously test our security defenses through realistic simulated attacks across our infrastructure, including corporate, cloud and production environments\. These continuous assessments enable us to identify vulnerabilities proactively, enhance our detection capabilities, and strengthen our response strategies against sophisticated threats\. Beyond these assessments, we are also collaborating to generate advanced skills training to improve our model capabilities into additional techniques for better protecting our products and models\. **Disrupting threat actors and proactively combating malicious AI abuse**We continuously monitor and[disrupt attempts by malicious actors to exploit our technologies⁠](https://openai.com/global-affairs/disrupting-malicious-uses-of-ai/)\. When we identify threats targeting us, such as a[recent spear phishing campaign aimed at our employees⁠\(opens in a new window\)](https://cdn.openai.com/threat-intelligence-reports/influence-and-cyber-operations-an-update_October-2024.pdf), we don’t just defend ourselves, we share tradecraft with other AI labs to strengthen our collective defenses\. By sharing these emerging risks and collaborating across industry and government, we help ensure AI technologies are developed and deployed securely\. **Securing emerging AI agents**As we introduce advanced AI agents, such as[Operator⁠](https://openai.com/index/introducing-operator/)and[deep research⁠](https://openai.com/index/introducing-deep-research/), we invest in understanding and mitigating the unique security and resilience challenges that arise with such technology\. Our efforts include developing robust alignment methods to defend against prompt injection attacks, strengthening underlying infrastructure security, and implementing agent monitoring controls to quickly detect and mitigate unintended or harmful behaviors\. As part of this, we're building a unified pipeline and modular framework to provide scalable, real\-time visibility and enforcement across agent actions and form\-factors\. **Security for future AI initiatives**Security is a cornerstone in the design and implementation of next\-generation AI projects such as[Stargate⁠](https://openai.com/index/announcing-the-stargate-project/)\. We work with our partners to adopt industry\-leading security practices such as zero\-trust architectures and hardware\-backed security solutions\. Where we are substantially expanding our physical infrastructure, we closely partner to ensure our physical safeguards evolve in tandem with our AI capabilities\. These strategies include implementing advanced access controls, comprehensive security monitoring, cryptographic protections, and defense in depth\. These practices, combined with a focus on securing software and hardware supply chains, help build foundational security from the ground up\. **Expanding our security program**We are growing our security program across several dimensions, and are looking for passionate engineers in several areas\. If you are interested in protecting OpenAI and our customers – and building the future of secure and trustworthy AI—[we’d love to hear from you⁠\(opens in a new window\)](https://jobs.ashbyhq.com/openai/form/security-blog-post-march-2025)\!

Similar Articles

Planning for AGI and beyond

OpenAI Blog

OpenAI outlines its strategy for preparing for AGI, emphasizing gradual deployment with real-world feedback loops, increasing caution as systems approach AGI capabilities, and development of better alignment techniques to ensure AI systems remain steerable and safe.

Taking a responsible path to AGI

Google DeepMind Blog

DeepMind publishes a comprehensive approach to AGI safety and security, outlining a systematic framework to address misuse, misalignment, accidents, and structural risks as artificial general intelligence approaches reality within the coming years.

Cybersecurity in the Intelligence Age

OpenAI Blog

OpenAI has published a comprehensive Action Plan aimed at democratizing AI-powered cyber defense and coordinating with government and industry to address evolving cyber threats.

Working with US CAISI and UK AISI to build more secure AI systems

OpenAI Blog

OpenAI announces collaborative security improvements with US CAISI and UK AISI, highlighting joint red-teaming efforts that discovered and helped remediate novel vulnerabilities in ChatGPT Agent systems through multidisciplinary cybersecurity and AI agent security approaches.

OpenAI safety practices

OpenAI Blog

OpenAI outlines 10 safety practices it actively uses and improves upon, including empirical red-teaming, alignment research, abuse monitoring, and voluntary commitments shared at the AI Seoul Summit. The company emphasizes a balanced, scientific approach to safety integrated into development from the outset.