OpenAI Enhances AI Security Through US and UK Partnerships
OpenAI collaborates with US and UK research bodies to improve AI system security and deployment, sharing updates on joint efforts and outcomes.
OpenAI has deepened its collaboration with the US Center for AI Standards and Innovation (CAISI) and the UK AI Security Institute (UK AISI) to enhance the security of its AI systems. These partnerships focus on red-teaming, vulnerability testing, and rapid feedback loops to ensure safer AI deployment.
Key Collaborations and Findings
1. Joint Red-Teaming with CAISI
- OpenAI and CAISI conducted a groundbreaking red-teaming exercise in July, targeting the security of OpenAI’s agentic AI systems, including ChatGPT Agent.
- CAISI identified two novel vulnerabilities that could have allowed attackers to bypass security protections and impersonate users. These were fixed within one business day.
- The team used a multidisciplinary approach, combining traditional cybersecurity with AI-specific exploits, and even leveraged ChatGPT Agent to aid in vulnerability discovery. Read more about the CAISI evaluation.
2. Biosecurity Safeguards with UK AISI
- UK AISI red-teamed OpenAI’s safeguards against biological misuse, testing systems like GPT-5 and ChatGPT Agent.
- The collaboration involved non-public prototypes, custom configurations, and frequent feedback loops to strengthen safeguards.
- UK AISI uncovered over a dozen vulnerabilities, leading to improvements in OpenAI’s monitoring stack and product configurations. Explore GPT-5’s system card.
Why These Efforts Matter
- Public-Private Model: These collaborations set a precedent for how governments and industry can work together to secure frontier AI.
- Real-World Impact: The partnerships led to tangible security enhancements for widely used AI products.
- Expert Accountability: External evaluations help identify blind spots and build trust in AI systems.
Looking Ahead
OpenAI plans to continue its collaborations with CAISI and UK AISI, emphasizing the importance of ongoing evaluations over one-time checks. These efforts aim to raise the bar for AI security across the industry.
For more details on ChatGPT Agent’s safeguards, visit its .
Related News
AI Agents Fuel Identity Debt Risks Across APAC
Organizations must adopt secure authorization flows for AI environments rather than relying on outdated authentication methods to mitigate identity debt and stay ahead of attackers.
Dynamic Context Firewall Enhances AI Security for MCP
A Dynamic Context Firewall for Model Context Protocol offers adaptive security for AI agent interactions, addressing risks like data exfiltration and malicious tool execution.
About the Author

Dr. Lisa Kim
AI Ethics Researcher
Leading expert in AI ethics and responsible AI development with 13 years of research experience. Former member of Microsoft AI Ethics Committee, now provides consulting for multiple international AI governance organizations. Regularly contributes AI ethics articles to top-tier journals like Nature and Science.