ChatGPT Launches Enhanced Security Features with Lockdown Mode

OpenAI has introduced two major security features for ChatGPT to combat prompt injection attacks: Lockdown Mode for high-risk users requiring enhanced protection, and "Elevated Risk" labels to help users make informed decisions about potentially vulnerable features. These protections build upon OpenAI's existing security infrastructure and will initially be available to enterprise customers before expanding to consumers.

openai Feb 13, 2026

As AI systems handle increasingly complex operations involving web connectivity and integrated applications, security considerations have evolved significantly. A particularly concerning vulnerability has gained prominence: prompt injection attacks, where malicious actors attempt to manipulate conversational AI into executing harmful commands or exposing confidential data.

OpenAI has released two significant security enhancements aimed at mitigating prompt injection risks and improving transparency around potential threats:

  • Lockdown Mode for ChatGPT - An advanced security configuration targeted at users facing elevated security threats
  • "Elevated Risk" indicators - Clear labeling for functionalities within ChatGPT, ChatGPT Atlas, and Codex that could present heightened security exposure

These measures complement OpenAI's comprehensive security framework, which incorporates sandboxing technologies, safeguards against URL-based information extraction, continuous system monitoring, and enterprise-level management features including role-based permissions and activity tracking.

Safeguarding High-Priority Personnel from Sophisticated Attacks

Lockdown Mode represents an advanced security configuration specifically engineered for select high-profile users-including organizational leaders and cybersecurity professionals at major institutions-requiring enhanced defense against sophisticated threats. This feature restricts ChatGPT's ability to communicate with external platforms, significantly reducing vulnerability to prompt injection-mediated data breaches.

The mode operates by deterministically restricting tools and features that could potentially be weaponized for unauthorized data extraction from user interactions or integrated applications through prompt injection techniques.

In this configuration, web browsing capabilities are confined to cached information exclusively, preventing live network communications beyond OpenAI's secure infrastructure. This limitation aims to block potential data transmission to attackers via browsing activities. When deterministic data protection cannot be ensured, certain functionalities are completely disabled.

ChatGPT's business offerings already include enterprise-level data protection. Lockdown Mode enhances these existing measures and is accessible through ChatGPT Enterprise, ChatGPT Edu, ChatGPT for Healthcare, and ChatGPT for Teachers. Administrators can activate this feature through Workspace Settings by establishing appropriate user roles. Once active, Lockdown Mode imposes supplementary restrictions alongside current administrative configurations.

For operational flexibility, Workspace Administrators maintain detailed control capabilities. They can specify precisely which applications-and individual functions within them-remain accessible to users operating under Lockdown Mode. Furthermore, the Compliance API Logs Platform offers comprehensive monitoring of application utilization, information sharing, and source connections, enabling administrators to maintain effective supervision.

OpenAI intends to extend Lockdown Mode availability to individual consumers in upcoming releases.

Empowering Users with Risk Transparency

AI systems demonstrate enhanced functionality when integrated with applications and web services, prompting significant investment in securing connected information. However, certain network-dependent features present emerging risks that current industry security protocols don't fully address. Some users may accept these risks, and OpenAI believes users should retain decision-making authority regarding feature utilization, particularly when handling sensitive information.

OpenAI's strategy involves providing clear guidance within products for features potentially introducing elevated risks. To improve clarity and consistency, the company is implementing standardized labeling for specific capabilities. These functionalities will display uniform "Elevated Risk" indicators throughout ChatGPT, ChatGPT Atlas, and Codex, ensuring users encounter consistent warnings across platforms.

For instance, within Codex (OpenAI's programming assistant), developers can authorize network connectivity enabling web-based actions such as documentation retrieval. The configuration interface displays the "Elevated Risk" indicator alongside comprehensive descriptions of modifications, potential vulnerabilities, and appropriate use scenarios.

Future Developments

OpenAI maintains ongoing investments in reinforcing safety and security protocols, particularly addressing emerging or expanding threats. As protective measures for these functionalities strengthen, the company will remove "Elevated Risk" designations after determining that technological improvements have adequately addressed general usage risks. The organization will continuously evaluate and update labeled features to effectively communicate potential vulnerabilities to users.