AI Feature Rollouts and Data-Handling Risks in Consumer and Developer Tools
Mozilla said an upcoming Firefox release will add centralized controls to disable generative-AI capabilities, including a single “Block AI enhancements” toggle intended to prevent current and future AI features (and related prompts) from being enabled in the desktop browser. The controls are expected to allow per-feature management of AI functions such as translations, PDF image alt-text generation, AI-assisted tab grouping, link previews, and sidebar chatbot access.
Separately, OpenAI announced product changes around its developer and ChatGPT ecosystems, including a Mac-only Codex app positioned as a multi-agent “command center” with sandboxing intended to limit file writes and network access, and plans to retire GPT-4o and several other ChatGPT models as usage shifts to GPT-5.2. In parallel, a security warning highlighted a report alleging two widely used AI coding assistants were exfiltrating all ingested code to China, underscoring the need for enterprise controls over AI developer tools, data residency, and code/IP handling.
Sources
Related Stories

Firefox Adds Centralized Controls and a Kill Switch for Built-in GenAI Features
Mozilla is adding new desktop-browser controls in *Firefox 148* that let users **manage or fully disable built-in generative AI features**, including a single “**Block AI enhancements**” toggle designed to turn off current and future AI capabilities in one action. When enabled, the toggle is intended to suppress AI-related prompts and notifications and prevent exposure of AI features in the UI, addressing privacy and security concerns from users and organizations that want to reduce potential attack surface introduced by AI-driven functionality. The per-feature controls allow users to independently enable/disable specific AI functions, including **page translation**, **automatic alt-text generation for images in PDFs**, **AI-assisted tab grouping**, and **link previews** that summarize destination pages before navigation. Firefox also includes an **AI chatbot sidebar** that can integrate with multiple third-party models/services (e.g., **Anthropic Claude**, **ChatGPT**, **Microsoft Copilot**, **Google Gemini**, and others), with the centralized kill switch providing an organization-friendly way to opt out of these capabilities entirely.
1 months ago
OpenAI Adds ChatGPT Lockdown Mode and Elevated Risk Labels to Reduce Prompt-Injection Exfiltration
OpenAI introduced **Lockdown Mode** and **Elevated Risk** labels in *ChatGPT* to reduce exposure to **prompt injection** and related data-exfiltration risks when AI features interact with external systems. Lockdown Mode is positioned as an optional, advanced setting for higher-risk users and environments (notably *ChatGPT Enterprise*, *Edu*, *for Healthcare*, and *for Teachers*) that restricts tool access and limits how ChatGPT can reach outside systems; reported controls include disabling or constraining capabilities attackers could abuse via conversations or connected apps, and limiting browsing so that no live network requests leave OpenAI-controlled infrastructure (with browsing constrained to cached content). Admins can enable the setting via workspace controls and apply additional restrictions through dedicated roles, while Elevated Risk labels provide in-product warnings and guidance for features that increase risk when connecting to apps or the web, including across *ChatGPT*, *ChatGPT Atlas*, and *Codex*. Separate research highlighted how AI assistants with web-browsing/URL-fetching features can be abused as stealthy **command-and-control (C2) relays**, demonstrating a technique against **Microsoft Copilot** and **xAI Grok** that tunnels operator commands and victim data through legitimate AI web interfaces and can work without an API key or registered account. In parallel, the **European Parliament** reportedly disabled built-in AI tools on lawmakers’ work devices due to cybersecurity and privacy concerns about uploading sensitive correspondence to third-party cloud AI providers and uncertainty about what data is shared and retained. Other referenced material focused on general productivity customization of ChatGPT via “Custom Instructions,” rather than a specific security event or disclosure.
3 weeks ago
AI Safety Concerns Around Copilot and ChatGPT Content Controls
Recent reporting highlights **AI safety and governance risks** in mainstream generative AI tools, with concerns spanning both enterprise and consumer use. Gartner warned that **Microsoft 365 Copilot** can amplify existing data exposure problems by making over-shared SharePoint and Microsoft 365 content easier to discover, and also flagged the risk of users distributing inaccurate or culturally inappropriate output without proper review. The guidance emphasized enabling Microsoft’s filters, tightening document permissions, and training users to validate generated content before sharing it. Separate reporting on **OpenAI’s ChatGPT** described internal opposition to expanded “adult mode” capabilities, with former safety personnel reportedly warning that age-gating and content controls were not reliable enough to prevent minors from accessing prohibited material. The article also cited prior filter failures that allegedly allowed graphic erotic content outside intended policy boundaries. Both reports point to a broader governance issue: organizations and platform providers are struggling to keep **content moderation, access controls, and user safeguards** aligned with rapidly expanding AI functionality.
Today