OpenAI has openly acknowledged that prompt injection attacks—a sophisticated vulnerability where malicious instructions are concealed in web content, emails, or documents to manipulate AI agents—pose an intractable, long-term security threat to its ChatGPT Atlas browser, with no prospect of complete elimination.
In a comprehensive blog post published Monday titled “Continuously hardening ChatGPT Atlas against prompt injection attacks,” the company likened the issue to “ever-evolving online scams that target humans,” stating unequivocally: “Prompt injection, much like scams and social engineering on the web, is unlikely to ever be fully ‘solved.’”
The admission underscores fundamental challenges in securing “agentic” AI systems that autonomously interact with the open web. Launched on October 21, 2025, for macOS (with Windows, iOS, and Android versions forthcoming), Atlas integrates ChatGPT directly into browsing via a persistent sidebar for contextual queries, summarization, and analysis. Its standout “agent mode”—available in preview for Plus, Pro, Business, and select Enterprise/Edu users—allows the AI to perform multi-step tasks, such as navigating sites, clicking links, filling forms, managing emails, or automating workflows like meal planning and grocery ordering.
Register for Tekedia Mini-MBA edition 19 (Feb 9 – May 2, 2026): big discounts for early bird.
Tekedia AI in Business Masterclass opens registrations.
Join Tekedia Capital Syndicate and co-invest in great global startups.
Register for Tekedia AI Lab: From Technical Design to Deployment (next edition begins Jan 24 2026).
However, this autonomy dramatically “expands the security threat surface,” OpenAI noted.
Indirect prompt injections exploit the agent’s need to process untrusted content: attackers embed hidden commands (e.g., in invisible text on webpages or crafted emails) that override user intent, potentially leading to data exfiltration, unauthorized actions, or privacy breaches. Vulnerabilities surfaced immediately post-launch. Security researchers demonstrated exploits, including one where hidden text in a Google Doc altered browser behavior, and another via the Omnibox (combined address/search bar), treating malicious URLs as trusted prompts. Brave’s contemporaneous analysis deemed indirect prompt injection a “systematic challenge” for all AI browsers, citing Perplexity’s Comet as similarly affected.
Additional reports highlighted issues like “ChatGPT Tainted Memories” (a CSRF flaw that injects persistent instructions) and low phishing detection rates (Atlas blocking only ~5.8% of malicious sites versus Chrome’s 47%). Aligning with broader concerns, the U.K.’s National Cyber Security Centre (NCSC) warned earlier in December that prompt injections “may never be totally mitigated,” advising focus on risk reduction rather than eradication—echoing OpenAI’s stance. To counter this “Sisyphean” challenge, OpenAI detailed a proactive defense framework:
- LLM-based Automated Attacker: A reinforcement learning-trained bot simulates advanced hackers, iteratively crafting multi-step (tens to hundreds) attack chains in simulated environments. Leveraging access to the target’s internal reasoning traces—unavailable to external attackers—it uncovers novel exploits missed by human red teams.
- Rapid Response Loop: Discoveries feed into adversarial training for updated models, system-level safeguards, and monitoring enhancements. A recent update, triggered by automated red teaming, rolled out an adversarially trained checkpoint to all users.
- A demo illustrated the attacker’s prowess: It seeded a user’s inbox with a malicious email containing injected instructions. When the agent scanned for an out-of-office reply, it instead drafted a resignation to the CEO. Post-update, Atlas detected and flagged the attempt.
OpenAI has collaborated with third-party experts pre-launch and emphasizes user mitigations: mandatory confirmations for sensitive actions (e.g., emails, payments), “logged-out” mode to avoid credential exposure, narrow task instructions, and optional features like Browser Memories (opt-in, deletable, no training use for paid users).
Industry parallels abound. Anthropic employs similar red-teaming for Claude, while Google prioritizes architectural controls in agentic tools. Yet, experts remain cautious. Rami McCarthy, principal security researcher at Wiz, framed risk as “autonomy multiplied by access,” noting agentic browsers occupy a “challenging” high-access space.
“For most everyday use cases, agentic browsers don’t yet deliver enough value to justify their current risk profile,” McCarthy told TechCrunch, citing exposure to emails and payments.
As agentic AI proliferates, spanning OpenAI’s Atlas, Perplexity Comet, and emerging tools from Google/Anthropic, the prompt injection conundrum highlights systemic hurdles in trusting autonomous agents on the unfiltered internet.



