AI Browsers Face Permanent Cyber Risk: OpenAI Warns of Unavoidable Prompt Attacks

OpenAI: OpenAI has warned that "prompt injection attacks" in AI browsers are a cyber threat that may never be completely eradicated. According to the company, these attacks are similar to scams and social engineering prevalent on the web.

Tue, 23 Dec 2025 02:24 PM (IST)
 0
AI Browsers Face Permanent Cyber Risk: OpenAI Warns of Unavoidable Prompt Attacks
AI Browsers Face Permanent Cyber Risk: OpenAI Warns of Unavoidable Prompt Attacks

OpenAI is working hard to protect its new Atlas AI browser against cyber attacks. However, the organization has recognized the bitter truth as well. According to OpenAI, prompt injection attacks, a hacking technique used on AIs, are a threat which can never completely be eliminated. This is a rather interesting comment on the security of operating AIs on the web.

OpenAI explained on their blog post that “prompt injection attacks are somewhat similar to scamming and social engineering on the web that are difficult to completely eliminate.” To put it shortly, prompt injection is a hack where hackers embed malicious intent inside web pages or emails. When an AI agent reads that page, it unknowingly follows those instructions. OpenAI has acknowledged that their browser's "Agent Mode" increases the security risk. Not only OpenAI, but Brave and the UK's National Cyber ​​Security Centre have also warned that completely preventing such attacks may never be possible.

Prompt injection attacks cannot be completely eliminated, so OpenAI is taking a different approach to manage them. The company has developed a "Large Language Model-Based Automated Attacker." This is essentially a bot that OpenAI has designed to play the role of a "hacker" through reinforcement learning (RL). This bot attacks the AI ​​agent in a simulated environment and finds new vulnerabilities. This helps OpenAI understand what the AI ​​will think and how it will react if attacked. The advantage of this is that the company can strengthen its security before real hackers attack.

OpenAI shared a demo showing how their automated attacker placed an email containing hidden instructions in a user's inbox. When the AI ​​agent scanned the inbox, it followed the hidden instructions and, instead of drafting an "out of office" reply, sent a resignation letter on the user's behalf. However, after a security update, the "Agent Mode" detected this attack and warned the user.

Advertisement

Want to get your story featured as above? click here!

Advertisement

Want to get your story featured as above? click here!

Rami McCarthy, a researcher at the cybersecurity firm Wiz, says that AI browsers pose a significant risk because they have extensive access to email, payments, and other services. However, they are not fully autonomous.

Don't allow AI to automatically send messages or make payments; require user approval for every action. Instead of giving AI instructions like "do whatever you want," assign it precise and limited tasks. Avoid giving AI full access to your sensitive accounts. McCarthy believes that for now, the risks of AI browsers for the general public outweigh their benefits, especially when it comes to personal information like email and payments.

Muskan Kumawat Journalist & Writer