SAN FRANCISCO: Cybersecurity researchers are raising alarms over a new wave of threats emerging from AI agents โ advanced artificial intelligence systems designed to perform tasks online autonomously โ warning that these tools could be hijacked and exploited by hackers to carry out malicious operations.
AI agents, unlike traditional chatbots, can perform a wide range of digital tasks such as booking flights, scheduling meetings, managing emails, or even handling financial transactions โ all through natural language commands. But this same convenience, experts say, has introduced a dangerous new attack surface for cybercriminals.
โWeโre entering an era where cybersecurity is no longer about protecting users from bad actors with a highly technical skillset,โ said AI startup Perplexity in a blog post. โFor the first time in decades, weโre seeing new and novel attack vectors that can come from anywhere.โ
The rise of โprompt injectionโ attacks
The primary concern centers around injection attacks โ a decades-old hacking technique that has evolved alongside AI. In the context of AI agents, these attacks occur when malicious prompts or instructions are secretly embedded in data or online content.
When an AI agent encounters such hidden commands, it can be manipulated into performing unintended actions, from sending unauthorized payments to exfiltrating sensitive data.
Cybersecurity firm NeuralTrust engineer Marti Jorda Roca explained that the problem lies in the agentโs autonomy. โPeople need to understand there are specific dangers using AI in the security sense,โ he said, adding that the risk increases as agents are given more independence online.
Meta has publicly labeled this vulnerability a โquery injection threat,โ while OpenAIโs Chief Information Security Officer Dane Stuckey has called it โan unresolved security issue.โ Both tech giants are reportedly investing heavily to mitigate these risks as the use of AI agents expands rapidly.
When convenience meets vulnerability
Query injections can occur in multiple ways. In one example, a user command such as โbook me a hotel reservationโ could be maliciously modified into โtransfer $100 to this account.โ
In another, AI agents integrated into browsers may unknowingly interact with compromised web pages containing hidden prompts that override user intent.
Eli Smadja, cybersecurity researcher at Check Point, described these attacks as the โnumber one security problemโ facing large language models (LLMs). โOne huge mistake that I see happening a lot is to give the same AI agent all the power to do everything,โ he warned.
Industry response and growing concern
Tech firms are already deploying preventive measures. Microsoft has added tools to detect malicious commands by analyzing their origins and context, while OpenAI now alerts users when AI agents attempt to access sensitive sites, requiring human supervision for high-risk tasks.
Some experts recommend stricter oversight โ for instance, requiring explicit user approval for sensitive actions such as transferring funds or exporting data.
But cybersecurity analyst Johann Rehberger, known in the industry as โWunderwuzzi,โ cautions that current AI systems are far from trustworthy for unsupervised use.
โI donโt think we are in a position where you can have an agentic AI go off for a long time and safely do a certain task,โ he said. โIt just goes off track.โ
The growing sophistication of hackers, coupled with the accelerating deployment of autonomous AI systems, means that the next frontier of cybersecurity will depend not just on stronger code โ but on smarter, safer design principles for AI itself.





