AI Agents Introduce New Cybersecurity Risks

Representational image of AI

This post is also available in: עברית (Hebrew)

The emergence of AI agents – autonomous programs that can perform online tasks like booking flights or managing calendars – has created new cybersecurity concerns. Experts warn that these agents, which operate on natural language commands, could be exploited by malicious actors, opening fresh attack vectors that require little technical skill to execute.

Unlike traditional cyberattacks that demanded coding expertise, AI agents can be manipulated through carefully crafted prompts, a technique known as “query injection”. In these attacks, a seemingly benign instruction, such as “schedule a hotel reservation”, can be altered by a hacker to carry out unauthorized actions, including financial fraud. In some cases, agents encounter hidden malicious instructions embedded online, potentially executing harmful tasks without the user’s knowledge.

Cybersecurity specialists are increasingly concerned that this shift changes the landscape of digital defense. “We’re entering an era where cybersecurity is no longer about protecting users from highly skilled hackers,” noted AI startup Perplexity. Instead, the threats can now originate from AI itself if it is misdirected.

According to TechXplore, industry observers highlight the difficulty of balancing convenience with security. AI agents are designed for efficiency, but giving them unrestricted access to sensitive data or systems can magnify risk. Solutions such as real-time supervision, user approvals before critical actions, and monitoring for malicious prompts are being deployed by major AI providers. For example, Microsoft uses detection tools to flag suspicious instructions, while OpenAI restricts agent access to sensitive websites until human verification occurs.

Researchers emphasize that AI agents are not yet ready to manage critical tasks independently. According to cybersecurity experts, query injection represents the most significant vulnerability for large language models driving these systems. Similarly, they point out that attacks are rapidly evolving and can easily outpace defensive measures.

As AI agents become more widespread, the challenge for developers and users will be maintaining the convenience of autonomous tools while preventing them from being hijacked. Until security safeguards mature, experts recommend caution when assigning agents tasks that involve sensitive data or financial transactions, underscoring that AI agents, though promising, remain prone to being “taken off track.”