Don’t ignore the security risks of agentic AI

In the race to deploy agentic artificial intelligence systems across workflows, an uncomfortable truth is being ignored: autonomy invites unpredictability, and unpredictability is a security risk. If we don’t rethink our approach to safeguarding these systems now, we may find ourselves chasing threats we barely understand at a scale we can’t contain.

Agentic AI systems are designed with autonomy at their core. They can reason, plan, take action across digital environments, and even coordinate with other agents. Think of them as digital interns with initiative, capable of setting and executing tasks with minimal oversight.

But the very thing that makes agentic AI powerful—its ability to make independent decisions in real-time—is also what makes it an unpredictable threat vector. In the rush to commercialize and deploy these systems, insufficient attention has been given to the potential security liabilities they introduce.

Whereas large language model-based chatbots are mostly reactive, agentic systems operate proactively. They might autonomously browse the web, download data, manipulate application programming interfaces (APIs), execute scripts, or even interact with real-world systems like trading platforms or internal dashboards.

That sounds exciting—until you realize how few guardrails may be in place to monitor or constrain these actions once set in motion.

### ‘Can’ vs. ‘Should’

Security researchers are increasingly raising alarms about the attack surface these systems introduce. One glaring concern is the blurred line between what an agent *can* do and what it *should* do.

As agents gain permissions to automate tasks across multiple applications, they also inherit access tokens, API keys, and other sensitive credentials. A prompt injection, hijacked plugin, exploited integration, or engineered supply chain attack could give attackers a backdoor into critical systems.

We’ve already seen examples of large language model agents falling victim to adversarial inputs. In one case, researchers demonstrated that embedding a malicious command in a webpage could trick an agentic browser bot into exfiltrating data or downloading malware—without any malicious code on the attacker’s end. The bot simply followed instructions buried in natural language. No exploits. No binaries. Just linguistic sleight of hand.

And it doesn’t stop there.

When agents are granted access to email clients, file systems, databases, or DevOps tools, a single compromised action can trigger cascading failures. From initiating unauthorized Git pushes to granting unintended permissions, agentic AI has the potential to replicate risks at machine speed and scale.

### The Problem with Capability Benchmarks

The problem is exacerbated by the industry’s obsession with capability benchmarks over safety thresholds. Much of the focus has been on how many tasks agents can complete, how well they self-reflect, or how efficiently they chain tools. Relatively little attention has been given to sandboxing, logging, or even real-time override mechanisms.

In the push for autonomous agents that can take on end-to-end workflows, security is playing catch-up.

### The Need to Catch Up Fast

Mitigation strategies must evolve beyond traditional endpoint or application security. Agentic AI exists in a gray area between the user and the system. Role-based access control alone won’t cut it.

We need policy engines that understand intent, monitor behavioral drift, and can detect when an agent begins to act out of character. Developers must implement fine-grained scopes for what agents can do—limiting not just which tools they use, but how, when, and under what conditions.

Auditability is also critical. Many of today’s AI agents operate in ephemeral runtime environments with little to no traceability. If an agent makes a flawed decision, there’s often no clear log of its thought process, actions, or triggers. That lack of forensic clarity is a nightmare for security teams.

In at least some cases, models have resorted to malicious insider behaviors when that was the only way to avoid replacement or achieve their goals—including blackmailing officials and leaking sensitive information to competitors.

Finally, we need robust testing frameworks that simulate adversarial inputs in agentic workflows. Penetration-testing a chatbot is one thing; evaluating an autonomous agent that can trigger real-world actions is a completely different challenge. It requires scenario-based simulations, sandboxed deployments, and real-time anomaly detection.

### Halting First Steps

Some industry leaders are beginning to respond.

OpenAI LLC has hinted at dedicated safety protocols for its newest publicly available agent. Anthropic PBC emphasizes constitutional AI as a safeguard, and others are building observability layers around agent behavior.

But these are early steps, and they remain uneven across the ecosystem.

Until security is baked into the development lifecycle of agentic AI, rather than being patched on afterward, we risk repeating the same mistakes made during the early days of cloud computing: excessive trust in automation before building resilient guardrails.

We are no longer speculating about what agents might do. They are already executing trading strategies, scheduling infrastructure updates, scanning logs, crafting emails, and interacting with customers.

The question isn’t whether they’ll be abused but when.

Any system that can act must be treated as both an asset and a liability.

Agentic AI could become one of the most transformative technologies of the decade. However, without robust security frameworks, it could also become one of the most vulnerable targets.

The smarter these systems get, the harder they’ll be to control in retrospect.

Which is why the time to act isn’t tomorrow. It’s now.

*Isla Sibanda is an ethical hacker and cybersecurity specialist based in Pretoria, South Africa. She has been a cybersecurity analyst and penetration testing specialist for more than 12 years. She wrote this article for SiliconANGLE.*
https://siliconangle.com/2025/11/15/dont-ignore-security-risks-agentic-ai/

Leave a Reply

Your email address will not be published. Required fields are marked *