How to explain OpenClaw, prompt injection, and agentic AI risk without losing the room
.png)
How to explain OpenClaw, prompt injection, and agentic AI risk without losing the room
.png)
Your team is talking about AI agents. Maybe they're already experimenting with OpenClaw, the open-source assistant that's taken the tech world by storm this past week. Maybe they're asking for budget to explore "agentic AI." Maybe someone just sent you a breathless article about the future of autonomous assistants.
You need a mental model that demystifies AI, something you can use in your next board meeting and which makes the risks intuitive without requiring a computer science degree.
Here it is.
AI agents like OpenClaw (previously ClawBot) are best understood as obedient monkeys.
A monkey is very capable since it can use tools and pull levers. It can really do things in the real world. People are using OpenClaw to negotiate car prices with dealerships, make restaurant reservations by phone, book flights, clear inboxes, and manage calendars - all autonomously and without human input.

But here's the thing about your monkey: it's not very discerning.
It's trying to listen to your instructions and do its best to get them done. But it's also willing to listen to anybody else's instructions. It doesn't have a strong sense of who should be giving it orders. If someone else tells it to do something, there's a good chance it will just go and do it.
This is the core risk with AI agents. The monkey is powerful, helpful, and eager to please - but it can't always tell the difference between you and someone pretending to be you.
When security researchers talk about "prompt injection," this is what they mean in plain terms:
Somebody else has put an instruction where they're hoping your monkey will see it. And when it sees it, there's a good chance it will just go and do whatever that instruction says.
That instruction might be hidden in an email. It might be buried in a document. It might be embedded in a webpage the agent visits on your behalf. Your monkey sees it, thinks it's a legitimate task, and acts.
.png)
The consequences depend on what your monkey has access to. If it can read your email, it can forward sensitive messages. If it can manage your calendar, it can accept meeting invites. If it has access to your bank account (and some people are giving agents exactly that kind of access) it can transfer money.
This isn't theoretical.
Security researchers have demonstrated extracting cryptocurrency private keys from compromised OpenClaw instances in under five minutes using prompt injection attacks.
To understand the risk, you need to understand what the monkey has been given access to.
OpenClaw and similar AI agents work by connecting to the tools you already use: email, calendar, Slack, WhatsApp, file systems, even your terminal. The more connections you give it, the more capable it becomes, but also the more damage it can do if compromised.
Think of each integration as a lever your monkey can pull:
The monkey doesn't distinguish between "pull this lever because my owner asked" and "pull this lever because someone in an email told me to." It just pulls levers.
.png)
AI assistants aren't new. Siri, Alexa, and Google Assistant have been around for years. But those assistants are limited by design: they can set timers and play music, but they can't take meaningful action in your digital life.
OpenClaw is different.
It's the first widely adopted AI agent that can actually do things across your most sensitive systems. It can book flights, negotiate purchases, manage your inbox, and execute code, all without you being present.
This is genuinely useful. It's also genuinely risky.
And right now, in early February 2026, we're watching what happens when a powerful new capability goes viral before the security practices catch up. In the past week alone:
The monkey is out of the cage. The question is whether your organisation is ready.
When AI agents come up in conversation, here are the questions that matter:
Map every integration. If someone in your organisation is running an AI agent, what systems does it have access to? Email? Calendar? File storage? Financial systems? The blast radius of a compromised agent equals every tool it can touch.
AI agents that connect to email, messaging apps, or the web are exposed to untrusted input. Anyone who can send your agent a message or embed instructions in a document it might read, can potentially influence its behaviour. That's a fundamentally different threat model than traditional software.
Your people are curious.
They're going to experiment whether you've figured out the guardrails or not. The question is whether they're doing it on work devices connected to corporate systems, or on isolated personal machines with no sensitive data. One of those is manageable. The other is a breach waiting to happen.
AI agents are coming. They represent a genuine leap in what's possible: autonomous assistants that can take action on your behalf, around the clock, across every system you use.
But the same capabilities that make them powerful make them dangerous when misconfigured or compromised. And right now, the technology is moving faster than the security practices.
The obedient monkey metaphor captures the essential dynamic: you have a capable, eager assistant that will do what it's told, by you or by anyone else who figures out how to talk to it.
To enable experimentation safely, we all have the responsibility to ensure the monkey is on a very short leash until we figure out how to make it safe.
If your organisation is grappling with AI agent risk, or the broader wave of agentic AI heading your way, we can help.
Metomic offers 1:1 AI Readiness Strategy Workshops (with our CTO, Ben van Enckevort) for enterprise teams, designed to help you:
The technology will mature but right now, you need a plan to experiment safely.