The OpenClaw Security Wake-Up Call
Local AI agents like OpenClaw run on your workstation with filesystem access, shell privileges, and network connectivity. They execute commands, write files, and call APIs. They operate like admins with root access.
The problem? Security guidance is scattered across LLM app security, endpoint protection, and academic AI safety papers. No one owns this space. You're figuring out how to secure something that can rm -rf / prompted creatively enough.
Treat AI agents like privileged operators. Apply the same supply chain and runtime controls you'd give production servers.
Why Agents are Different (and Riskier)
Agents chain prompts, execute shell commands, parse untrusted documents, and load third-party plugins. This creates attack surfaces traditional apps don't have.
- Prompt injection: OWASP lists this in its LLM Top 10 for good reason. One malicious document can make your agent exfiltrate credentials or delete files.
- Plugin supply chain: A compromised npm package or Python module becomes a remote execution backdoor. Same risk as any dependency, now running with agent privileges.
- Memory poisoning: Corrupt the agent's long-term context and you control future behavior. Still theoretical, deeply unsettling.
Classic injection vulnerabilities: PATH manipulation, SSH command injection, template injection leading to remote code execution. We've seen all of these in production.
Real advisories prove the risk. OpenClaw published three security fixes in six months: one-click RCE via token exfiltration (GHSA-g8p2-7wf7-98mq), command injection through Docker PATH manipulation (GHSA-mc68-q9jw-2h3v), and SSH remote connection exploits (GHSA-q284-4pvr-m585).
LangChain and AutoGPT have similar records. Serialization injection leading to secret extraction (CVE-2025-68664), template injection (CVE-2025-65106), RCE via disabled block execution (CVE-2026-24780), and server-side request forgery through IPv6 parsing (CVE-2025-22603).
These are all patched. Would you have caught them before public disclosure?
If you're running a local agent, here's a security checklist you can knock out in under 20 minutes:
- Lock down inbound channels. Allowlist which messaging apps, email parsers, and document types your agent can process: for example, if it doesn't need to read PDFs from strangers, disable that capability.
- Default-deny for tools. Disable exec, browser automation, and file writes unless explicitly required. Most agents ship with these enabled by default, flip that to opt-in.
- Sandbox everything. Run sessions in isolated containers with ephemeral credentials wherever possible.
- Separate accounts. Agent credentials should never overlap with your personal access tokens. Treat them as untrusted by design.
Enterprise Controls: When You Need More Than Duct Tape
Teams running agents at scale need policy enforcement, not just best practices.
Start with policy-as-code for agent capabilities: define allowlists for commands, file paths, and API endpoints, and enforce them at runtime—not in documentation.
Then add audit logging: log every command, file write, and network call. Treat these logs like privileged access logs, because that’s what they are.
Finally, build supply chain + provenance visibility: know what’s in the runtime (dependencies), know whether vulnerabilities are actually exploitable in your setup, and know where models came from and what data trained them. If you're running a fine-tuned model from Hugging Face, you need an audit trail.
SBOM, VEX, and AIBOM: the Acronyms You Actually Need
SBOM lists every package, library, and dependency in your agent's runtime. It's the baseline for vulnerability scanning and patching. CISA has solid guidance on why this matters.
VEX tells you whether a CVE is actually exploitable in your setup. Not every vulnerability matters. VEX filters the noise. OpenVEX is the spec to watch.
AIBOM extends SBOM concepts to models and datasets. It tracks model fingerprints, training data provenance, licensing, and dataset lineage. CycloneDX already supports this.
When CVE-2026-XXXXX drops, you need answers in under ten minutes: which agents are using the affected component, whether it is actually exploitable in your environment, and what the blast radius looks like if you are wrong. SBOM, VEX, and AIBOM provide those answers with precision and speed, and spreadsheets do not.
Mistakes to Avoid
- Don’t treat agents like normal apps: They have higher privileges and broader access than typical applications.
- Don’t grant tool access by default: This is like running everything as root because it's easier. Don't.
- Don’t ignore model provenance: If you don't know where the model came from or what data trained it, you can't assess risk.
- Don’t trust sandboxes completely: Sandboxes fail. Plan for it.
- Don’t chase every CVE: Use VEX to focus on exploitable vulnerabilities, not theoretical ones.
What We're Seeing in the Wild
Security teams are tracking the same pattern families repeatedly: token theft via crafted inputs, execution hijack via environment/path behavior, lateral movement via remote command surfaces, and persistence via plugin supply chain compromise.
A representative example: token exfiltration via crafted files. The agent reads a malicious file, extracts an API token, writes it to a shared location, and the attacker uses the token remotely. No zero-days required, just broken trust assumptions.
The Bottom Line
AI agents are undiscerning privileged operators. Secure them that way.
Start with hygiene: lock channels, default-deny tools, sandbox sessions, and separate credentials. For teams, add policy-as-code, runtime monitoring, and supply chain transparency via SBOM, VEX, and AIBOM. The advisories are piling up, and the attack surface keeps growing.
An Easy Way to Get Started
See how Manifest tracks AI supply chain risk with integrated SBOM, VEX, and AIBOM workflows through Manifest AI Risk. Treat agents like privileged users. Close the gap now.



.png)

