NanoClaw + Docker Sandboxes: MicroVM Isolation Is Now the Default for AI Agents
Update (March 17, 2026): NVIDIA just announced NemoClaw at GTC 2026 â Jensen Huang called OpenClaw âthe operating system for personal AIâ and announced enterprise-grade security and privacy controls built on top of it.
The McKinsey Lilli breach took two hours. An AI agent found 22 unauthenticated endpoints, ran a SQL injection, and had full read-write access to 46 million messages â all without credentials.
The uncomfortable follow-up question: if youâre running AI agents that can execute code, install packages, and modify configurations, whatâs stopping a compromised or misbehaving agent from doing the same to your own systems?
NanoClaw just answered that question with a Docker partnership thatâs currently trending on Hacker News.
What NanoClaw + Docker Sandboxes actually does
NanoClaw is a lightweight open-source agent framework built around one core idea: a codebase small enough to actually audit. 15 source files. Up to 100x fewer lines of code than comparable alternatives. If you canât read and understand what your agent is doing, you canât trust it.
The Docker Sandboxes integration takes this further: every NanoClaw agent now runs inside a disposable MicroVM with full OS-level isolation from the host machine.
The practical difference from a regular Docker container matters. Containers share the host kernel â a container escape can reach your system. MicroVMs run their own lightweight kernel. The isolation is hardware-level, not just process-level. Your agent can:
- Install packages and tools freely
- Modify configurations
- Run Docker-in-Docker
- Execute any command without approval prompts
And your host machine remains completely untouched. When the task is done, the MicroVM is discarded.
The --dangerously-skip-permissions argument
This is the detail that changes the production calculus.
Coding agents typically require constant approval prompts as they move through tasks â install this package? create this file? run this script? These prompts exist because the agent is operating on your real system. Skip them and you get faster, more autonomous execution but youâre trusting the agent fully.
Inside a Docker Sandbox, --dangerously-skip-permissions becomes safe in production. The agent can move fast and operate autonomously â and if it does something destructive, the sandbox is disposable. You throw it away and start a new one. The risk is fully bounded.
This is what NanoClaw and Docker mean by âgreater autonomy without greater exposure.â
Why minimalism is a security feature
The auditable codebase argument is worth taking seriously in the context of recent events.
When CodeWallâs agent breached Lilli, one of the things it exfiltrated was McKinseyâs AI system prompts and model configurations â the behavioral instructions for the entire platform. A complex, opaque agent codebase is an attack surface in the same way: if you canât read what it does, you canât know what a compromised version of it would do.
NanoClawâs 15-file architecture is a direct answer to this. Itâs the same logic behind Agent Safehouseâs kernel-level sandboxing approach â reduce the attack surface, make the blast radius explicit and bounded.
How it fits in the isolation landscape
Weâve covered several approaches to running agents safely:
Agent Safehouse â macOS kernel sandbox (Seatbelt/sandbox-exec) for AI coding agents. OS-level syscall filtering, fine-grained filesystem permissions. Best for: Mac developers who want to constrain what a coding agent can touch without virtualization overhead.
Crust â Security gateway layer that sits between your agent and the outside world. Intercepts and filters outbound calls, enforces allow/deny policies on tool use. Best for: controlling what an agent can reach, not just what it can do locally.
OpenClaw â Full-featured personal AI agent running natively with configurable gateway-based routing and permission policies. Best for: power users who want deep capability with manual oversight control.
NanoClaw + Docker Sandboxes â MicroVM isolation, 15-file auditable codebase, disposable execution environments. Best for: enterprise teams that need to answer security audit questions, or anyone who wants to run aggressive autonomous modes safely.
These approaches arenât mutually exclusive â you can run NanoClaw inside a Docker Sandbox and still route its outbound calls through Crust. Defense in depth.
The Hacker News conversation
The HN thread is worth reading. The most interesting debate: skills-as-specs (NanoClawâs approach â ship a specification for how to write an integration, not the integration itself) vs. battle-tested shared implementations. The argument against: every user re-invents the wheel and introduces their own bugs. The argument for: âfork and customizeâ means each deployment has a different attack surface, making systematic exploitation harder.
Both sides have merit. The âheterogenous ecosystem as immune systemâ framing from one commenter is genuinely interesting â if every NanoClaw instance looks slightly different, an attack that works on one doesnât necessarily work on all of them.
The timing
NanoClaw launched this the week after the McKinsey breach went public. Thatâs not a coincidence. The market for âagent framework we can actually trust in productionâ just got a lot more urgent, and NanoClaw is explicitly pitching into that gap.
The question of how to run AI agents safely â with real isolation, real auditability, real bounded blast radius â is no longer an academic one. Itâs what every enterprise team shipping internal AI needs to answer before the next CodeWall-style disclosure lands.
Related: McKinseyâs Lilli Got Hacked in 2 Hours ¡ Agent Safehouse: Kernel-Level Sandboxing ¡ Crust: Security Gateway for AI Agents ¡ AGENTS.md Is an Attack Surface ¡ OpenClaw: Run Your Own Personal AI Assistant
Sources: Docker Blog ¡ The Register ¡ HN thread