NanoClaw + Docker Sandboxes: MicroVM Isolation Is Now the Default for AI Agents

By Prahlad Menon 5 min read

Update (March 17, 2026): NVIDIA just announced NemoClaw at GTC 2026 — Jensen Huang called OpenClaw “the operating system for personal AI” and announced enterprise-grade security and privacy controls built on top of it.

The McKinsey Lilli breach took two hours. An AI agent found 22 unauthenticated endpoints, ran a SQL injection, and had full read-write access to 46 million messages — all without credentials.

The uncomfortable follow-up question: if you’re running AI agents that can execute code, install packages, and modify configurations, what’s stopping a compromised or misbehaving agent from doing the same to your own systems?

NanoClaw just answered that question with a Docker partnership that’s currently trending on Hacker News.

What NanoClaw + Docker Sandboxes actually does

NanoClaw is a lightweight open-source agent framework built around one core idea: a codebase small enough to actually audit. 15 source files. Up to 100x fewer lines of code than comparable alternatives. If you can’t read and understand what your agent is doing, you can’t trust it.

The Docker Sandboxes integration takes this further: every NanoClaw agent now runs inside a disposable MicroVM with full OS-level isolation from the host machine.

The practical difference from a regular Docker container matters. Containers share the host kernel — a container escape can reach your system. MicroVMs run their own lightweight kernel. The isolation is hardware-level, not just process-level. Your agent can:

  • Install packages and tools freely
  • Modify configurations
  • Run Docker-in-Docker
  • Execute any command without approval prompts

And your host machine remains completely untouched. When the task is done, the MicroVM is discarded.

The --dangerously-skip-permissions argument

This is the detail that changes the production calculus.

Coding agents typically require constant approval prompts as they move through tasks — install this package? create this file? run this script? These prompts exist because the agent is operating on your real system. Skip them and you get faster, more autonomous execution but you’re trusting the agent fully.

Inside a Docker Sandbox, --dangerously-skip-permissions becomes safe in production. The agent can move fast and operate autonomously — and if it does something destructive, the sandbox is disposable. You throw it away and start a new one. The risk is fully bounded.

This is what NanoClaw and Docker mean by “greater autonomy without greater exposure.”

Why minimalism is a security feature

The auditable codebase argument is worth taking seriously in the context of recent events.

When CodeWall’s agent breached Lilli, one of the things it exfiltrated was McKinsey’s AI system prompts and model configurations — the behavioral instructions for the entire platform. A complex, opaque agent codebase is an attack surface in the same way: if you can’t read what it does, you can’t know what a compromised version of it would do.

NanoClaw’s 15-file architecture is a direct answer to this. It’s the same logic behind Agent Safehouse’s kernel-level sandboxing approach — reduce the attack surface, make the blast radius explicit and bounded.

How it fits in the isolation landscape

We’ve covered several approaches to running agents safely:

Agent Safehouse — macOS kernel sandbox (Seatbelt/sandbox-exec) for AI coding agents. OS-level syscall filtering, fine-grained filesystem permissions. Best for: Mac developers who want to constrain what a coding agent can touch without virtualization overhead.

Crust — Security gateway layer that sits between your agent and the outside world. Intercepts and filters outbound calls, enforces allow/deny policies on tool use. Best for: controlling what an agent can reach, not just what it can do locally.

OpenClaw — Full-featured personal AI agent running natively with configurable gateway-based routing and permission policies. Best for: power users who want deep capability with manual oversight control.

NanoClaw + Docker Sandboxes — MicroVM isolation, 15-file auditable codebase, disposable execution environments. Best for: enterprise teams that need to answer security audit questions, or anyone who wants to run aggressive autonomous modes safely.

These approaches aren’t mutually exclusive — you can run NanoClaw inside a Docker Sandbox and still route its outbound calls through Crust. Defense in depth.

The Hacker News conversation

The HN thread is worth reading. The most interesting debate: skills-as-specs (NanoClaw’s approach — ship a specification for how to write an integration, not the integration itself) vs. battle-tested shared implementations. The argument against: every user re-invents the wheel and introduces their own bugs. The argument for: “fork and customize” means each deployment has a different attack surface, making systematic exploitation harder.

Both sides have merit. The “heterogenous ecosystem as immune system” framing from one commenter is genuinely interesting — if every NanoClaw instance looks slightly different, an attack that works on one doesn’t necessarily work on all of them.

The timing

NanoClaw launched this the week after the McKinsey breach went public. That’s not a coincidence. The market for “agent framework we can actually trust in production” just got a lot more urgent, and NanoClaw is explicitly pitching into that gap.

The question of how to run AI agents safely — with real isolation, real auditability, real bounded blast radius — is no longer an academic one. It’s what every enterprise team shipping internal AI needs to answer before the next CodeWall-style disclosure lands.


Related: McKinsey’s Lilli Got Hacked in 2 Hours · Agent Safehouse: Kernel-Level Sandboxing · Crust: Security Gateway for AI Agents · AGENTS.md Is an Attack Surface · OpenClaw: Run Your Own Personal AI Assistant

Sources: Docker Blog ¡ The Register ¡ HN thread