Maybe this is a very basic question. But we know that giving local models tool call access and filesystem mounts is inherently risky — the model itself might hallucinate into a dangerous action, or get hit with a prompt injection from external content it reads. We usually just rely on the agent framework's built-in sandboxing to catch whatever slips through.
I was reading through the recent OpenClaw security audit by Ant AI Security Lab, and it got me thinking. They found that the framework's message tool could be tricked into reading arbitrary local files from the host machine by bypassing the sandbox parameter validation (reference: https://github.com/openclaw/openclaw/security/advisories/GHSA-v8wv-jg3q-qwpq).
If a framework's own parameter validation can fail like this, and a local model gets prompt-injected or goes rogue — how are you all actually securing your local agent setups?
Are you relying on strict Docker configs? Dedicated VMs? Or just trusting the framework's built-in isolation?
[link] [comments]



