If you've been following the MemSpren series, you know the core thesis: AI agents don't execute reliably because we ask them to do too much at once. For months, I've argued that the solution to non-determinism is moving orchestration out of the LLM's "vibes" and into a structured code pipeline.
This is the story of what happened when I actually tried to do that, and why it convinces me that the "AI will replace humans" narrative is fundamentally hollow.
The Stack: OpenClaw + Lobster
OpenClaw is the AI agent platform I'm building MemSpren on. Lobster is its workflow engine. The pitch is elegantly simple: define your steps in YAML, pass data between them as JSON, and let the pipeline handle the sequencing.
The division of labor is clear:
The LLM does what LLMs are good at: generating, analyzing, and transforming text.
Lobster does what code is good at: sequencing, retrying, and routing.
In theory, this eliminates the "hallucinated loop" where an agent gets stuck in a logic trap. But my use case is more than a three-step "Hello World." It involves multi-stage extraction, transformation, and writing, with structured JSON outputs at every turn. That complexity is where the friction lived.
The Myth of the "Easy" Integration
There is a massive amount of FOMO right now, especially among non-technical people, to integrate AI into every facet of their workflow. Big AI companies would have you believe this is a "plug and play" evolution.
My experience over the last week suggests the opposite. While standard, linear workflows are becoming easier to automate, the moment you attempt something creative or non-standard, the difficulty spikes exponentially. If you aren't prepared to spend hours debugging silent failures and observational mishaps, you won't get a "seamless" integration; you'll get a broken system.
When "Explicit Trust" Becomes a Wall
Throughout March 2026, OpenClaw tightened its permissions model. The releases around v2026.3.31 and v2026.4.1 hit me hard.
The update moved from implicit trust to explicit permissioning across exec, gateway auth, and node permissions. Even after I granted "Full Power" in the config, the system kept prompting for manual approval on every tool invocation. For a pipeline designed to run autonomously, this was a death sentence.
I spent hours diving through release notes and toggling flags. Eventually, I made a choice that every developer recognizes: I rolled back. I needed to execute, not audit. But that rollback triggered a cascade of secondary failures.
The Lobster Installation Workflow
The entire installation of Lobster is quite a workflow in itself. You have to install it as a plugin, you have to enable it, and you also need to install it at the global level individually for it to work. If you miss any of these steps or if it is somehow not connected to your main agent, it will simply not work.
I missed some of these instructions initially when I was reading the documentation. It took me a significant amount of time to figure out why Lobster was unable to execute from within OpenClaw as a subprocess, even though I was able to run the entire workflow directly from the terminal command line. When OpenClaw invokes Lobster, it launches the CLI in "tool mode" and expects to parse a JSON envelope from stdout. If the binary is not in the global PATH, the gateway fails silently. There are no logs in the UI and no error message in the console, just a "Step Failed" status and a void where the data should be.
Character Limits: The Silent Killer
This was the most painful bug to squash. I use llm-task directives to force the LLM into returning structured JSON. I had set a maxLength of 3,000 characters on the content field in my schema, and I had reinforced this as a critical requirement within the prompt itself.
The LLM, as LLMs do, ignored both the schema constraint and the explicit prompt instruction, generating 3,200 characters.
The LLM, as LLMs do, ignored the constraint and generated 3,200 characters.
The result? A generic 500 Internal Server Error. No partial content was returned, and there was no UI indication that the failure was a schema mismatch. I eventually found the culprit by tailing the gateway log at /tmp, filtering through the 1,440 "heartbeat" lines the gateway generates every day. Tucked away in the noise was the smoking gun:
LLM JSON did not match schema: /content must NOT have more than 3000 characters
The Anthropic Squeeze
Today, April 4, 2026, Anthropic officially ended Claude subscription support for third-party tools like OpenClaw.
For months, I've been running Opus and Sonnet through my flat-rate subscription. Now, it's pay-as-you-go or bust. I'm currently experimenting with KIMI K2.5 and OpenAI Codex. My concern isn't just price; it's reliability. Claude has been the gold standard for following complex JSON schemas. If the model layer gets flakier while I'm still fighting for determinism at the orchestration layer, I'm fighting a war on two fronts.
The New Frontier: Why AI Won't Replace You (Yet)
The pipeline finally runs. But reaching this point required a level of patience and technical forensic work that most people simply don't have the time or interest to perform.
This leads me to a few firm conclusions:
AI isn't autonomous: it's high-maintenance. The "autonomy" of these agents is an illusion that shatters the moment you step off the well-trodden path of documented examples.
The "Replacement" Narrative is Wrong. Anyone who says AI is going to replace humans has no idea what is coming at them. The complexity of making these systems actually work, not just demo well, is staggering.
The Rise of the AI Specialist. I believe we are about to see a massive surge in AI-related jobs. These are not "prompt engineers," but people who can resolve the silent failures, navigate the learning cycles, and bridge the gap between creative intent and deterministic execution.
People want to do what they are good at. They don't want to tail /tmp logs for four hours to find a schema mismatch. As long as AI remains this "tricky," there will be a massive market for humans who can tame it.
The core thesis of the MemSpren series still holds: tighter scopes and more governance is the only way forward. But I've learned that governance is a human job.
The fight for reliability continues.
If you want the personal and philosophical context behind MemSpren, I write about that on Odyssey (Substack). For the technical deep dives, follow me on dev.to.



