We're giving AI agents access to production infrastructure and behaving as if we're simply releasing a new feature. I need to talk about this.
Recently, Cloudflare introduced a set of tools that allow AI agents to write code, run it, and deploy it - all on their own. There's no human involved in the process. They just announced this and the developer community seems... excited? 🤔
Why This Is Different
We have been using AI code helpers for some time now. Copilot recommends a line of code. ChatGPT writes a function. You then inspect it, test it, and deploy it on your own.
This is different. Here, the agent not only writes the code but also runs it on the production server. You are not the pilot here, you are more like a passenger who might check the flight path through the window sometimes.
What Cloudflare Actually Built
So, using these Cloudflare tools:
→ Project Think — long-running stateful AI agents that persist across sessions and maintain context over time. Not a one-shot prompt-response. A thinking entity that remembers what it's doing.
→ Dynamic Workers — AI-generated code gets executed inside sandboxed isolates. The agent writes something, and it runs. In Cloudflare's infrastructure. At the edge.
→ Codemode — instead of making individual sequential tool calls, models are encouraged to write and run code that orchestrates those predefined tools as their primary way of interacting with the world. The agent doesn't pick items from the menu one at a time. It writes a script that combines them.
Each component individually? Neat engineering. All three together? That's an autopilot deployment pipeline for inanimate software agents.
The Sandboxing Argument Doesn't Comfort Me
I can already hear the arguments: "It's all compartmentalized! Isolates are secure!"
Of course. Sandboxes are useful until they're no longer effective. Throughout the history of computing, every sandbox has been evaded, circumvented, or incorrectly configured by an exhausted engineer at 2am.
Even assuming the sandbox remains intact forever — that's not the real problem. I'm worried about what the agent decides to deploy in the first place. A sandboxed isolate that runs horrendous business logic is still horrendous business logic. It's just isolated horrendous business logic. 💀
We're Normalizing Without Discussing
What bugs me isn't the technology itself. It's how casual we are about "AI writes and ships its own code" this quickly.
We sorted deployment guardrails for decades. Code review. Staging environments. Feature flags. Canary releases. All because humans make mistakes when shipping code.
And now we're skipping most of that for a system that hallucinates confidently, calling it "developer productivity."
I'm not anti-AI. I use AI tools daily. But there's a meaningful difference between "AI helps me write code faster" and "AI writes and deploys code without me." We're blurring that line and pretending it's fine.
Where This Goes
I think we end up in one of two places:
→ Agents get real guardrails — approval workflows, automated testing gates, human checkpoints — and this becomes genuinely useful infrastructure.
→ Or we speedrun past the safety conversations because shipping fast feels too good, and we learn the hard way why those deployment ceremonies existed.
Right now, the industry seems to be sprinting toward option two. 🚀
The tooling is impressive. Cloudflare's engineering here is legitimately clever. But clever infrastructure serving an unexamined workflow is how you get elegant disasters.
Here's my question for you: At what point does "AI-assisted development" become "AI-autonomous development," and who should be drawing that line — platform providers, engineering teams, or regulators?

Top comments (0)