Poor Man's Prompt

Good week for OpenAI. GPT 5.2 Codex is 40% faster than last week. They’re offering 2x rate limits if you use the macOS desktop app. I gave it three days, basically full time.
However.
A colleague switched from Claude Code to the Codex app because of the higher limits. Within a week he messaged me: “Codex is an absolute pain to talk to compared to Claude Code.” It keeps questioning his decisions. Claude Code just gets him.
I think the Codex Desktop app is basically a poor man’s Claude Code prompt.
Here’s what I mean. The Codex app lets you click to start new conversations that spin up fresh worktrees. At first that seems elegant - each ticket gets its own isolated environment, everything wired up automatically.
Until you realise the models have no clue what workflow they’re in.
The context is implicit in the harness code. The model can’t see it. So you keep having to explain what’s actually happening, which comes as a surprise to them. You’re educating the models about their own environment.
A harness is great for improving today’s workflow. But it’s behavior you can encode through natural language instead. And when you do, today’s models can actually reason about it. Notice conflicts. Adapt to edge cases. This wasn’t true a few months ago - we’ve reached a level of reliability, particularly with Opus 4.5, where you can encode an awful lot through natural language and expect it to be adhered to. A prompt is visible to the thing doing the work. A GUI isn’t.
So I’m back to using Claude Code as my concierge with an orchestrator prompt - it manages worktrees and coordinates headless Codex instances running in the background. Better at talking anyway. More flexible than hard-coded buttons. With the caveat that none of my workflows seem to last very long given the rate of progress. The best workflow tracks the frontier. The training data is always behind the frontier. So your workflow is out of distribution by construction, not by accident.
And in the meantime, I’m conscious that the bitter lesson is coming for all of this. Models will absorb these workflows. I’d be loath to become reliant on scaffolding I don’t control when it’s only a matter of time till the models consume the majority of it anyway.

Nothing against OpenAI here. The problem is putting intelligence in code the model can’t see.