GPT 5.5 and GPT 5.5 Pro are not consumer upgrades. Claire Vo ran three weeks of real tests and found one clear verdict: these models belong in agentic developer workflows, not chat. A nearly six-hour autonomous run inside OpenAI Codex one-shotted 98% of edge cases across a migration of millions of chat threads in the ChatPRD codebase and dropped Sentry error rates to zero. Claude Code and GPT 5.4 both failed the same tasks.

The intelligence gap shows up in two specific places. First, reverse engineering: Vo combined a Bluetooth packet sniffer with GPT 5.5 to crack the proprietary protocol on a Divoom MiniToo pixel speaker after every other model quit. Second, long-running autonomous loops: the episode details the exact prompt pattern Vo uses to keep the model working without human checkpoints, which is the part worth reading in full. She also frames GPT 5.5 Pro's pricing explicitly against engineering time, calling it an 'intelligence tax' that is only worth paying for specific high-complexity jobs.

The practical guidance here is direct: Vo is now routing tech debt, flaky tests, and security backlogs to GPT 5.5 first. She treats GPT 5.5 as a developer model and could find no consumer use case that justified its capability ceiling. The episode also covers the /personality command inside Codex, a small but telling detail about what it actually takes to make these tools usable at scale.

[READ ORIGINAL →]