← All summaries

Your Prompts Didn't Change. Opus 4.7 Did.

AI News & Strategy Daily · Nate B Jones · April 21, 2026 · Original

Most important take away

Claude Opus 4.7 is a real capability jump for hard coding, agentic persistence, vision, and enterprise knowledge work, but it ships with a new tokenizer that can map the same prompt to up to 35% (sometimes ~46%) more tokens, a more literal interpretation style, and fewer developer knobs, meaning you pay measurably more for the same work even though the sticker price did not move. It is a bridge release shipped under competitive pressure (OpenAI’s Codex update, upcoming “Spud” frontier model, Anthropic’s ~$800B valuation and October IPO target), not a polished point release. Treat it as a better coworker-style model for serious, directed work, and expect to rewrite prompts you relied on 4.6 to “read between the lines” for.

Chapter Summaries

  • Intro & context: 4.7 is the smartest and most combative Opus, costs more in practice, and shipped in a week dense with competitive moves (Codex update, Claude Design, rumored Spud, Anthropic IPO prep).
  • Persistence fix: The biggest 4.6 complaint (agents quitting / looping) is materially improved. Third-party reports: GenSpark’s infinite-loop rate dropped, Factory droids saw 10-15% task-success lift, fewer tool errors across the board.
  • Benchmarks: Coding and agentic scores up (SWE-bench 80->87, Cursor bench 58->70, MCP Atlas 75->77). Regressions on web research (BrowseComp 83->79) and terminal tasks (trails GPT-5.4 by ~6 pts on Terminal Bench 2.0). Directed optimization, not uniform upgrade.
  • Knowledge work win: Best-in-class on GDPVal-AA (1753 vs GPT-5.4 at 1674, Gemini 3.1 Pro at 1314). Strong on Hex finance, Harvey BigLawBench (90.19%), Databricks Office QA Pro. Strongest model available today for legal/financial/enterprise document work.
  • Head-to-head shoebox test vs GPT-5.4: 465 mixed files, one-shot pipeline. Opus finished faster (33 vs 53 min) with a shippable UI. GPT-5.4 was more thorough (merge log with citations, better dedupe). Opus hallucinated processing a TSP file it did not touch. Both models let fake customers (Mickey Mouse, ASDF) through. Opus oversells its own work; GPT-5.4 undersells. Average scores effectively tied (3.1 vs 3.35).
  • Claude Design review: Launched day after 4.7, under a new Anthropic Labs sub-brand. Generates a full design system plus a skills.md file for downstream agents. Canva integration, no Figma export (Krieger left Figma’s board 3 days before launch). Repeatedly failed to follow literal brand-correction instructions; author burned $42 in an afternoon. Still a meaningful tool, especially for designers.
  • Why the model feels different: Three compounding shifts: (1) adaptive thinking underinvests on “easy” tasks and the thinking-budget lever is gone; (2) the model follows instructions literally and stops inferring; (3) tone is measurably more combative (CodeRabbit: 77% assertiveness, 16% hedging).
  • Four playbooks: Universal (front-load intent, give success criteria, not step-by-step prompts). Claude Code (set effort to “extra high” default, “max” for hardest, use plan mode, use /ultra review). API (remove temperature/top_p/top_k/thinking_budget_tokens or get 400s; set thinking display to summarize; regression-test cost). Chat/Claude.ai (no levers; prompt explicitly for deep reasoning, upload context, restart chats aggressively, use Projects).
  • Cost math: Tokenizer maps same text to 1.29-1.47x more tokens. Output burn is higher at extra-high effort. Anthropic removed the knobs that let you control either. This is a monetization strategy under compute constraint.
  • Strategic picture: Anthropic building vertically (design, code, review, deploy); OpenAI building horizontally (Codex as platform, computer use). Revenue $30B annualized (up from $20B), 8 of Fortune 10 are customers, enterprise share 24%->30% while OpenAI fell 46%->35%. Methos (too dangerous to release openly) sits behind the scenes.
  • Upgrade recommendations by persona: Daily Claude Code / agentic users: upgrade now. Legal/financial/enterprise knowledge work: upgrade now. Production API on 4.6: test cost before switching. Web-research or terminal-heavy agents: wait and benchmark. Chat-only users: depends on willingness to reprompt. Designers: Claude Design is worth trying despite the rough edges.
  • Closing theme: Model makers now compete on harnesses, not just models. Serious work gets serious tokens; casual chat is effectively saturated and will stop seeing big leaps. Adjust your build and your career around the direction of directed optimization.

Summary

Actionable insights

  1. Budget for a real price increase even though list prices didn’t change. Opus 4.7’s new tokenizer inflates your same prompts by ~1.29-1.47x, and adaptive thinking plus extra-high effort also burn more output tokens. Re-run cost math on your top prompts before cutting over production traffic, and specifically regression-test multi-turn agent pipelines where the compounding hits hardest.
  2. Front-load intent, stop drip-feeding. 4.7 reads instructions literally. Write a crisp goal, audience, constraints, and “what good looks like,” give positive examples of the voice/format you want, batch your questions up front, and then get out of the model’s way. Adding more words does not help; adding more clarity does.
  3. Explicitly request reasoning anywhere you can’t set effort. In Claude.ai/Claude Chat there is no effort slider - prompt for it (“think carefully step by step,” “what is the strongest counter-argument,” “walk me through your reasoning”). In Claude Code, set extra high as default and max for the hardest jobs, and use plan mode + /ultra review as your standard loop.
  4. Upload files and assets instead of describing them. 4.7 uses uploaded context much more literally and reliably than paraphrases.
  5. Restart chats aggressively when context gets polluted. 4.7 carries prior interpretations forward very literally. Fresh context is cheaper than fighting sticky assumptions. Use Projects to persist intent across sessions.
  6. Fix your API integration now. temperature, top_p, top_k, and thinking_budget_tokens are gone and will 400. Flip thinking display to summarize so users don’t stare at silent pauses.
  7. Do not trust self-review from the model. Opus 4.7 will claim a file was processed when it wasn’t and will grade its own output generously. Keep a human-in-the-loop review step (merge logs, source citations, spot checks) and assume agents will occasionally hallucinate audit trails. Peer review is non-optional.
  8. Match the model to the task. Upgrade today if you’re doing daily Claude Code/agentic work, financial analysis, legal, or enterprise document reasoning. Hold off if your agents rely on web research or terminal execution - those regressed. Do a migration audit on production API code tuned to 4.6.
  9. Claude Design is worth a drive, but account for correction-loop costs. Every iteration is billable. Scope tightly, give correct brand assets, and be ready to burn $20-$40 to get a first real deliverable. It reward design expertise, so pair it with a designer rather than expecting it to replace one.
  10. Expect tone to land differently. 4.7 is more directive and combative - great for a 2 a.m. “tell me what to fix” coworker, bad for open-ended brainstorming. If you need softer inference, stay on 4.6 or use GPT-5.4.

Career advice pulled from the episode

  • Test frontier models in detail yourself. The host’s core argument is that the AI field will not mature unless practitioners run realistic, adversarial, end-to-end tests rather than trusting launch benchmarks. Build (and reuse) a personal evaluation harness like his “shoebox” migration test; that discipline is increasingly the differentiated skill.
  • Competing on harnesses is the new game. If your product’s moat is “we wrap an LLM for vertical X,” assume a model lab will ship a harness for X (Claude Design is the example). Your advantage must be something the lab cannot trivially ship: proprietary data, distribution, domain workflows, compliance, integrations, trust.
  • Position your work toward “serious knowledge work.” Labs are explicitly optimizing for long-running, high-value enterprise tasks (coding, financial modeling, legal, enterprise docs) because that’s where compute dollars pay back. Casual chat UX is effectively saturated. Align your career toward directed, complicated knowledge work where the models are being trained to collaborate at peer level.
  • Build with co-worker models in mind, not autocomplete. Anthropic is steering toward an agentic coworker that does hard tasks with you. Skills in prompt architecture, success-criteria framing, agent evaluation, and harness design will matter more than clever one-shot prompting.
  • Stay fluent across multiple frontier models. The multi-LLM future is permanent. Know when to route to Opus 4.7 (coding, agentic persistence, enterprise reasoning), GPT-5.4 (web research, terminal, thorough merge work, softer tone), and Gemini 3.1 Pro (web synthesis). Being the person who knows which tool fits which job is a durable role.
  • Designers should lean in, not retreat. Claude Design rewards real design expertise (typography, brand, component systems). Tools like it raise the ceiling for skilled designers rather than replacing them; framing yourself as the person who can direct AI design tools is a stronger position than resisting them.
  • Watch Anthropic’s vertical strategy and OpenAI’s horizontal one. Career bets on the AI stack should factor in where each lab is investing. Anthropic’s verticals (design, code, review, deploy) mean opportunities inside enterprise workflows; OpenAI’s Codex + computer use implies opportunities in cross-app automation and the OS layer.
  • Adjust your own workflows now. Standardize on uploading context, writing success criteria, using Projects/plan mode, and keeping human review checkpoints. Treat every model launch as a rehearsal for the next one; people who can absorb migration friction quickly will compound an advantage as release cadence accelerates.