The Work Primitive: What Every AI Product Leader Gets Wrong
Most important take away
The future of AI products is not agents that click buttons better, but software that exposes semantically meaningful units of work, actions that are described, permissioned, reviewable, and reversible. When evaluating any AI product, do not just ask whether the agent can act; ask whether the product knows what that action means.
Summary
Actionable insights and career advice from this episode:
For product leaders and builders:
- Stop treating “computer use” as the destination. It is a bridge, a universal adapter for the messy middle period before software becomes agent-native. Build for the layer underneath.
- Architect agents to use the richest semantic interface available. Hierarchy of preference: typed/permissioned action > proper protocol > connector/MCP > browser > desktop control. Only fall back to UI control when richer interfaces do not exist.
- Replace the simplistic “trusted to write” framing. Design granular states: read vs write, draft vs send, stage vs deploy, recommend vs approve, sandbox vs production, refund-from-Shopify vs refund-from-Stripe. Each requires semantic awareness.
- For every AI feature, answer these product (not governance) questions: What is this object? Who owns it? What happens on success/failure? Is it reversible? Does it touch money, customer data, or production? Does it create external obligations? Can another agent review it? Can the system verify the outcome?
- Software you build in 2026 should be agent-readable from the start: tell the agent what exists, what can be done, what each action means, what permissions are required, how results are checked, and what happens next.
For startup founders (this is a free roadmap):
- The valuable, defensible problem is making the semantic meaning of work legible to agents. Do not just rely on standard MCP interfaces. Stress-test them, find where prompts cannot reliably drive levers, and fill the gap.
- Pick a lane. If you are not a hyperscaler, your lane is picked for you: start from the semantic meaning of real-world work and build back toward agents (the Perplexity approach). Hyperscalers can start from the model and code primitives and work outward.
- Coding agents arrived first not because code is text, but because codebases already have rich work semantics (modules, tests, type systems, linters, git history). That feedback loop is what made the wedge work. Look for or create similar feedback density in your domain.
For users of agents today (practical takeaway):
- Add every available connector and plugin to your ChatGPT, Codex, and Claude installations. The more semantic plugins your agent has, the less it has to fall back on slower, fuzzier computer use.
Strategic lens for the platform fight:
- The coming fight is a multi-layer negotiation: model companies want broad bridging agents, browser companies want cross-app orchestration, SaaS companies want to keep authority over domain semantics, identity providers want to govern authorization. Whoever owns the meaning of work owns durable platform power.
- Expose too little semantic access and generic agents will operate clumsily through your UI. Expose too much and you become back-end infrastructure for someone else’s interface. Salesforce (leaning in, headless-ready) is positioned better than SAP (locking agents out).
- Perplexity’s move toward Comet and a personal computer product makes sense through this lens. The browser sees cross-domain context, but a browser-only play is incomplete; reaching the file/desktop layer is necessary to hold a durable grip on the semantic meaning of work.
Career advice embedded in the piece:
- Do not build demos that “look good in a Twitter video.” Spend your time on the hard problem of making work semantically legible, because hard problems are the valuable, defensible ones.
- Train yourself to evaluate every new AI product through one question: does this give the model access, or does it give the model a meaningful set of levers it can actually use to drive the product? Use this lens to allocate your attention, your bets, and your career.
Chapter Summaries
1. The Visible Work is a Distraction. Agents clicking buttons feels like a breakthrough, but the real platform shift is happening underneath. The fight is not over who can click for you, but over who defines what the button means.
2. Three Layers: Access, Meaning, Authority. Computer use grants access. Semantic work primitives grant meaning. Authority sits deeper still. Hyperscalers understand this stack; most builders are stuck on the first layer.
3. Calendars, Checkout, and Deletion. Concrete examples show identical-looking actions can have radically different consequences. A “save” on a calendar can break commitments; a “buy” button represents money, consent, fraud risk, and disputes; deleting a file can be cleanup or destruction of the only signed copy.
4. Why Computer Use Still Matters. Most software was built for humans. Without computer use, agents cannot reach the bulk of real work. It is a necessary universal adapter, but a shallow one—it does not reveal underlying structure, reversibility, or domain meaning.
5. The Hierarchy of Meaning. Agents should always reach for the richest available interface: typed objects and permissioned actions first, protocols and connectors next, browser/desktop only as fallback. Practical takeaway: install every plugin and MCP connector available to your agent.
6. From Access to Meaning. Trusted-write framing is too crude. Real autonomy depends on semantic distinctions: draft vs send, stage vs production, refund vs refund-from-the-wrong-store. Without semantics, agents make catastrophic mistakes.
7. The Real Primitive Is a Unit of Work. Refunds, reschedules, payment authorizations, compliance exceptions, meeting briefs—these are the primitives. Human software hides them behind buttons. Agent-native software must expose them directly.
8. Why Coding Agents Arrived First. Not because code is text, but because codebases already carry rich semantics: modules, tests, dependencies, linters, git history. The environment gives agents perceptible state and feedback. Most knowledge work lacks that density.
9. Coding as Wedge, Codex as Lab. Coding is legible enough that agents can participate without constant supervision. Tools like Codex are previews of how the rest of work will eventually be structured.
10. Harnesses and the Startup Roadmap. Models alone are not products. Harnesses give agents access; semantic legibility makes the work actually doable. Making semantic meaning legible is a defensible problem space and a free roadmap for startups.
11. The Perplexity Strategy. Perplexity’s move from search to Comet to a personal computer product is a play to control the semantic meaning of work. The browser captures cross-app context; the personal computer touches files and primitives. Both are needed for durable stickiness.
12. Two Plays: Bottom-Up vs Top-Down. Hyperscalers push outward from models and code primitives. Non-hyperscalers (like Perplexity) push inward from real-world work semantics. The bridge between has gaps, and where you stand determines which gaps you must close.
13. The Coming Platform Fight. Models, browsers, SaaS, and identity providers all want to own the meaning of work. Each software company must decide how much semantic access to expose. Salesforce is leaning in; SAP is locking agents out and will lose.
14. The Final Test. The bar for software in 2026: does it tell the agent what exists, what can be done, what each action means, what permissions apply, how to verify the result, and what comes next? Stop asking whether agents can act. Ask whether the product knows what the action means.