← All summaries

AI-Native Healthcare: 100M Doctor Visits, 10–20 Hours Saved, Prior Auth in Minutes — Janie Lee & Chai Asawa, Abridge

Latent Space · Latent.Space — Janie Lee, Chai Asawa · May 14, 2026 · Original

Most important take away

Abridge is evolving from an ambient AI scribe into a full “clinical intelligence layer” that sits before, during, and after the patient visit — saving doctors 10–20 hours/week of documentation while collapsing workflows like prior authorization from 45 days to minutes. The most actionable lesson for builders: in high-stakes verticals, the winning strategy is proprietary data (100M+ medical conversations), deep EHR/workflow integration, rigorous evals, and progressive rollout — not weekend prototypes.

Summary

Actionable insights and career advice from the conversation:

Product and engineering insights

  • Prefer proactive intelligence over reactive alerts. Over 90% of healthcare alerts are ignored — design like “air conditioning” running in the background and only intervene when the clinical stakes truly justify breaking the doctor’s attention (e.g., prior-auth gaps before the patient leaves the room).
  • Build a “constellation of models” rather than betting on one. Use cheap/fast models to triage, hand off to larger models for intelligence, and post-train on proprietary data where third-party models won’t improve (e.g., transcription, diarization, note generation).
  • Treat the EHR like a file system. As frontier models become better coding agents, agentic data manipulation over EHR records will compound — invest in clean tool/context interfaces now.
  • Personalize at three layers: individual (style, templates, phrasing), specialty (cardiology vs. dermatology notes), and health system (embedding hospital-specific guidelines). Memory belongs in an external store (not baked into weights) so it survives model swaps.
  • Evals are operational, not just ML. Use LFD (“look at the f-ing data”), calibrated LLM judges, in-house clinician scientists, and progressive rollout — Waymo-style — to make offline distributions match real-world distributions.
  • “Earn the right” with every feature. Two extra clicks kills adoption with clinicians who have 15-minute appointments. Deep EHR interoperability is table stakes.
  • Prototypes are not the end-all. Janie’s contrarian take: in complex, high-stakes domains, written clarity (PRDs, strategy docs) matters more than ever because implementation, security, compliance, and edge cases never appear in a prototype. “Go slow to go fast.”
  • Latency optimization is a competitive moat. Real-time agents in the conversation require model efficiency, batched triggers, and clever orchestration (Kafka, Temporal, sockets, CRDT-style conflict handling for multi-agent systems).
  • The things built for humans (collaborative editing primitives, durable workflow engines) are the things that scale to agents.

Career advice

  • Hire “mutants” — domain experts who are also technical (e.g., MDs who code). They raise the ceiling on evals and product quality and become disproportionately valuable as AI tooling improves.
  • Don’t avoid regulated verticals. Healthcare’s high bar forces the hardest AI problems (zero-error evals, multi-step workflows) to be solved first — and regulatory tailwinds (interoperability mandates, updated FDA CDS guidance) are now favorable.
  • The “vertical AI” thesis holds: narrower persona variance lets you go deeper on product and evals than horizontal players.
  • Pattern-match across waves. Chai notes context engineering at Glean translated directly to Abridge — the meta-skill is recognizing that context, not raw model power, is what makes agents useful.
  • Use Claude Code and similar agentic coding tools to onboard and ramp faster at new companies.

Investments / stocks mentioned

  • No public-market stock tickers are recommended. Abridge is a private company; Redpoint (host) and Andreessen Horowitz are disclosed investors. The implicit investment thesis: vertical AI companies with proprietary data flywheels, deep workflow integration, and regulated-market moats are durable bets, while horizontal foundation-model territory (consumer, code) belongs to the model labs.

Chapter Summaries

  • Introduction and company overview: Abridge is a clinical intelligence layer for health systems, starting with ambient documentation and expanding to before/during/after-visit workflows. Doctors spend 10–20 hours/week on documentation (“pajama time”); Abridge aims to save time, save/make money, and eventually save lives.
  • Chai’s transition from Glean: Context engineering lessons translate, but healthcare has higher downside risk, narrower (vertical) variance enabling deeper product focus, and ambient-first modality as the “Jarvis” form factor.
  • Alerts vs. proactive intelligence: Over 90% of healthcare alerts are ignored. Abridge prefers pre-visit prep and selective in-visit nudges (e.g., prior-authorization criteria) over interrupting clinicians.
  • The prior authorization moat: Combining EHR data, payer policies (often in 50-page PDFs), and real-time conversation context to collapse 45-day workflows into minutes.
  • Form factors: Mobile and desktop today; exploring in-room devices and AR glasses (already used in surgical visualization) for ambient multi-modal capture.
  • Customer dynamics: Buyers (CMIOs, CFOs, CIOs) vs. users (clinicians) vs. downstream patients and payers — each requires different ROI framing.
  • Hardest AI problems: Real-time prior-auth requires quality + latency + cost balance; modeling combinatorial payer/procedure policies in intermediate representations.
  • Model strategy: Constellation of proprietary + third-party models. Proprietary data (~100M conversations) trains specialized models where off-the-shelf won’t improve.
  • Real-time architecture: Currently batched, prototyping native real-time. Voice-in/text-out is the design choice — patients don’t want a third voice in the room.
  • Personalization: Three layers (individual style, specialty, health system). Memory lives in an external store with sub-agents collating it in the background.
  • Evals and rollout: LFD process, LLM judges, internal and third-party evaluators, monthly customer release cycles with pre-GA design-partner customers. Progressive rollout modeled on Waymo.
  • HIPAA and de-identification: One-way de-identification models for training and evals; customer contracts govern PHI retention and access.
  • Scale challenges: At 100M conversations, cost and efficiency dominate over leaderboard maxing. Post-training optimization is where ROI compounds.
  • Future product surface: One conversation serves clinicians, patients, payers, pharma, and clinical trials. Latency-to-care reductions (background lab agents, real-time prior auth) are the central thesis.
  • EHR partnerships: Tight integration is table stakes; Abridge positions outside core EHR scope (provider/payer/pharma intelligence layer).
  • Healthcare regulation: Surprisingly favorable tailwinds — interoperability mandates, updated FDA CDS guidance — make this a special moment for AI in healthcare.
  • Long-tail and team composition: Clinician scientists (“mutants”) embedded in product teams raise the eval bar; scale of offline evals catches long-tail issues pre-production.
  • Glean reflections: Strong technical foundations (Google-class search) underpinned Glean’s success; the durable lesson is quality unlocks new markets.
  • Durable AI infrastructure: Async/real-time systems, Kafka, Temporal, sockets, CRDT-style conflict resolution — primitives built for humans scale to agents.
  • Changed minds: Janie pushes back on “prototype-first, PRDs are dead” — in high-stakes domains, written clarity is more important than ever. Chai recalibrates toward agentic models reducing need for custom scaffolding.
  • Tools and closing: Claude Code and Cursor are daily drivers. Abridge publishes white papers (hallucination reduction with stats professors on staff) and hosts an upcoming AI-in-healthcare event with a16z.