AI-Native Healthcare: 100M Doctor Visits, 10–20 Hours Saved, Prior Auth in Minutes — Janie Lee & Chai Asawa, Abridge
Most important take away
Abridge is evolving from an ambient AI scribe into a full “clinical intelligence layer” that sits before, during, and after the patient visit — saving doctors 10–20 hours/week of documentation while collapsing workflows like prior authorization from 45 days to minutes. The most actionable lesson for builders: in high-stakes verticals, the winning strategy is proprietary data (100M+ medical conversations), deep EHR/workflow integration, rigorous evals, and progressive rollout — not weekend prototypes.
Summary
Actionable insights and career advice from the conversation:
Product and engineering insights
- Prefer proactive intelligence over reactive alerts. Over 90% of healthcare alerts are ignored — design like “air conditioning” running in the background and only intervene when the clinical stakes truly justify breaking the doctor’s attention (e.g., prior-auth gaps before the patient leaves the room).
- Build a “constellation of models” rather than betting on one. Use cheap/fast models to triage, hand off to larger models for intelligence, and post-train on proprietary data where third-party models won’t improve (e.g., transcription, diarization, note generation).
- Treat the EHR like a file system. As frontier models become better coding agents, agentic data manipulation over EHR records will compound — invest in clean tool/context interfaces now.
- Personalize at three layers: individual (style, templates, phrasing), specialty (cardiology vs. dermatology notes), and health system (embedding hospital-specific guidelines). Memory belongs in an external store (not baked into weights) so it survives model swaps.
- Evals are operational, not just ML. Use LFD (“look at the f-ing data”), calibrated LLM judges, in-house clinician scientists, and progressive rollout — Waymo-style — to make offline distributions match real-world distributions.
- “Earn the right” with every feature. Two extra clicks kills adoption with clinicians who have 15-minute appointments. Deep EHR interoperability is table stakes.
- Prototypes are not the end-all. Janie’s contrarian take: in complex, high-stakes domains, written clarity (PRDs, strategy docs) matters more than ever because implementation, security, compliance, and edge cases never appear in a prototype. “Go slow to go fast.”
- Latency optimization is a competitive moat. Real-time agents in the conversation require model efficiency, batched triggers, and clever orchestration (Kafka, Temporal, sockets, CRDT-style conflict handling for multi-agent systems).
- The things built for humans (collaborative editing primitives, durable workflow engines) are the things that scale to agents.
Career advice
- Hire “mutants” — domain experts who are also technical (e.g., MDs who code). They raise the ceiling on evals and product quality and become disproportionately valuable as AI tooling improves.
- Don’t avoid regulated verticals. Healthcare’s high bar forces the hardest AI problems (zero-error evals, multi-step workflows) to be solved first — and regulatory tailwinds (interoperability mandates, updated FDA CDS guidance) are now favorable.
- The “vertical AI” thesis holds: narrower persona variance lets you go deeper on product and evals than horizontal players.
- Pattern-match across waves. Chai notes context engineering at Glean translated directly to Abridge — the meta-skill is recognizing that context, not raw model power, is what makes agents useful.
- Use Claude Code and similar agentic coding tools to onboard and ramp faster at new companies.
Investments / stocks mentioned
- No public-market stock tickers are recommended. Abridge is a private company; Redpoint (host) and Andreessen Horowitz are disclosed investors. The implicit investment thesis: vertical AI companies with proprietary data flywheels, deep workflow integration, and regulated-market moats are durable bets, while horizontal foundation-model territory (consumer, code) belongs to the model labs.
Chapter Summaries
- Introduction and company overview: Abridge is a clinical intelligence layer for health systems, starting with ambient documentation and expanding to before/during/after-visit workflows. Doctors spend 10–20 hours/week on documentation (“pajama time”); Abridge aims to save time, save/make money, and eventually save lives.
- Chai’s transition from Glean: Context engineering lessons translate, but healthcare has higher downside risk, narrower (vertical) variance enabling deeper product focus, and ambient-first modality as the “Jarvis” form factor.
- Alerts vs. proactive intelligence: Over 90% of healthcare alerts are ignored. Abridge prefers pre-visit prep and selective in-visit nudges (e.g., prior-authorization criteria) over interrupting clinicians.
- The prior authorization moat: Combining EHR data, payer policies (often in 50-page PDFs), and real-time conversation context to collapse 45-day workflows into minutes.
- Form factors: Mobile and desktop today; exploring in-room devices and AR glasses (already used in surgical visualization) for ambient multi-modal capture.
- Customer dynamics: Buyers (CMIOs, CFOs, CIOs) vs. users (clinicians) vs. downstream patients and payers — each requires different ROI framing.
- Hardest AI problems: Real-time prior-auth requires quality + latency + cost balance; modeling combinatorial payer/procedure policies in intermediate representations.
- Model strategy: Constellation of proprietary + third-party models. Proprietary data (~100M conversations) trains specialized models where off-the-shelf won’t improve.
- Real-time architecture: Currently batched, prototyping native real-time. Voice-in/text-out is the design choice — patients don’t want a third voice in the room.
- Personalization: Three layers (individual style, specialty, health system). Memory lives in an external store with sub-agents collating it in the background.
- Evals and rollout: LFD process, LLM judges, internal and third-party evaluators, monthly customer release cycles with pre-GA design-partner customers. Progressive rollout modeled on Waymo.
- HIPAA and de-identification: One-way de-identification models for training and evals; customer contracts govern PHI retention and access.
- Scale challenges: At 100M conversations, cost and efficiency dominate over leaderboard maxing. Post-training optimization is where ROI compounds.
- Future product surface: One conversation serves clinicians, patients, payers, pharma, and clinical trials. Latency-to-care reductions (background lab agents, real-time prior auth) are the central thesis.
- EHR partnerships: Tight integration is table stakes; Abridge positions outside core EHR scope (provider/payer/pharma intelligence layer).
- Healthcare regulation: Surprisingly favorable tailwinds — interoperability mandates, updated FDA CDS guidance — make this a special moment for AI in healthcare.
- Long-tail and team composition: Clinician scientists (“mutants”) embedded in product teams raise the eval bar; scale of offline evals catches long-tail issues pre-production.
- Glean reflections: Strong technical foundations (Google-class search) underpinned Glean’s success; the durable lesson is quality unlocks new markets.
- Durable AI infrastructure: Async/real-time systems, Kafka, Temporal, sockets, CRDT-style conflict resolution — primitives built for humans scale to agents.
- Changed minds: Janie pushes back on “prototype-first, PRDs are dead” — in high-stakes domains, written clarity is more important than ever. Chai recalibrates toward agentic models reducing need for custom scaffolding.
- Tools and closing: Claude Code and Cursor are daily drivers. Abridge publishes white papers (hallucination reduction with stats professors on staff) and hosts an upcoming AI-in-healthcare event with a16z.