UnitedHealth bets $3B on AI 🏥, Yuzu rebuilds insurance plumbing 🔧
The nation's largest insurer just told you where health tech is going. The question is whether clinician-builders are building the tools — or getting built around.
UnitedHealth Group Is Spending $3 Billion on AI — and Avery Is Already Live
UnitedHealth Group is deploying $3 billion into AI across its operations, making it one of the largest AI investments in healthcare to date. The company employs 22,000 software engineers globally, with over 80% now using AI to write code or build agents. The first major product: Avery, a generative AI companion for members that’s agentic, HIPAA-compliant, and already live for 6.5 million employer-sponsored plan members and 160,000 Medicare Advantage members — scaling to 20.5 million by year-end.
Avery isn’t a chatbot that answers FAQs. It learns from member interactions and integrates into insurance workflows: claims status, benefits navigation, prior authorization support. Optum Insight CEO Sandeep Dadlani framed the goal as replacing human-driven processes in claims processing, billing code selection, and fraud detection with AI-driven algorithms.
😤 Haters
“UnitedHealth automating claims decisions with AI is terrifying — they already deny too many claims.” This is the right concern. UHG has faced lawsuits over algorithmic denials before. An AI that processes claims faster isn’t inherently better for patients — it’s faster. The question is whether Avery-style tools increase transparency or just increase throughput. Clinician-builders should be watching this closely because the counter-tool — the thing that audits and challenges AI-generated denials on the provider side — doesn’t exist yet at scale.
“$3B is just a number. Every big company says they’re investing in AI.” Fair, but UHG isn’t announcing a lab or a partnership. They’re reporting that 80% of their engineers are already building with AI agents. This is operational, not aspirational. The spending is already happening.
“This is payer infrastructure — it doesn’t affect how I practice.” It does. When the largest payer automates prior auth and billing code selection, the rules your practice operates under change. The appeal process changes. The documentation requirements change. If you build tools that interact with payer systems, the API on the other side just got a lot more algorithmic.
💡 80/20: UHG is building the AI that decides what gets paid and what doesn’t. Clinician-builders have a window to build the tools that sit on the provider side — audit agents that flag questionable denials, documentation assistants tuned to payer-specific requirements, appeal generators that understand the algorithm’s logic. Try: pick one payer denial pattern you see repeatedly and sketch what an automated appeal would look like.
→ Full write-up
📡 Builder’s Radar
Yuzu Health Raises $35M to Rebuild Insurance Plumbing from Scratch
Yuzu Health raised a $35M Series A led by General Catalyst and Chemistry, with Anthropic’s Anthology Fund participating. The company is a next-generation third-party administrator (TPA) — the behind-the-scenes engine that powers claims processing, payments, and member administration for health plans. Founded in 2022, Yuzu pivoted from building a health plan to building the infrastructure that health plans run on. They’ve processed over $1B in claims across all 50 states, supporting thousands of employers.
😤 Haters
“TPAs are boring back-office plumbing — why should clinician-builders care?” Because the plumbing determines what’s possible. Every health plan innovation — new benefit designs, AI-driven care navigation, real-time eligibility checks — runs through the TPA layer. If the TPA is built on 1990s batch-processing infrastructure, nothing moves fast. Yuzu rebuilding this with modern APIs and unified data is the kind of invisible infrastructure that unlocks what clinician-builders can ship on top.
“Anthropic investing in a TPA? That’s a stretch.” It’s a signal. Anthropic’s Anthology Fund specifically backs companies where AI infrastructure creates leverage. A TPA with unified data and modern architecture is exactly the kind of system where AI agents can automate claims adjudication, stop-loss submissions, and reconciliation — all workflows that currently require humans reading faxes.
💡 80/20: If you’re building anything that touches payer workflows — eligibility checks, claims, benefits verification — watch the TPA layer. Companies like Yuzu are creating the API-first infrastructure that will make real-time payer integrations possible. Reframe: the boring plumbing is what makes the exciting tools work.
🛠️ From the Workbench
Gemma 4 Runs Locally on Your MacBook via LM Studio
Google’s Gemma 4 is a 26B-parameter open model (Apache 2.0 license) that activates only 4B parameters at inference time, making it runnable on consumer hardware. LM Studio 0.4.0 supports it natively — reports of 51 words/sec on a MacBook. It supports tool use, vision input, and a 256K context window. Not medical-specific, but the Apache 2.0 license and local execution make it interesting for clinician-builders who need to experiment with patient-adjacent workflows without sending data to cloud APIs.
⚠️ Verify: Gemma 4 is a general-purpose model, not fine-tuned for clinical use. Do not use for clinical decision-making without thorough evaluation against your specific use case. Local execution reduces data exposure but does not eliminate risk — validate outputs against clinical ground truth before any patient-facing deployment.
😤 Haters
“A general model isn’t useful for clinical work.” Not directly. But a 26B model running locally with tool use and 256K context is a prototyping sandbox. You can test clinical NLP workflows, experiment with structured data extraction from notes, and iterate on prompts — all without an API bill or a data processing agreement.
“51 words/sec is too slow for production.” It’s not for production. It’s for Saturday morning prototyping. The value is zero-cost local iteration before you commit to a cloud API for the real thing.
💡 80/20: If you run Ollama or LM Studio, pull Gemma 4 this week. Use it as a local testing sandbox for clinical NLP experiments — medication extraction, note summarization, structured data parsing. The Apache 2.0 license means you can fine-tune it later if something clicks. If you’re not running local models, you are missing a huge privacy focused aspect of AI building.
What are you building this week? Reply and tell me — I read every one.
— Kevin


