Chatbot overload is real đ§ , FHIR gets (another) MCP bridge đ
The Chatbot Is the Bottleneck, Not the Model
Ethan Mollickâs latest analysis cites research showing financial professionals using GPT-4o saw real productivity gains â but the chatbot interface itself created cognitive overload that partially cancelled them out. âGiant walls of text, offers to pursue new topics, and sprawling discussionsâ hurt less experienced workers most â the exact people whoâd benefit most from AI. Mollick argues the next AI leap isnât bigger models but better delivery: specialized interfaces, familiar platforms, and context-specific tools. Claude Dispatch â send a task from your phone, get results later â is his example of âpost-chatbot AI.â
đ¤ Haters
âThis is obvious. Everyone knows chatbots arenât the final form.â Knowing it and quantifying it are different. The research shows interface overhead measurably reduces AI productivity gains. If youâre building a clinical AI tool and your delivery mechanism is a chat window, youâre leaving value on the table â and the people who need it most (residents, new nurses, rural clinicians with less support) are the ones losing out.
âPurpose-built interfaces are just more expensive to build.â They are. But the vibe coding era changes the cost equation. A clinician who understands the workflow can now build a focused interface for their specific use case faster than an enterprise vendor can ship a generic one. Thatâs the entire thesis of clinicians.build.
đĄ 80/20: If youâve built or are building a clinical AI tool delivered through a chat interface, the Mollick research suggests youâre underselling your own tool. Reframe: what would a single-purpose interface look like for your most common use case? One input, one output, no menu.
AI Products Need Failure Mode Report Cards
Automate Clinic published a concept called AI Failure Mode Literacy â the ability to understand when and how AI tools succeed or fail. Right now, developing this intuition requires what they call an âinsane amountâ of hands-on use. Their proposal: every AI product should ship with a consistently updated report card documenting its failure modes. Not just accuracy benchmarks, but contextual failure patterns.
đ¤ Haters
âVendors will never voluntarily publish their failure modes.â Probably not â at least not the incumbents. But imagine a clinical AI marketplace where the tools that publish failure data earn trust faster than the ones that donât. Transparency becomes a competitive advantage. The first vendor to ship real failure documentation in a clinical context will differentiate on something no one else is offering.
âThis is just model cards rebranded.â Model cards describe training data and performance metrics. Failure mode reports would describe contextual behavior â when does this tool break in practice, under what clinical conditions, with what patient populations? Thatâs operationally useful in a way model cards arenât.
đĄ 80/20: Next time you evaluate a clinical AI tool, ask the vendor: âWhat does this tool get wrong, and how often?â If they canât answer, thatâs your answer. Try: keep a running log of failure modes for every AI tool you use clinically â even just a shared doc. Youâre building the report card that doesnât exist yet.
â Full write-up
đ ď¸ From the Workbench
LangCare â Open Source MCP Server for FHIR + AI Agents
LangCare (formerly AgentCare) is an open-source FHIR MCP server written in Go that connects AI agents â Claude, ChatGPT, Gemini â to Epic, Cerner, and any FHIR R4 EMR. Ships with 40+ clinical agentic skills (medication management, lab interpretation, clinical decision support) and supports 150+ FHIR R4 resources. MIT licensed. Install via npm: npm install -g @langcare/langcare-mcp-fhir. Shared in the HTN community by creator Hari Kolasani â already past 1,600 installs in under three months.
â ď¸ Verify: âHIPAA-compliantâ is a vendor claim. Before routing any real patient data through this, confirm BAA availability, review the PHI scrubbing implementation, verify TLS configuration, and understand the zero-persistent-storage architecture yourself. Open source means you can audit it â which is an advantage, but also a responsibility.
đ¤ Haters
âAn npm-installed MCP server handling PHI is a security nightmare waiting to happen.â Itâs a reasonable concern. But the architecture is a stateless proxy â it translates MCP requests to FHIR calls and passes back structured responses. No persistent storage means no data at rest to breach. The real risk surface is in transit: OAuth2 token handling, TLS implementation, and whether the PHI scrubbing actually catches everything. Being open source means you can verify all of this, which is more than you can say for most commercial FHIR middleware.
â1,600 installs doesnât mean production-ready.â Correct. But 1,600 installs in under three months for a healthcare-specific MCP server means thereâs real demand for this connector layer. If youâre experimenting with AI agents in a sandbox environment with synthetic FHIR data, this is worth testing. Production with real PHI is a different conversation.
đĄ 80/20: If youâve wanted to connect Claude or another LLM to a FHIR sandbox but didnât want to build the plumbing, this is your shortcut. Try: set up LangCare against a HAPI FHIR test server with synthetic data this weekend. Zero patient risk, full agent capability.
â Full write-up
đŻ Clinician-Builder Tip of the Day
Before you build the feature, build the test. Not a unit test â a clinical scenario test. Write down five specific patient encounters where your tool should help, and five where it should stay out of the way. Run those scenarios against your prototype before you write another line of code. The five âstay out of the wayâ cases will teach you more about your toolâs actual value than the five where it shines. A tool that knows when to be quiet is more trustworthy than one that always has an answer.
What are you building this week? Reply and tell me â I read every one.
â Kevin


