Why Most Physician-Built AI Tools Will Fail (And How to Build the Ones That Won’t)
5 min read

Why Most Physician-Built AI Tools Will Fail (And How to Build the Ones That Won’t)

Doctors Who Code

The adoption crisis in clinical AI isn’t technical — it’s generational. Here’s what 18 months building CodeCraftMD taught me about who actually uses physician-built tools.

By Chukwuma Onyeije, MD, FACOG
Medical Director, Atlanta Perinatal Associates | Founder, CodeCraftMD
Building HIPAA-compliant AI tools for real clinical workflows


I spent last Tuesday watching a third-year resident use ChatGPT to draft her clinic notes while her attending physician — fifteen years into practice — manually transcribed the same encounter into Epic.

Both physicians. Same patient. Same documentation burden.

Completely different approaches to solving it.

This moment crystallized something I’ve observed while building CodeCraftMD over the past 18 months: The question isn’t whether physicians will adopt AI tools. It’s which physicians will adopt them first — and why that matters for everyone else.

The Tale of Two Clinical Cultures

Building a medical billing automation platform has given me a front-row seat to how different generations of physicians approach new technology. What I’ve found contradicts most vendor assumptions about clinical AI adoption.

Clinical Veterans: The Domain-Insight Advantage

After 20+ years in maternal-fetal medicine, I can diagnose systemic workflow problems in seconds:

  • The billing code that never captures the actual clinical complexity
  • The EHR click sequence that wastes 90 seconds per patient
  • The handoff protocol that consistently loses critical information
  • The documentation template that forces unnatural clinical reasoning

This pattern recognition is irreplaceable. No software engineer — however talented — can identify these friction points without living inside clinical flow for years.

We experienced physicians know exactly what’s broken. We’ve complained about it in physician lounges for decades.

But here’s our Achilles heel: Most of us trained when medical software was stable, certified, and changed once per decade. The modern AI ecosystem — with its weekly model updates, shifting APIs, and “move fast and break things” ethos — feels antithetical to medical culture.

Our strength: We know which problems actually matter.
Our barrier: We expect stability in a space that rewards rapid iteration.

AI-Native Trainees: The Workflow-Fluency Advantage

Meanwhile, today’s residents and medical students entered training already fluent in:

  • Prompt engineering for clinical reasoning
  • No-code automation for workflow hacks
  • Rapid prototyping of custom tools
  • Iterative experimentation without fear of “breaking” things

They don’t ask “Is this tool approved?” They ask “Does this solve my problem right now?”

Imagine a resident builds a custom triage tool that auto-prioritizes patient callbacks based on keyword extraction from nursing messages. Technically elegant. Built over a weekend. Saves 30 minutes daily.

Then imagine it flags “patient reports decreased fetal movement at 32 weeks” as low-priority because the message doesn’t contain words like “urgent” or “emergency.”

Any MFM specialist knows that’s a potential sentinel event.

Their strength: Technical fluency meets operational fearlessness.
Their limitation: They haven’t yet attended enough funerals to know which shortcuts kill.

They can build fast — but clinical pattern recognition can’t be compressed into a weekend hackathon.

The Bridge Is the Business Model

Here’s what most clinical AI companies miss: The adoption wedge isn’t veterans OR trainees. It’s veterans THROUGH trainees.

The most successful physician-built tools I’ve seen follow this pattern:

  1. Experienced clinician identifies authentic pain point (knows what to build)
  2. AI-native trainee prototypes rapid solution (knows how to build)
  3. Tool gets stress-tested in real clinical flow (bottom-up adoption)
  4. Senior physicians adopt after seeing it work (peer validation)
  5. Administration buys in once ROI is obvious (institutional scale)

Innovation in medicine doesn’t flow top-down. It spreads laterally through trusted peers.

This is why CodeCraftMD’s initial pilot is with MFM fellows who code. They have enough clinical depth to recognize real problems, enough technical fluency to iterate solutions, and enough institutional credibility to influence attending adoption.

The Real Question: Who Controls Clinical Software?

Here’s what keeps me up at night:

Every specialty will soon have AI-powered cognitive instruments. The question is who builds them.

If physicians don’t participate in this process, we’ll inherit tools designed by people who’ve never:

  • Managed a precipitous breech delivery at 2 AM
  • Explained a devastating diagnosis to a family
  • Made a split-second clinical decision with incomplete information
  • Navigated the cognitive load of 40 patients across three hospitals

Learning to code as a physician isn’t about becoming a software engineer.

It’s about:

  • Preserving clinical epistemology in automated reasoning
  • Ensuring safety constraints are clinically meaningful, not just technically compliant
  • Designing interfaces that augment — not replace — physician judgment
  • Retaining agency over the tools that mediate our practice

The physicians who learn to build now will shape the clinical software ecosystem for the next generation.

The rest will inherit whatever gets built for them.

Three Actions You Can Take This Week

If you’re an experienced clinician frustrated by broken workflows:

Join the next Doctors Who Code workshop where we translate clinical pain points into buildable prototypes. (Next session: February 2026 — [join the waitlist here])

If you’re a trainee already tinkering with AI tools:

Document what you’re building. Your workflow hacks today become tomorrow’s institutional solutions. Share them in the DWC community forum.

If you’re a medical educator shaping next-generation training:

Advocate for “Clinical Software Design” electives in your residency program. The physicians you’re training will spend more time interacting with software than with patients. They need to understand how it’s built.


The future of medicine isn’t physician vs. AI. It’s physician-built AI vs. vendor-built AI.

Which side of that divide do you want to be on?

— Chukwuma Onyeije, MD, FACOG
Writing about physician-built tools, clinical AI, and the future of medical software at lightslategray-turtle-256743.hostingersite.com

Next post: Inside CodeCraftMD — Why I built billing automation first (and what it taught me about clinical AI adoption)