AI-Assisted Diagnosis: Why It Makes Your Doctor Smarter, Not Obsolete
By Dr. RP, MD | Analog Precision Medicine
The question I hear most often when I describe how Analog Precision Medicine uses AI is: "So is the computer making the diagnosis?" The answer is no — and the way that question gets asked reveals something important about how the public understands artificial intelligence in medicine, which is mostly through a lens of replacement rather than augmentation.
Where AI Demonstrably Improves Clinical Performance
- —Medical imaging — Google's DeepMind matched or exceeded top ophthalmology consultants on detecting 50+ ocular conditions from OCT scans; deep learning classifies skin cancer with accuracy comparable to board-certified dermatologists
- —Cardiology — AI interpretation of 12-lead ECGs can identify hypertrophic cardiomyopathy, cardiac amyloidosis, and low-EF heart failure from the surface ECG alone — genuinely novel capabilities
- —Colonoscopy — AI-assisted real-time computer vision has demonstrated significant increases in adenoma detection rate in randomized controlled trials, with reductions in miss rates for small polyps
- —LLM clinical performance — GPT-4 passed the USMLE with a high-performing score; multiple studies have found LLM-generated differential diagnoses rated comparable to attending physicians for structured clinical scenarios
- —Sepsis prediction — ML-based early warning systems can identify sepsis risk hours before conventional triggers, though clinical impact depends heavily on workflow design
Where AI Fails — and Why the Limitations Matter
- —Training data bias — AI systems perpetuate and can amplify historical clinical biases; a 2019 Science study found a widely used algorithm was significantly less accurate for Black patients than white patients
- —Distribution shift — systems trained in one clinical environment often underperform when deployed in another due to differing patient demographics, equipment, and documentation styles
- —Explainability gaps — many high-performing ML systems are not interpretable; a physician who cannot understand why an algorithm flagged a patient cannot critically evaluate the recommendation
- —Narrow scope — AI systems are trained for specific tasks and do not generalize; clinical medicine requires integrating findings across multiple domains simultaneously
- —Automation bias — the documented tendency to over-rely on algorithmic recommendations, reducing independent assessment; a real risk in AI-assisted clinical settings
AI as Clinical Amplifier
The accurate model for AI's role in precision medicine is amplification, not replacement:
- —Data synthesis — multi-omic profiling generates volumes that exceed what any physician can hold in working memory; AI does the analytical legwork so the physician makes a better-informed judgment
- —Literature currency — the biomedical literature grows by ~1.5 million articles per year; AI-assisted literature review ensures recommendations reflect current evidence
- —Cognitive offloading — documentation and routine monitoring can be substantially automated, freeing physician cognitive resources for contextual judgment, nuanced communication, and integrating patient values
- —Pattern recognition assistance — in complex diagnostic cases, AI tools can surface overlooked possibilities and flag relevant historical literature
What AI Cannot Replace
- —Contextual clinical judgment — integrating a patient's history, values, family situation, psychological state, and risk tolerance into a coherent recommendation is irreducibly human
- —The therapeutic relationship — patient trust, therapeutic alliance, and the ability to deliver difficult information with emotional attunement are not computational problems
- —Ethical reasoning — clinical medicine routinely navigates competing values and decisions under irreducible uncertainty
- —Accountability — a physician is accountable legally, ethically, and professionally for clinical decisions; AI is a tool
The central paradox: we use more technology than any large health system deploys for individual patients, specifically so the human element of medicine can be more present, not less.
Bottom Line
AI is demonstrably improving clinical performance in specific, well-defined domains — particularly medical imaging, ECG interpretation, colonoscopy, and structured clinical benchmarks. The right framework is amplification: AI extends what a skilled physician can know and synthesize, enabling better-informed judgment rather than replacing it. Patients should be skeptical of practices that use AI to thin the physician relationship and enthusiastic about practices that use it to enrich it.
References
- 1. De Fauw J, et al. Clinically applicable deep learning for diagnosis and referral in retinal disease. Nat Med. 2018;24(9):1342–1350.
- 2. Esteva A, et al. Dermatologist-level classification of skin cancer with deep neural networks. Nature. 2017;542(7639):115–118.
- 3. Attia ZI, et al. Screening for cardiac contractile dysfunction using an AI-enabled electrocardiogram. Nat Med. 2019;25(1):70–74.
- 4. Wong A, et al. External validation of a widely implemented proprietary sepsis prediction model. JAMA Intern Med. 2021;181(8):1065–1070.
- 5. Wang P, et al. Effect of a deep-learning CAD system on adenoma detection during colonoscopy. JAMA. 2020;323(14):1383–1392.
- 6. Nori H, et al. Capabilities of GPT-4 on medical challenge problems. arXiv. 2023;arXiv:2303.13375.
- 7. Obermeyer Z, et al. Dissecting racial bias in an algorithm used to manage the health of populations. Science. 2019;366(6464):447–453.
- 8. Goddard K, et al. Automation bias: a systematic review of frequency, effect mediators, and mitigants. J Am Med Inform Assoc. 2012;19(1):121–127.
Dr. RP, MD is dual board-certified in Emergency Medicine and Critical Care Medicine and is the founder of Analog Precision Medicine, a precision medicine practice in Southern California. This article is for educational purposes only and does not constitute medical advice or establish a physician-patient relationship.
