Navigating the New landscape of Medical Advice: AI, Misinformation, and the Evolving Doctor-Patient Relationship
The practice of medicine is undergoing a rapid transformation, not just in treatment modalities, but in how patients seek and interpret health information. As a primary care physician, I’m witnessing this shift firsthand, and it presents a complex set of challenges - and opportunities – for clinicians today.It’s a landscape increasingly populated by AI-generated suggestions, readily available (and frequently enough inaccurate) information from social media, and a growing tendency for individuals to self-diagnose and seek advice outside of customary healthcare channels.
This isn’t entirely new. Physicians have always had to address patient beliefs shaped by external sources. But the sheer volume and accessibility of information, coupled with the perceived authority of algorithms and online personalities, is amplifying the issue. A recent article highlighted this tension, noting the need for “patience and curiosity” when addressing patient requests informed by non-evidence-based sources. While admirable in theory, the reality of packed schedules and demanding workloads makes that level of sustained engagement feel increasingly unrealistic. We’re being asked to not just be doctors,but to be constant debunkers,educators,and patient advocates against a tide of digital noise.
The Promise and peril of AI in Dermatology: A Case Study
The integration of Artificial Intelligence into healthcare isn’t just a future prospect; it’s happening now. A recent JAMA Editor’s Note explored the potential of AI to improve the cost-effectiveness of 3D total-body photography for melanoma screening. As someone who has personally navigated the anxiety and inconvenience of numerous skin biopsies, this topic resonated deeply.
The research, building on a randomized clinical trial published in JAMA Dermatology, revealed a captivating paradox. While 3D photography led to more biopsies,it didn’t actually increase the detection rate of melanomas. This suggests the technology is highly sensitive, but perhaps not specific enough to justify its current cost. A companion study in JAMA Dermatology confirmed this, finding the procedure currently isn’t cost-effective. However, the authors optimistically suggest that AI enhancements could possibly bridge the gap, making it a more viable screening tool in the future. For now, for many high-risk patients, “usual care” remains the standard.
This example illustrates a crucial point: AI isn’t a replacement for clinical judgment,but a potential tool to augment it. The challenge lies in ensuring that AI is deployed responsibly, with a clear understanding of its limitations and a focus on improving patient outcomes, not simply increasing testing volume.
The Curious Case of the Online Medical Consultant: When Patients Turn to Google (and Friends)
The demand for readily available information extends beyond specialized areas like dermatology. I frequently encounter situations where friends and acquaintances seek my medical opinion outside of a formal patient-physician relationship. It’s a common phenomenon – the assumption that a primary care physician possesses a universal understanding of all things medical. Frequently enough, these inquiries stem from a desire to avoid “bothering” their own doctor, a sentiment I find both perplexing and concerning. Why share protected health information with someone who isn’t directly involved in their care?
I recently tested this tendency by exploring how AI tools responded to a friend’s question about a radiology report finding: “pleural based opacity.” Both Google and Copilot provided explanations that were surprisingly accurate and aligned with my own understanding. However, the crucial difference lay in the follow-up advice. While I emphasized the importance of discussing the finding with the ordering physician within the context of their individual clinical picture, the AI sources universally recommended “further investigation,” a phrase that many patients would interpret as a need for additional, potentially needless, testing.
This highlights a critical limitation of AI: its inability to provide nuanced, personalized advice. It can offer information, but it can’t replicate the critical thinking and contextual understanding of a trained clinician.
The Limits of Empathy (and the Surprisingly Accurate Google)
My response to another unsolicited medical inquiry – a detailed question about back pain treatments from a high school acquaintance via Facebook – was admittedly less patient. Overwhelmed and short on time, I offered a generic, albeit empathetic, response: “so many factors play into the choice of treatments and it really depends on the patient.”
Later, out of curiosity, I ran the same question through Google. The results were remarkably thorough, outlining various treatment options and concluding with a crucial disclaimer: “Notable note: The choice of treatment depends on the specific nature and severity of the herniated disc,







