Start of Main Content
Author(s)

Xuhai Xu

Haoyu Hu

Haoran Zhang

Will Wang

Reina Wang

Luis Soenksen

Omar Badri

Sheharbano Jafry

Elise Burger

Lotanna Nwandu

Apporva Mehta

Eric Duhaime

Asif Qasim

Hause Lin

Janis Pereira

Jonathon Hershon

Matthew Groh

Philipp Tschandl

Roxana Daneshjou

Marzyeh Ghassemi

Artificial intelligence (AI) is increasingly permeating healthcare, from physician assistants to consumer applications. Since AI algorithm's opacity challenges human interaction, explainable AI (XAI) addresses this by providing AI decision-making insight, but evidence suggests XAI can paradoxically induce over-reliance or bias. We present results from two large-scale experiments (623 lay people; 153 primary care physicians, PCPs) combining a fairness-based diagnosis AI model and different XAI explanations to examine how XAI assistance, particularly multimodal large language models (LLMs), influences diagnostic performance. AI assistance balanced across skin tones improved accuracy and reduced diagnostic disparities. However, LLM explanations yielded divergent effects: lay users showed higher automation bias - accuracy boosted when AI was correct, reduced when AI erred - while experienced PCPs remained resilient, benefiting irrespective of AI accuracy. Presenting AI suggestions first also led to worse outcomes when the AI was incorrect for both groups. These findings highlight XAI's varying impact based on expertise and timing, underscoring LLMs as a "double-edged sword" in medical AI and informing future human-AI collaborative system design.
Date Published: 2025
Citations: Xu, Xuhai, Haoyu Hu, Haoran Zhang, Will Wang, Reina Wang, Luis Soenksen, Omar Badri, Sheharbano Jafry, Elise Burger, Lotanna Nwandu, Apporva Mehta, Eric Duhaime, Asif Qasim, Hause Lin, Janis Pereira, Jonathon Hershon, Matthew Groh, Philipp Tschandl, Roxana Daneshjou, Marzyeh Ghassemi. 2025. Explainable AI as a Double-Edged Sword in Dermatology: The Impact on Clinicians versus The Public Authors.