Would you trust an AI doctor? New research shows patients are split

May 19, 2023

A University of Arizona Health Sciences-led study found that more than 50% of people don’t fully trust AI-powered medical advice, but many put faith in AI if it’s monitored and guided by human touch.

Image
Participants in a new study trusted medical advice generated by artificial intelligence more when their doctors expressed support of the new technology.

Participants in a new study trusted medical advice generated by artificial intelligence more when their doctors expressed support of the new technology.

Artificial intelligence-powered medical treatment options are on the rise and have the potential to improve diagnostic accuracy, but a new study led by University of Arizona Health Sciences researchers found that about 52% of participants would choose a human doctor rather than AI for diagnosis and treatment.

Image
portrait of Marvin Slepian, MD

Marvin Slepian, MD, JD, is a Regents Professor in the University of Arizona College of Medicine – Tucson and founding director of the Arizona Center for Accelerated Biomedical Innovation, which focuses on developing solutions for unmet medical needs.

The paper, “Diverse Patients Attitudes Towards Artificial Intelligence (AI) in Diagnosis,” was published today in the journal PLOS Digital Health.

The research was led by Marvin J. Slepian, MD, JD, Regents Professor of Medicine at the UArizona College of Medicine – Tucson and member of the BIO5 Institute, and Christopher Robertson, JD, professor of law and associate dean for strategic initiatives at Boston University. The research team found that most patients aren’t convinced the diagnoses provided by AI are as trustworthy of those delivered by human medical professionals.

“While many patients appear resistant to the use of AI, accuracy of information, nudges and a listening patient experience may help increase acceptance,” Dr. Slepian said of the study’s other primary finding: that a human touch can help clinical practices use AI to their advantage and earn patients’ trust. “To ensure that the benefits of AI are secured in clinical practice, future research on best methods of physician incorporation and patient decision making is required.”

In the National Institutes of Health-funded study, participants were placed into scenarios as mock patients and asked whether they would prefer to have an AI system or a physical doctor for diagnosis and treatment, and under what circumstances. 

In the first phase, researchers conducted structured interviews with actual patients, testing their reactions to current and future AI technologies. In the second phase of the study, researchers polled 2,472 participants across diverse ethnic, racial and socioeconomic groups using a blinded, randomized survey that tested eight variables.  

Overall, participants were almost evenly split, with more than 52% choosing human doctors as a preference versus approximately 47% choosing an AI diagnostic method. If study participants were prompted that their primary care physicians felt AI was superior and helpful as an adjunct to diagnosis or otherwise nudged to consider AI as good, the acceptance of AI by study participants on re-questioning increased. This signaled the significance of the human physician in guiding a patient’s decision.

Disease severity – leukemia versus sleep apnea – did not affect participants’ trust in AI. Compared to white participants, Black participants selected AI less often and Native Americans selected it more often. Older participants were less likely to choose AI, as were those who self-identified as politically conservative or viewed religion as important.

The racial, ethnic and social disparities identified suggest that differing groups will warrant specific sensitivity and attention as to informing them as to the value and utility of AI to enhance diagnoses.

“I really feel this study has the import for national reach. It will guide many future studies and clinical translational decisions even now,” Dr. Slepian said. “The onus will be on physicians and others in health care to ensure that information that resides in AI systems is accurate, and to continue to maintain and enhance the accuracy of AI systems as they will play an increasing role in the future of health care.”

Co-authors include Andrew Woods, JD, Milton O. Riepe professor of law and co-director of the TechLaw program at the UArizona James E. Rogers College of Law; Kelly Bergstrand, PhD, associate professor of sociology and anthropology at the University of Texas at Arlington; Jess Findley, JD, PhD, professor of practice and director of bar and academic success at UArizona James E. Rogers College of Law; and Cayley Balser, JD,  postgraduate at Innovation for Justice, housed at both the UArizona James E. Rogers College of Law and the University of Utah David Eccles School of Business.

This research was funded in part by the National Institutes of Health under award no. 3R25HL126140-05S1.
 

Our Experts

Contact

Phil Villarreal
University of Arizona Health Sciences Office of Communications
520-621-8659
pvillarreal@arizona.edu