This article has been reviewed according to Science X's editorial process and policies. Editors have highlighted the following attributes while ensuring the content's credibility:

fact-checked

peer-reviewed publication

trusted source

proofread

Would you trust an AI doctor? New research shows patients are split

Would you trust an AI doctor? New research shows patients are split
Participants in a new study trusted medical advice generated by artificial intelligence more when their doctors expressed support of the new technology. Credit: University of Arizona

Artificial intelligence-powered medical treatment options are on the rise and have the potential to improve diagnostic accuracy, but a new study led by University of Arizona Health Sciences researchers found that about 52% of participants would choose a human doctor rather than AI for diagnosis and treatment.

The paper, "Diverse Patients Attitudes Towards Artificial Intelligence (AI) in Diagnosis," was published May 19 in the journal PLOS Digital Health.

The research was led by Dr. Marvin J. Slepian, Regents Professor of Medicine at the UArizona College of Medicine–Tucson and member of the BIO5 Institute, and Christopher Robertson, professor of law and associate dean for strategic initiatives at the Boston University School of Law. The research team found that most patients aren't convinced the diagnoses provided by AI are as trustworthy of those delivered by human medical professionals.

"While many patients appear resistant to the use of AI, accuracy of information, nudges and a listening patient experience may help increase acceptance," Slepian said of the study's other primary finding: that a human touch can help use AI to their advantage and earn patients' trust. "To ensure that the benefits of AI are secured in , future research on best methods of physician incorporation and patient decision making is required."

In the study, participants were placed into scenarios as mock patients and asked whether they would prefer to have an AI system or a physical doctor for diagnosis and treatment, and under what circumstances.

In the first phase, researchers conducted structured interviews with actual patients, testing their reactions to current and future AI technologies. In the second phase of the study, researchers polled 2,472 participants across diverse ethnic, racial and socioeconomic groups using a blinded, randomized survey that tested eight variables.

Overall, participants were almost evenly split, with more than 52% choosing human doctors as a preference versus approximately 47% choosing an AI diagnostic method. If study participants were prompted that their felt AI was superior and helpful as an adjunct to diagnosis or otherwise nudged to consider AI as good, the acceptance of AI by study participants on re-questioning increased. This signaled the significance of the human physician in guiding a patient's decision.

Disease severity—leukemia versus sleep apnea—did not affect participants' trust in AI. Compared to white participants, Black participants selected AI less often and Native Americans selected it more often. Older participants were less likely to choose AI, as were those who self-identified as politically conservative or viewed religion as important.

The racial, ethnic and social disparities identified suggest that differing groups will warrant specific sensitivity and attention as to informing them as to the value and utility of AI to enhance diagnoses.

"I really feel this study has the import for national reach. It will guide many future studies and clinical translational decisions even now," said Slepian, who also holds a J.D. "The onus will be on physicians and others in health care to ensure that information that resides in AI systems is accurate, and to continue to maintain and enhance the accuracy of AI systems as they will play an increasing role in the future of ."

More information: Christopher Robertson et al, Diverse patients' attitudes towards Artificial Intelligence (AI) in diagnosis, PLOS Digital Health (2023). DOI: 10.1371/journal.pdig.0000237

Journal information: PLOS Digital Health

Citation: Would you trust an AI doctor? New research shows patients are split (2023, May 22) retrieved 4 October 2023 from https://medicalxpress.com/news/2023-05-ai-doctor-patients.html
This document is subject to copyright. Apart from any fair dealing for the purpose of private study or research, no part may be reproduced without the written permission. The content is provided for information purposes only.

Explore further

Understanding what prevents shared decision making from wider implementation with Black patients

25 shares

Feedback to editors