News

ChatGPT helpful for breast cancer screening advice with certain caveats, new study finds

5 Apr 2023
ChatGPT helpful for breast cancer screening advice with certain caveats, new study finds

As more consumers turn to the newly available ChatGPT for health advice, researchers are eager to see whether the information provided by the artificial intelligence chatbot is reliable and accurate.

A new study conducted by researchers at the University of Maryland School of Medicine (UMSOM) indicates that the answers generated provide correct information the vast majority of the time; sometimes, though, the information is inaccurate or even fictitious.

Findings were published today in the journal Radiology.

In February 2023, UMSOM researchers created a set of 25 questions related to advice on getting screened for breast cancer.

They submitted each question to ChatGPT three times to see what responses were generated.

(The chatbot is known for varying its response each time a question is posed.)

Three radiologists fellowship-trained in mammography evaluated the responses; they found that the responses were appropriate for 22 out of the 25 questions.

The chatbot did, however, provide one answer based on outdated information.

Two other questions had inconsistent responses that varied significantly each time the same question was posed.

“We found ChatGPT answered questions correctly about 88 percent of the time, which is pretty amazing,” said study corresponding author Paul Yi , MD, Assistant Professor of Diagnostic Radiology and Nuclear Medicine at UMSOM and Director of the UM Medical Intelligent Imaging Center (UM2ii). 

“It also has the added benefit of summarising information into an easily digestible form for consumers to easily understand.” 

ChatGPT correctly answered questions about the symptoms of breast cancer, who is at risk, and questions on the cost, age, and frequency recommendations concerning mammograms.

The downside is that it is not as comprehensive in its responses as what a person would normally find on a Google search.  

“ChatGPT provided only one set of recommendations on breast cancer screening, issued from the American Cancer Society, but did not mention differing recommendations put out by the Centers for Disease Control and Prevention (CDC) or the US Preventative Services Task Force (USPSTF),” said study lead author Hana Haver, MD, a radiology resident at University of Maryland Medical Center.

In one response deemed by the researchers to be inappropriate, ChatGPT provided an outdated response to planning a mammogram around COVID-19 vaccination.

The advice to delay a mammogram for four to six weeks after getting a COVID-19 shot was changed in February 2022, and the CDC endorses the USPSTF guidelines, which don’t recommend waiting.

Inconsistent responses were given to questions concerning an individual’s personal risk of getting breast cancer and on where someone could get a mammogram.

“We’ve seen in our experience that ChatGPT sometimes makes up fake journal articles or health consortiums to support its claims,” said Dr. Yi. 

“Consumers should be aware that these are new, unproven technologies, and should still rely on their doctor, rather than ChatGPT, for advice.” 

He and his colleagues are now analysing how ChatGPT fares for lung cancer screening recommendations and identifying ways to improve the recommendations made by ChatGPT to be more accurate and complete – as well as understandable to those without a high level of education.

“With the rapid evolution of ChatGPT and other large language models, we have a responsibility as a medical community to evaluate these technologies and protect our patients from potential harm that may come from incorrect screening recommendations or outdated preventive health strategies,” said Mark T. Gladwin, MD, Dean, University of Maryland School of Medicine, Vice President for Medical Affairs, University of Maryland, Baltimore, and the John Z. and Akiko K. Bowers Distinguished Professor.

Jean Jeudy, MD, Professor of Diagnostic Radiology and Nuclear Medicine and Radiology Vice Chair of Informatics at UMSOM was a co-author on the study. Researchers from the Massachusetts General Hospital and the Johns Hopkins University School of Medicine also participated in this study.

The study was funded by the individual institutions.

Source: University of Maryland School of Medicine