This article has been reviewed according to Science X's editorial process and policies. Editors have highlighted the following attributes while ensuring the content's credibility:


peer-reviewed publication

reputable news agency


ChatGPT diagnoses ER patients 'like a human doctor': Study

Credit: Unsplash/CC0 Public Domain

Artificial intelligence chatbot ChatGPT diagnosed patients rushed to emergency at least as well as doctors and in some cases outperformed them, Dutch researchers have found, saying AI could "revolutionize the medical field".

But the report published Wednesday also stressed ER doctors needn't hang up their scrubs just yet, with the chatbot potentially able to speed up diagnosis but not replace human medical judgment and experience.

Scientists examined 30 cases treated in an in the Netherlands in 2022, feeding in anonymized patient history, lab tests and the doctors' own observations to ChatGPT, asking it to provide five possible diagnoses.

They then compared the chatbot's shortlist to the same five diagnoses suggested by ER doctors with access to the same information, then cross-checked with the correct diagnosis in each case.

Doctors had the correct diagnosis in the top five in 87 percent of cases, compared to 97 percent for ChatGPT version 3.5 and 87 percent for version 4.0.

"Simply put, this indicates that ChatGPT was able to suggest medical diagnoses much like a human doctor would," said Hidde ten Berg, from the emergency medicine department at the Netherlands' Jeroen Bosch Hospital.

Co-author Steef Kurstjens told AFP the survey did not indicate that computers could one day be running the ER, but that AI can play a vital role in assisting under-pressure medics.

"The key point is that the chatbot doesn't replace the physician but it can help in providing a diagnosis and it can maybe come up with ideas the doctor hasn't thought of," Kurstjens told AFP.

Large language models such as ChatGPT are not designed as , he stressed, and there would also be about feeding confidential and sensitive medical data into a chatbot.


And as in other fields, ChatGPT showed some limitations.

The chatbot's reasoning was "at times medically implausible or inconsistent, which can lead to misinformation or incorrect diagnosis, with significant implications," the report noted.

The scientists also admitted some shortcomings with the research. The sample size was small, with 30 cases examined. In addition, only relatively simple cases were looked at, with patients presenting a single primary complaint.

It was not clear how well the chatbot would fare with more complex cases. "The efficacy of ChatGPT in providing multiple distinct diagnoses for patients with complex or rare diseases remains unverified."

Sometimes the chatbot did not provide the correct in its top five possibilities, Kurstjens explained, notably in the case of an abdominal aneurysm, a potentially life-threatening complication where the aorta artery swells up.

The only consolation for ChatGPT: in that case the doctor got it wrong too.

The report sets out what it calls the medical "bloopers" the chatbot made, for example diagnosing anemia (low hemoglobin levels in the blood) in a patient with a normal hemoglobin count.

"It's vital to remember that ChatGPT is not a and there are concerns over privacy when using ChatGPT with medical data," concluded ten Berg.

"However, there is potential here for saving time and reducing waiting times in the emergency department. The benefit of using could be in supporting with less experience, or it could help in spotting rare diseases," he added.

The findings—published in the medical journal Annals of Emergency Medicine—will be presented at the European Emergency Medicine Congress (EUSEM) 2023 in Barcelona.

More information: Hidde ten Berg et al, ChatGPT and Generating a Differential Diagnosis Early in an Emergency Department Presentation, Annals of Emergency Medicine (2023). DOI: 10.1016/j.annemergmed.2023.08.003

Journal information: Annals of Emergency Medicine

© 2023 AFP

Citation: ChatGPT diagnoses ER patients 'like a human doctor': Study (2023, September 13) retrieved 20 June 2024 from
This document is subject to copyright. Apart from any fair dealing for the purpose of private study or research, no part may be reproduced without the written permission. The content is provided for information purposes only.

Explore further

Exploring potential applications for ChatGPT in nuclear medicine and molecular imaging


Feedback to editors