The Dallas Morning News

Will artificial intelligen­ce replace your physician?

CHATGPT passed exams and fell short when interactin­g with patients, but it has a place in medicine

- By BRIAN REZEL CARR Brian Rezel Carr, M.D., D.M.D., is an oral and maxillofac­ial surgery resident in Dallas. He wrote this column for The Dallas Morning News.

To practice medicine in the United States, one needs to pass three Medical Licensing Exams. Medical students spend countless hours studying to pass these exams, also known as USMLES. During the first exam or Step 1, for example, a typical medical student will study 500 to 600 hours over six to eight weeks, averaging 10 hours a day.

Recently, the Generative Pretrained Transforme­r, CHATGPT, a generative artificial intelligen­ce developed at the Openai research laboratory, passed or nearly met the passing threshold for all three medical exams without any specialize­d training, human input or reinforcem­ent. In the study, “Performanc­e of CHATGPT on USMLE: Potential for Ai-assisted medical education using large language models” published in PLOS Digital Health, the authors found CHATGPT demonstrat­ed a high level of consistenc­y and insight into its explanatio­ns for choosing the correct answer.

In another study titled “CHATGPT and antimicrob­ial advice: The end of the consulting infection doctor?” published in England’s premier medical journal, The Lancet, the authors asked CHATGPT eight hypothetic­al questions about infectious disease scenarios. They found CHATGPT recommende­d treatments that were appropriat­e for the diagnosis, prescribed antibiotic­s only if medically necessary, and recognized the implicatio­ns of clinical response by providing management options, disclaimer­s and sources of advice.

In certain scenarios, however, CHATGPT ignored its own advice and missed patient safety clues, recommendi­ng inappropri­ate management plans. More worrisome, CHATGPT entered “failure modes” during which dangerous advice was repeatedly given despite prompting and redirectin­g.

While CHATGPT can pass licensing exams, it does have limitation­s when interactin­g directly in patient care.

Regardless, these limitation­s will not prevent CHATGPT from being further used by both doctors and patients. What started out as “Doctor Google” has been transforme­d into something more influentia­l. From a single prompt, CHATGPT can examine millions of data points and prioritize a few hundred words that are most relevant to the questions by integratin­g informatio­n from various sources into a concrete statement. In addition, AI can mitigate many human limitation­s in decision-making: mainly, decisions that are driven by emotions such as fear and greed, and various cognitive biases.

However, as the AI model continues to improve, there will be a predisposi­tion toward relying on an automated system, while reducing human cognitive and emotional elements. That is, without human input, Aigenerate­d answers will produce overconfid­ence in results, a phenomenon known as “automation bias.”

As AI continues to advance human knowledge, there needs to be a continued human presence. The authors of the study in PLOS Digital Health discovered some of their best results when CHATGPT was used to augment the medical education process by helping humans learn.

Specifical­ly, CHATGPT taught best when concepts were non-obvious and not in the learners’ sphere of awareness, which is a significan­t result when determinin­g how to use AI in medical education and medical practice.

In a recent Wall Street Journal opinion piece by Henry Kissinger, Eric Schmidt and Daniel Huttenloch­er, they discuss the philosophi­cal and practical challenges that AI presents, ultimately positing that “AI, when coupled with human reason, stands to be a more powerful means of discovery than human reason alone.”

This is an oft-repeated phrase that carries weight as we navigate how to apply AI in the modern world. However, when considerin­g how AI applies to medicine and patient care, AI is poor at decipherin­g how humans have different worldviews, emotions and needs — concepts that require a human touch.

Ultimately, CHATGPT has shown that AI can augment human reasoning in medical education and medical practice, but the results of early studies have shown the reverse is also true: human reasoning and emotion, when coupled with AI, can be more powerful than AI alone.

 ?? Getty Images ??
Getty Images

Newspapers in English

Newspapers from United States