We've updated our Privacy Policy to make it clearer how we use your personal data. We use cookies to provide you with a better experience. You can read our Cookie Policy here.

Advertisement

ChatGPT Can’t Replace Your Doctor Quite Yet

Screenshot of ChatGPT website.
Credit: Jonathan Kemper / Unsplash.
Listen with
Speechify
0:00
Register for free to listen to this article
Thank you. Listen to this article using the player above.

Want to listen to this article for FREE?

Complete the form below to unlock access to ALL audio articles.

Read time: 1 minute

ChatGPT, a sophisticated chatbot driven by artificial intelligence (AI) technology, has been increasingly used in health care contexts, one of which is assisting patients in self-diagnosing before seeking medical help. Although it seems very useful at first glance, AI may cause more harm than good to the patient if it is not accurate in its diagnosis and recommendations. A research team from Japan and the United States recently found that the precision of ChatGPT’s diagnoses and the degree to which it recommends medical consultation require further development.

In a study published in September, the multi-institutional research team led by Tokyo Medical and Dental University (TMDU) evaluated the accuracy (percentage of correct responses) and precision of ChatGPT’s response to five common orthopedic diseases (including carpal tunnel syndrome, cervical myelopathy, and hip osteoarthritis) because orthopedic complaints are very common in clinical practice and comprise up to 26% of the reasons why patients seek care. Over a 5-day course, each of the study researchers submitted the same questions to ChatGPT. The reproducibility between days and researchers was also calculated, and the strength of the recommendation that the patient seek medical attention was evaluated. 

Want more breaking news?

Subscribe to Technology Networks’ daily newsletter, delivering breaking science news straight to your inbox every day.

Subscribe for FREE
“We found that accuracy and reproducibility of ChatGPT’s diagnosis are not consistent over the five conditions. ChatGPT's diagnosis was 100% accurate for carpal tunnel syndrome, but only 4% for cervical myelopathy,” says lead author Tomoyuki Kuroiwa. Additionally, reproducibility between days and researchers varied from “poor” to “almost perfect” among the five conditions even though researchers entered the same questions every time. 

ChatGPT was also inconsistent in recommending medical consultation. Although almost 80% of ChatGPT’s answers recommended medical consultation, only 12.8% included a strong recommendation as set by the study standards. “Without direct language, it is possible that the patient is left confused after self-diagnosis, or worse, experience harm from a misdiagnosis,” says Kuroiwa.

This is the first study to evaluate the reproducibility and degree of the medical consultation recommendation of ChatGPT’s ability to self-diagnose. “In its current form, ChatGPT is inconsistent in both accuracy and precision to help patients diagnose their disease,” explains senior author Koji Fujita. “Given the risk of error and potential harm from misdiagnosis, it is important for any diagnostic tool to include clear language alerting patients to seek expert medical opinions for confirmation of a disease.” 

The researchers also note some limitations of the study including the use of questions simulated by the research team and not patient-derived questions; focusing on only five orthopedic diseases; and using only ChatGPT. While it is still too early to use AI intelligence for self-diagnosis, the training of ChatGPT on diseases of interest could change this. Future studies can help shed light on the role of AI as a diagnostic tool. 


Reference: Kuroiwa T, Sarcon A, Ibara T, et al. The Potential of ChatGPT as a Self-Diagnostic Tool in Common Orthopedic Diseases: Exploratory Study. J Med Internet Res. 2023;25:e47621. doi: 10.2196/47621


This article has been republished from the following materials. Note: material may have been edited for length and content. For further information, please contact the cited source.