Can Popular AI Large Language Models Provide Reliable Answers to Frequently Asked Questions About Rotator Cuff Tears

Loading...
Publication Logo

Date

2025

Journal Title

Journal ISSN

Volume Title

Publisher

Elsevier

Open Access Color

GOLD

Green Open Access

Yes

OpenAIRE Downloads

OpenAIRE Views

Publicly Funded

No
Impulse
Top 10%
Influence
Top 10%
Popularity
Top 10%

Research Projects

Journal Issue

Abstract

Background: Rotator cuff tears are common upper-extremity injuries that significantly impair shoulder function, leading to pain, reduced range of motion, and a decrease in quality of life. With the increasing reliance on artificial intelligence large language models (AI LLMs) for health information, it is crucial to evaluate the quality and readability of the information provided by these models. Methods: A pool of 50 questions was generated related to rotator cuff tear by querying popular AI LLMs (ChatGPT 3.5, ChatGPT 4, Gemini, and Microsoft CoPilot) and using Google search. After that, responses from the AI LLMs were saved and evaluated. For information quality the DISCERN tool and a Likert Scale was used, for readability the Patient Education Materials Assessment Tool for Printable Materials (PEMAT) Understandability Score and the Flesch-Kincaid Reading Ease Score was used. Two orthopedic surgeons assessed the responses, and discrepancies were resolved by a senior author. Results: Out of 198 answers, the median DISCERN score was 40, with 56.6% considered sufficient. The Likert Scale showed 96% sufficiency. The median PEMAT Understandability score was 83.33, with 77.3% sufficiency, while the Flesch-Kincaid Reading Ease score had a median of 42.05 with 88.9% sufficiency. Overall, 39.8% of the answers were sufficient in both information quality and readability. Differences were found among AI models in DISCERN, Likert, PEMAT Understandability, and Flesch-Kincaid scores. Conclusion: AI LLMs generally cannot offer sufficient information quality and readability. While they are not ready for use in medical field, they show a promising future. There is a necessity for continuous reevaluation of these models due to their rapid evolution. Developing new, comprehensive tools for evaluating medical information quality and readability is crucial for ensuring these models can effectively support patient education. Future research should focus on enhancing readability and consistent information quality to better serve patients. (c) 2024 The Author(s). Published by Elsevier Inc. on behalf of American Shoulder and Elbow Surgeons. This is an open access article under the CC BY-NC-ND license (http://creativecommons.org/licenses/by-nc-nd/4.0/).

Description

Karademir, Orhan Mete/0009-0002-6254-7071

Keywords

Artificial Intelligence, Large Language Models, Rotator Cuff Tears, Frequently Asked Questions, Patient Information, AI Tools in Healthcare, ChatGPT, Shoulder

Fields of Science

Citation

WoS Q

N/A

Scopus Q

Q2
OpenCitations Logo
OpenCitations Citation Count
N/A

Source

JSES International

Volume

9

Issue

2

Start Page

390

End Page

397
PlumX Metrics
Citations

Scopus : 11

PubMed : 4

Captures

Mendeley Readers : 23

SCOPUS™ Citations

11

checked on Feb 14, 2026

Web of Science™ Citations

9

checked on Feb 14, 2026

Page Views

1

checked on Feb 14, 2026

Google Scholar Logo
Google Scholar™
OpenAlex Logo
OpenAlex FWCI
8.19073818

Sustainable Development Goals