Gemini 1.5 Flash Provides the Most Reliable Content While ChatGPT-4o Offers the Highest Readability for Patient Education on Meniscal Tears
| dc.authorscopusid | 57241132400 | |
| dc.authorscopusid | 57210134406 | |
| dc.authorscopusid | 58304737100 | |
| dc.authorscopusid | 60259208900 | |
| dc.authorscopusid | 57192870929 | |
| dc.authorwosid | Ziroglu, Nezih/N-2480-2019 | |
| dc.authorwosid | Aloglu Ciftci, Ebru/Lxb-2688-2024 | |
| dc.authorwosid | Koluman, Ali Can/Izq-0097-2023 | |
| dc.authorwosid | Ciftci, Mehmet/Nxb-9001-2025 | |
| dc.contributor.author | Cakmur, Basar Burak | |
| dc.contributor.author | Koluman, Ali Can | |
| dc.contributor.author | Ciftci, Mehmet Utku | |
| dc.contributor.author | Ciftci, Ebru Aloglu | |
| dc.contributor.author | Ziroglu, Nezih | |
| dc.date.accessioned | 2026-01-15T15:12:34Z | |
| dc.date.available | 2026-01-15T15:12:34Z | |
| dc.date.issued | 2025 | |
| dc.department | Okan University | en_US |
| dc.department-temp | [Cakmur, Basar Burak] Beylikduzu State Hosp, Dept Orthoped & Traumatol, Istanbul, Turkiye; [Koluman, Ali Can] Bakirkoy Dr Sadi Konuk Training & Res, Dept Orthoped & Traumatol, Istanbul, Turkiye; [Ciftci, Mehmet Utku] Sultan Abdulhamid Han Training & Res Hosp, Dept Orthoped & Traumatol, Istanbul, Turkiye; [Ciftci, Ebru Aloglu] Istanbul Okan Univ, Fac Hlth Sci, Dept Physiotherapy & Rehabil, Istanbul, Turkiye; [Ciftci, Ebru Aloglu] Istinye Univ, Inst Grad Educ, Dept Physiotherapy & Rehabil, Istanbul, Turkiye; [Ziroglu, Nezih] Acibadem Mehmet Ali Aydinlar Univ, Vocat Sch Hlth Serv, Dept Orthoped Prosthet & Orthot, Istanbul, Turkiye; [Ziroglu, Nezih] Acibadem Univ, Atakent Hosp, Dept Orthopaed & Traumatol, Istanbul, Turkiye | en_US |
| dc.description.abstract | Purpose The aim of this study was to comparatively evaluate the responses generated by three advanced artificial intelligence (AI) models, ChatGPT-4o (OpenAI), Gemini 1.5 Flash (Google) and DeepSeek-V3, to frequently asked patient questions about meniscal tears in terms of reliability, usefulness, quality, and readability. Methods Responses from three AI chatbots, ChatGPT-4o (OpenAI), Gemini 1.5 Flash (Google) and DeepSeek-V3 (DeepSeek AI), were evaluated for 20 common patient questions regarding meniscal tears. Three orthopaedic specialists independently scored reliability and usefulness on 7-point Likert scales and overall response quality using the 5-point Global Quality Scale. Readability was analysed with six established indices. Inter-rater agreement was examined with intraclass correlation coefficients (ICCs) and Fleiss' Kappa, while between-model differences were tested using Kruskal-Wallis and ANOVA with Bonferroni adjustment. Results Gemini 1.5 Flash achieved the highest reliability, significantly outperforming both GPT-4o and DeepSeek-V3 (p = 0.001). While usefulness scores were broadly similar, Gemini was superior to DeepSeek-V3 (p = 0.045). Global Quality Scale scores did not differ significantly among models. In contrast, GPT-4o consistently provided the most readable content (p < 0.001). Inter-rater reliability was excellent across all evaluation domains (ICC > 0.9). Conclusion All three AI models generated high-quality educational content regarding meniscal tears. Gemini 1.5 Flash demonstrated the highest reliability and usefulness, while GPT-4o provided significantly more readable responses. These findings highlight the trade-off between reliability and readability in AI-generated patient education materials and emphasise the importance of physician oversight to ensure safe, evidence-based integration of these tools into clinical practice. | en_US |
| dc.description.woscitationindex | Science Citation Index Expanded | |
| dc.identifier.doi | 10.1002/ksa.70247 | |
| dc.identifier.issn | 0942-2056 | |
| dc.identifier.issn | 1433-7347 | |
| dc.identifier.pmid | 41451752 | |
| dc.identifier.scopus | 2-s2.0-105026030219 | |
| dc.identifier.scopusquality | Q1 | |
| dc.identifier.uri | https://doi.org/10.1002/ksa.70247 | |
| dc.identifier.uri | https://hdl.handle.net/20.500.14517/8700 | |
| dc.identifier.wos | WOS:001649493700001 | |
| dc.identifier.wosquality | Q1 | |
| dc.language.iso | en | en_US |
| dc.publisher | Wiley | en_US |
| dc.relation.ispartof | Knee Surgery Sports Traumatology Arthroscopy | en_US |
| dc.relation.publicationcategory | Makale - Uluslararası Hakemli Dergi - Kurum Öğretim Elemanı | en_US |
| dc.rights | info:eu-repo/semantics/openAccess | en_US |
| dc.subject | ChatGPT | en_US |
| dc.subject | DeepSeek | en_US |
| dc.subject | Gemini | en_US |
| dc.subject | Large Language Models | en_US |
| dc.subject | Meniscal Tear | en_US |
| dc.subject | Patient Education | en_US |
| dc.title | Gemini 1.5 Flash Provides the Most Reliable Content While ChatGPT-4o Offers the Highest Readability for Patient Education on Meniscal Tears | en_US |
| dc.type | Article | en_US |
| dspace.entity.type | Publication |