Dies ist eine Übersichtsseite mit Metadaten zu dieser wissenschaftlichen Arbeit. Der vollständige Artikel ist beim Verlag verfügbar.
Evaluation of the quality of information provided by ChatGPT on pelvic and acetabular surgery
1
Zitationen
6
Autoren
2025
Jahr
Abstract
Pelvic and Acetabular fractures are complex injuries commonly caused by both high-energy mechanisms in young patients and low velocity trauma in older patients, which can result in significant morbidity and mortality for both . This study aimed to assess the information provided by ChatGPT on pelvic and acetabular surgery using standard scoring systems to assess the quality, reliability, and readability of the content. We hypothesized that while ChatGPT would generate information of high quality, its readability would be low. An open AI model (ChatGPT) was used to answer 20 commonly asked questions from patients about Pelvic and Acetabular surgery. These answers were evaluated for medical accuracy, quality, and readability using the JAMA Benchmark criteria, DISCERN score, Flesch-Kincaid Reading Ease Score (FRES), and Grade Level (FKGL). The JAMA Benchmark criteria score was 0, the lowest score, indicating no reliable resources cited. The DISCERN score was 42.5, which scores information of fair quality. The area the open AI model scored lowest was the reliability portion of the DISCERN score due to a lack of resources. The FRES was 52.9, and the FKGL was at a 10th-12th grade reading level. A 10th to 12th grade reading level was required to comprehend the information provided by ChatGPT with regards to Pelvic and acetabular surgery, and the evidence supplied was of fair quality. With no citations provided, it remains unclear where these answers originate and how scientifically valid they are with respect to the latest literature. In spite of this, ChatGPT did safety-net patients by encouraging the importance of further discussion with a surgeon.
Ähnliche Arbeiten
Explainable Artificial Intelligence (XAI): Concepts, taxonomies, opportunities and challenges toward responsible AI
2019 · 8.418 Zit.
Stop explaining black box machine learning models for high stakes decisions and use interpretable models instead
2019 · 8.288 Zit.
High-performance medicine: the convergence of human and artificial intelligence
2018 · 7.726 Zit.
Proceedings of the 19th International Joint Conference on Artificial Intelligence
2005 · 5.781 Zit.
Peeking Inside the Black-Box: A Survey on Explainable Artificial Intelligence (XAI)
2018 · 5.516 Zit.