Dies ist eine Übersichtsseite mit Metadaten zu dieser wissenschaftlichen Arbeit. Der vollständige Artikel ist beim Verlag verfügbar.
Accuracy, Reliability, and Comprehensibility of ChatGPT-Generated Medical Responses for Patients With Nonalcoholic Fatty Liver Disease
81
Zitationen
14
Autoren
2023
Jahr
Abstract
Nonalcoholic fatty liver disease (NAFLD) is an increasing global health problem and is expected to become the leading indication for liver transplantation.<sup>1</sup> There are no approved NAFLD-specific pharmacotherapies, and lifestyle modification is the primary recommended therapy.<sup>2</sup> Innovative approaches to facilitate the implementation and long-term maintenance of lifestyle changes are needed to address the challenging and complex nature of the management of NAFLD, which recently was renamed as metabolic dysfunction-associated steatotic liver disease, to overcome the limitations and stigma of the previous name.<sup>3,4</sup> Artificial intelligence (AI)-powered chatbots have been shown to provide effective personalized support and education to patients, with the potential to complement health care resources. The OpenAI Foundation's AI chatbot, Chat Generative Pretrained Transformer (ChatGPT), has attracted worldwide attention for its remarkable performance in question-answer tasks.<sup>5-7</sup> This study evaluated the accuracy, completeness, and comprehensiveness of chatGPT's responses to NAFLD-related questions, with the aim of assessing its performance in addressing patients' queries about the disease and lifestyle behaviors.
Ähnliche Arbeiten
Explainable Artificial Intelligence (XAI): Concepts, taxonomies, opportunities and challenges toward responsible AI
2019 · 8.400 Zit.
Stop explaining black box machine learning models for high stakes decisions and use interpretable models instead
2019 · 8.261 Zit.
High-performance medicine: the convergence of human and artificial intelligence
2018 · 7.695 Zit.
Proceedings of the 19th International Joint Conference on Artificial Intelligence
2005 · 5.781 Zit.
Peeking Inside the Black-Box: A Survey on Explainable Artificial Intelligence (XAI)
2018 · 5.506 Zit.
Autoren
Institutionen
- Humanitas University(IT)
- IRCCS Humanitas Research Hospital(IT)
- Chinese University of Hong Kong(HK)
- Johannes Gutenberg University Mainz(DE)
- University Medical Center of the Johannes Gutenberg University Mainz(DE)
- Centro de Investigación Biomédica en Red(ES)
- Instituto de Biomedicina de Sevilla(ES)
- Hospital Universitario Virgen del Rocío(ES)
- Universidad de Sevilla(ES)
- McGill University Health Centre(CA)
- Inserm(FR)
- Université Paris Cité(FR)
- Hôpital Beaujon(FR)
- Sorbonne Paris Cité(FR)
- Assistance Publique – Hôpitaux de Paris(FR)
- Imperial College Healthcare NHS Trust(GB)
- University College London(GB)
- Università Cattolica del Sacro Cuore(IT)
- University of Palermo(IT)
- University of Milan(IT)
- Fondazione IRCCS Ca' Granda Ospedale Maggiore Policlinico(IT)
- Ospedale Maggiore(IT)
- Inova Health System(US)
- Inova Fairfax Hospital(US)