Dies ist eine Übersichtsseite mit Metadaten zu dieser wissenschaftlichen Arbeit. Der vollständige Artikel ist beim Verlag verfügbar.
O085 Estimating postoperative mortality in colorectal surgery- a systematic review of risk prediction models
0
Zitationen
6
Autoren
2023
Jahr
Abstract
Abstract Introduction Risk prediction models are frequently used to support decision-making in colorectal surgery but can be inaccurate. Machine learning (ML) is becoming increasingly popular, and its application may increase predictive accuracy. We compared conventional risk prediction models for postoperative mortality (based on regression analysis) with ML models to determine the benefit of the latter approach. Methods The study was registered in PROSPERO(CRD42022364753). Following the PRISMA guidelines, a systematic search of three databases (MEDLINE, EMBASE, WoS) was conducted (from 1/1/2000 to 29/09/2022). Studies were included if they reported the development of a risk model to estimate short-term postoperative mortality for patients undergoing colorectal surgery. Discrimination and calibration performance metrics were compared. Studies were evaluated against CHARMS and TRIPOD criteria. Results 3,052 articles were screened, and 45 studies were included. The total sample size was 1,356,058 patients. Six studies used ML techniques for model development. Most studies (n=42) reported the area under the receiver operating characteristic curve (AUROC) as a measure of discrimination. There was no significant difference in the mean AUROC values between regression models (0.833 s.d.±0.52) and ML (0.846 s.d.±0.55), p=0.539. Calibration statistics, which measure the agreement between predicted estimates and observed outcomes, were less consistent. Risk of bias assessment found most concerns in the data handling and analysis domains of eligible studies. Conclusion Our study showed comparable predictive performance between regression and ML methods in colorectal surgery. Integration of ML in colorectal risk prediction is promising but further refinement of the models is required to support routine clinical adoption.
Ähnliche Arbeiten
Explainable Artificial Intelligence (XAI): Concepts, taxonomies, opportunities and challenges toward responsible AI
2019 · 8.626 Zit.
Stop explaining black box machine learning models for high stakes decisions and use interpretable models instead
2019 · 8.532 Zit.
High-performance medicine: the convergence of human and artificial intelligence
2018 · 8.046 Zit.
BioBERT: a pre-trained biomedical language representation model for biomedical text mining
2019 · 6.843 Zit.
Proceedings of the 19th International Joint Conference on Artificial Intelligence
2005 · 5.781 Zit.