OpenAlex · Aktualisierung stündlich · Letzte Aktualisierung: 12.05.2026, 11:42

Dies ist eine Übersichtsseite mit Metadaten zu dieser wissenschaftlichen Arbeit. Der vollständige Artikel ist beim Verlag verfügbar.

<scp>ChatG</scp>‐<scp>PD</scp>? Comparing large language model artificial intelligence and faculty rankings of the competitiveness of standardized letters of evaluation

2024·3 Zitationen·AEM Education and TrainingOpen Access
Volltext beim Verlag öffnen

3

Zitationen

9

Autoren

2024

Jahr

Abstract

Background: While faculty have previously been shown to have high levels of agreement about the competitiveness of emergency medicine (EM) standardized letters of evaluation (SLOEs), reviewing SLOEs remains a highly time-intensive process for faculty. Artificial intelligence large language models (LLMs) have shown promise for effectively analyzing large volumes of data across a variety of contexts, but their ability to interpret SLOEs is unknown. Objective: The objective was to evaluate the ability of LLMs to rate EM SLOEs on competitiveness compared to faculty consensus and previously developed algorithms. Methods: Fifty mock SLOE letters were drafted and analyzed seven times by a data-focused LLM with instructions to rank them based on desirability for residency. The LLM was also asked to use its own criteria to decide which characteristics are most important for residency and revise its ranking of the SLOEs. LLM-generated rank lists were compared with faculty consensus rankings. Results: 0.86). Conclusions: The LLM generated rankings showed strong correlation with expert faculty consensus rankings with minimal input of faculty time and effort.

Ähnliche Arbeiten