Dies ist eine Übersichtsseite mit Metadaten zu dieser wissenschaftlichen Arbeit. Der vollständige Artikel ist beim Verlag verfügbar.
P714: Enhancing laboratory efficiency: Implementation of Revvity Transcribe AI for automated data entry in newborn screening
0
Zitationen
9
Autoren
2025
Jahr
Abstract
20 years of experience.A majority of the respondents reported interpreting more than 100 variants per year (71.6%), with 26.8% reporting reinterpreting more than 26 variants per year.80.5% of responders reported that insufficient data was frequently the cause for classifying variants as VUS.Overall, 78.4% of respondents reported engaging with functional evidence and most (77.4%)evaluating functional data for variant classification in the clinical setting.More than half of the respondents reported functional data was rarely or never available for their variants of interest (67.4%).A majority (90.5%) of respondents considered insufficient quality metrics or confidence in the accuracy of functional data to be a barrier to using functional evidence.When asked about options of interventions that could improve use of functional evidence, almost half of the respondents noted access to primary functional data (42.6%), and standardized interpretation of functional data through ClinVar (45.8%) would significantly improve their use.Conclusion: Overall, the results of this study provided insights on the functional evidence usage for individuals involved in variant interpretation.It also showed a demand for a robust database that provides more accessible information on the quality and accuracy of functional evidence to promote use of functional evidence in variant interpretation.
Ähnliche Arbeiten
Explainable Artificial Intelligence (XAI): Concepts, taxonomies, opportunities and challenges toward responsible AI
2019 · 8.663 Zit.
Stop explaining black box machine learning models for high stakes decisions and use interpretable models instead
2019 · 8.576 Zit.
High-performance medicine: the convergence of human and artificial intelligence
2018 · 8.091 Zit.
BioBERT: a pre-trained biomedical language representation model for biomedical text mining
2019 · 6.859 Zit.
Proceedings of the 19th International Joint Conference on Artificial Intelligence
2005 · 5.781 Zit.