Searched for: in-biosketch:yes
person:aphiny01
Methods and Impact for Using Federated Learning to Collaborate on Clinical Research
Cheung, Alexander T M; Nasir-Moin, Mustafa; Fred Kwon, Young Joon; Guan, Jiahui; Liu, Chris; Jiang, Lavender; Raimondo, Christian; Chotai, Silky; Chambless, Lola; Ahmad, Hasan S; Chauhan, Daksh; Yoon, Jang W; Hollon, Todd; Buch, Vivek; Kondziolka, Douglas; Chen, Dinah; Al-Aswad, Lama A; Aphinyanaphongs, Yindalon; Oermann, Eric Karl
BACKGROUND:The development of accurate machine learning algorithms requires sufficient quantities of diverse data. This poses a challenge in health care because of the sensitive and siloed nature of biomedical information. Decentralized algorithms through federated learning (FL) avoid data aggregation by instead distributing algorithms to the data before centrally updating one global model. OBJECTIVE:To establish a multicenter collaboration and assess the feasibility of using FL to train machine learning models for intracranial hemorrhage (ICH) detection without sharing data between sites. METHODS:Five neurosurgery departments across the United States collaborated to establish a federated network and train a convolutional neural network to detect ICH on computed tomography scans. The global FL model was benchmarked against a standard, centrally trained model using a held-out data set and was compared against locally trained models using site data. RESULTS:A federated network of practicing neurosurgeon scientists was successfully initiated to train a model for predicting ICH. The FL model achieved an area under the ROC curve of 0.9487 (95% CI 0.9471-0.9503) when predicting all subtypes of ICH compared with a benchmark (non-FL) area under the ROC curve of 0.9753 (95% CI 0.9742-0.9764), although performance varied by subtype. The FL model consistently achieved top three performance when validated on any site's data, suggesting improved generalizability. A qualitative survey described the experience of participants in the federated network. CONCLUSION/CONCLUSIONS:This study demonstrates the feasibility of implementing a federated network for multi-institutional collaboration among clinicians and using FL to conduct machine learning research, thereby opening a new paradigm for neurosurgical collaboration.
PMID: 36399428
ISSN: 1524-4040
CID: 5385002
Ten Years of Health Informatics Education for Physicians
Chapter by: Major, Vincent J.; Plottel, Claudia S.; Aphinyanaphongs, Yindalon
in: Proceedings - 2023 IEEE 11th International Conference on Healthcare Informatics, ICHI 2023 by
[S.l.] : Institute of Electrical and Electronics Engineers Inc., 2023
pp. 637-644
ISBN: 9798350302639
CID: 5630952
Enabling AI-Augmented Clinical Workflows by Accessing Patient Data in Real-Time with FHIR
Chapter by: Major, Vincent J.; Wang, Walter; Aphinyanaphongs, Yindalon
in: Proceedings - 2023 IEEE 11th International Conference on Healthcare Informatics, ICHI 2023 by
[S.l.] : Institute of Electrical and Electronics Engineers Inc., 2023
pp. 531-533
ISBN: 9798350302639
CID: 5630942
AI model transferability in healthcare: a sociotechnical perspective
Wiesenfeld, Batia Mishan; Aphinyanaphongs, Yin; Nov, Oded
SCOPUS:85139986644
ISSN: 2522-5839
CID: 5350312
Predicting Post-Operative C. difficile Infection (CDI) With Automated Machine Learning (AutoML) Algorithms Using the American College of Surgeons National Surgical Quality Improvement Program (ACS NSQIP) Database [Meeting Abstract]
Thangirala, A; Li, T; Abaza, E; Aphinyanaphongs, Y; Axelrad, J; Chen, J; Kelleher, A; Oeding, J; Hu, E; Martin, J; Katz, G; Brejt, S; Castillo, G; Ostberg, N; Kan, K
Introduction: Clostridium difficile infection (CDI) is one of the most common hospital-acquired infections leading to prolonged hospitalization and significant morbidity. Only a few prior studies have developed predictive risk models for CDI and all but one have utilized logistic regression (LR) models to identify risk factors. Automated machine learning (AutoML) programs consistently outperform standard LR models in non-medical contexts. This study aims to investigate the utility of AutoML methods in developing a model for post-operative CDI prediction.
Method(s): We used an AutoML system developed by Amazon, Autogluon v0.3.1, to evaluate the prediction accuracy of post-surgical CDI using the 2016-2018 ACS NSQIP database. A total of A total of 3,049,617 patients and 79 pre-operative features were included in the model. Post-operative CDI was defined as CDI within 30 days of surgery. Models were trained for 4 hours to optimize performance on the Brier score, with lower being better. Validation of all performance metrics was done using the 2019 NSQIP database.
Result(s): 0.36% of the patients (n = 11,001) developed post-operative CDI. Brier scores were calculated for each model with the top performing model being an ensembled neural net model having a Brier score of 0.0027 on the test set. The corresponding AUROC and AUC-PR was 0.840 and 0.015 respectively (Figure).
Conclusion(s): The models generated via AutoML to predict post-operative CDI had discriminatory characteristics greater than or equal to those models reported in the literature. Future post-operative CDI models may benefit from automated machine learning techniques
EMBASE:641287886
ISSN: 1572-0241
CID: 5514802
Development and Validation of a Machine Learning Model for Automated Assessment of Resident Clinical Reasoning Documentation
Schaye, Verity; Guzman, Benedict; Burk-Rafel, Jesse; Marin, Marina; Reinstein, Ilan; Kudlowitz, David; Miller, Louis; Chun, Jonathan; Aphinyanaphongs, Yindalon
BACKGROUND:Residents receive infrequent feedback on their clinical reasoning (CR) documentation. While machine learning (ML) and natural language processing (NLP) have been used to assess CR documentation in standardized cases, no studies have described similar use in the clinical environment. OBJECTIVE:The authors developed and validated using Kane's framework a ML model for automated assessment of CR documentation quality in residents' admission notes. DESIGN, PARTICIPANTS, MAIN MEASURES/UNASSIGNED:Internal medicine residents' and subspecialty fellows' admission notes at one medical center from July 2014 to March 2020 were extracted from the electronic health record. Using a validated CR documentation rubric, the authors rated 414 notes for the ML development dataset. Notes were truncated to isolate the relevant portion; an NLP software (cTAKES) extracted disease/disorder named entities and human review generated CR terms. The final model had three input variables and classified notes as demonstrating low- or high-quality CR documentation. The ML model was applied to a retrospective dataset (9591 notes) for human validation and data analysis. Reliability between human and ML ratings was assessed on 205 of these notes with Cohen's kappa. CR documentation quality by post-graduate year (PGY) was evaluated by the Mantel-Haenszel test of trend. KEY RESULTS/RESULTS:The top-performing logistic regression model had an area under the receiver operating characteristic curve of 0.88, a positive predictive value of 0.68, and an accuracy of 0.79. Cohen's kappa was 0.67. Of the 9591 notes, 31.1% demonstrated high-quality CR documentation; quality increased from 27.0% (PGY1) to 31.0% (PGY2) to 39.0% (PGY3) (p < .001 for trend). Validity evidence was collected in each domain of Kane's framework (scoring, generalization, extrapolation, and implications). CONCLUSIONS:The authors developed and validated a high-performing ML model that classifies CR documentation quality in resident admission notes in the clinical environment-a novel application of ML and NLP with many potential use cases.
PMCID:9296753
PMID: 35710676
ISSN: 1525-1497
CID: 5277902
Evaluating the Effect of a COVID-19 Predictive Model to Facilitate Discharge: A Randomized Controlled Trial
Major, Vincent J; Jones, Simon A; Razavian, Narges; Bagheri, Ashley; Mendoza, Felicia; Stadelman, Jay; Horwitz, Leora I; Austrian, Jonathan; Aphinyanaphongs, Yindalon
BACKGROUND: We previously developed and validated a predictive model to help clinicians identify hospitalized adults with coronavirus disease 2019 (COVID-19) who may be ready for discharge given their low risk of adverse events. Whether this algorithm can prompt more timely discharge for stable patients in practice is unknown. OBJECTIVES/OBJECTIVE: The aim of the study is to estimate the effect of displaying risk scores on length of stay (LOS). METHODS: We integrated model output into the electronic health record (EHR) at four hospitals in one health system by displaying a green/orange/red score indicating low/moderate/high-risk in a patient list column and a larger COVID-19 summary report visible for each patient. Display of the score was pseudo-randomized 1:1 into intervention and control arms using a patient identifier passed to the model execution code. Intervention effect was assessed by comparing LOS between intervention and control groups. Adverse safety outcomes of death, hospice, and re-presentation were tested separately and as a composite indicator. We tracked adoption and sustained use through daily counts of score displays. RESULTS: Enrolling 1,010 patients from May 15, 2020 to December 7, 2020, the trial found no detectable difference in LOS. The intervention had no impact on safety indicators of death, hospice or re-presentation after discharge. The scores were displayed consistently throughout the study period but the study lacks a causally linked process measure of provider actions based on the score. Secondary analysis revealed complex dynamics in LOS temporally, by primary symptom, and hospital location. CONCLUSION/CONCLUSIONS: An AI-based COVID-19 risk score displayed passively to clinicians during routine care of hospitalized adults with COVID-19 was safe but had no detectable impact on LOS. Health technology challenges such as insufficient adoption, nonuniform use, and provider trust compounded with temporal factors of the COVID-19 pandemic may have contributed to the null result. TRIAL REGISTRATION/BACKGROUND: ClinicalTrials.gov identifier: NCT04570488.
PMCID:9329139
PMID: 35896506
ISSN: 1869-0327
CID: 5276672
Automated interpretable discovery of heterogeneous treatment effectiveness: A COVID-19 case study
Lengerich, Benjamin J; Nunnally, Mark E; Aphinyanaphongs, Yin; Ellington, Caleb; Caruana, Rich
Testing multiple treatments for heterogeneous (varying) effectiveness with respect to many underlying risk factors requires many pairwise tests; we would like to instead automatically discover and visualize patient archetypes and predictors of treatment effectiveness using multitask machine learning. In this paper, we present a method to estimate these heterogeneous treatment effects with an interpretable hierarchical framework that uses additive models to visualize expected treatment benefits as a function of patient factors (identifying personalized treatment benefits) and concurrent treatments (identifying combinatorial treatment benefits). This method achieves state-of-the-art predictive power for COVID-19 in-hospital mortality and interpretable identification of heterogeneous treatment benefits. We first validate this method on the large public MIMIC-IV dataset of ICU patients to test recovery of heterogeneous treatment effects. Next we apply this method to a proprietary dataset of over 3000 patients hospitalized for COVID-19, and find evidence of heterogeneous treatment effectiveness predicted largely by indicators of inflammation and thrombosis risk: patients with few indicators of thrombosis risk benefit most from treatments against inflammation, while patients with few indicators of inflammation risk benefit most from treatments against thrombosis. This approach provides an automated methodology to discover heterogeneous and individualized effectiveness of treatments.
PMCID:9055753
PMID: 35504543
ISSN: 1532-0480
CID: 5216082
Open-Source Clinical Machine Learning Models: Critical Appraisal of Feasibility, Advantages, and Challenges
Harish, Keerthi B; Price, W Nicholson; Aphinyanaphongs, Yindalon
Machine learning applications promise to augment clinical capabilities and at least 64 models have already been approved by the US Food and Drug Administration. These tools are developed, shared, and used in an environment in which regulations and market forces remain immature. An important consideration when evaluating this environment is the introduction of open-source solutions in which innovations are freely shared; such solutions have long been a facet of digital culture. We discuss the feasibility and implications of open-source machine learning in a health care infrastructure built upon proprietary information. The decreased cost of development as compared to drugs and devices, a longstanding culture of open-source products in other industries, and the beginnings of machine learning-friendly regulatory pathways together allow for the development and deployment of open-source machine learning models. Such tools have distinct advantages including enhanced product integrity, customizability, and lower cost, leading to increased access. However, significant questions regarding engineering concerns about implementation infrastructure and model safety, a lack of incentives from intellectual property protection, and nebulous liability rules significantly complicate the ability to develop such open-source models. Ultimately, the reconciliation of open-source machine learning and the proprietary information-driven health care environment requires that policymakers, regulators, and health care organizations actively craft a conducive market in which innovative developers will continue to both work and collaborate.
PMCID:9039816
PMID: 35404258
ISSN: 2561-326x
CID: 5218922
PREDICTION OF PERIOPERATIVE MAJOR ADVERSE CARDIOVASCULAR AND CEREBROVASCULAR EVENTS (MACCE) USING AUTOMATED MACHINE LEARNING (AUTOML) ALGORITHMS WITH GOOGLE AUTOML TABLES (GAMLT) USING THE AMERICAN COLLEGE OF SURGEONS NATIONAL SURGICAL QUALITY IMPROVEMENT PROGRAM (ACS NSQIP) DATABASE [Meeting Abstract]
Thangirala, A; Aphinyanaphongs, Y; Kan, K; Talmor, N; Smilowitz, N R
Background: Risk calculators to predict perioperative major adverse cardiovascular and cerebrovascular events (MACCE) often rely on logistic regression (LR) analysis. Automated machine learning (AutoML) processes regularly outperform regular machine learning (ML) and LR methods for predictive accuracy. Commercial AutoML systems have not yet been applied to predict perioperative MACCE after non-cardiac and cardiac surgeries.
Method(s): We used a commercial AutoML system, Google AutoML Tables (GAMLT), to predict perioperative MACCE in the 2019 ACS NSQIP database. MACCE was defined as death, myocardial infarction, cardiac arrest, or stroke. Default AutoML settings were used, with 80% of cases randomly selected for training, 10% for validation, and 10% for testing. Global feature importance was determined through the Shapley method. Two models were generated: Model 1 included 81 pre-operative features; Model 2 included the top 21 features from Model 1 and was independently validated with 2016-2018 NSQIP data.
Result(s): Model 1 yielded an area under the receiver operating characteristic (AUROC) of 0.934 in the 2019 ACS NSQIP dataset. Model 2, an ensemble of 25 feedforward neural net models, yielded an AUROC of 0.914-0.920 for MACCE in 2016-2019 (Figure).
Conclusion(s): Compared to existing risk calculators, GAMLT-derived models offered novel feature detection and comparable predictive performance for MACCE. AutoML analyses should be considered for risk estimation of perioperative MACCE. [Formula presented]
Copyright
EMBASE:2017304686
ISSN: 1558-3597
CID: 5184022