Searched for: in-biosketch:true
person:genesn01
Early Insights Among Emergency Medicine Physicians on Artificial Intelligence: A National, Convenience-sample Survey of the American College of Emergency Physicians
Shy, Bradley D; Baloescu, Cristiana; Faustino, Isaac V; Taylor, R Andrew; Gottlieb, Michael; Sangal, Rohit B; Hood, Colton; Genes, Nicholas; Rabin, Elaine J; ,
OBJECTIVES/UNASSIGNED:This study aimed to assess the current utilization of artificial intelligence (AI) tools among emergency physicians, their attitudes toward AI in clinical practice, and how a national physician professional organization could best support its members regarding AI. METHODS/UNASSIGNED:A cross-sectional survey was emailed to American College of Emergency Physicians members and made available to conference attendees at a national symposium. The survey collected demographic information, details on the use of noninstitutional and institutional AI tools, attitudes toward AI, and desired forms of support. Descriptive statistics were used to summarize the data. RESULTS/UNASSIGNED:A total of 658 physicians responded, primarily practicing attendings (78%) and residents (9%), with 60% aged 35 to 54 years and 67% identifying as male; these respondents represented 2% of the membership of American College of Emergency Physicians. Noninstitutional AI tool use (eg, ChatGPT and independent electrocardiogram interpretation) was reported by 31% of respondents. Institutional AI was integrated into 52% of respondents' practices, with 18% regularly using ambient AI documentation and 22% using AI-assisted clinical decision support. AI tools for point-of-care ultrasound were available to 10%, and AI-assisted radiology interpretation was used by 14%, mainly for X-rays and computed tomography. Operational AI for triage, capacity management, and staff optimization were reported by 15%, while 9% used AI-assisted coding and billing. Moreover, 75% believed AI improves clinical efficiency, 57% felt it enhanced care quality, but 12% expressed concern about job displacement, 16% are unsure whether AI tools would adequately comply with Health Insurance Portability and Accountability Act regulations, and 38% noted potential biases. About half desire educational support and guidelines. CONCLUSION/UNASSIGNED:In this nonrepresentative emergency physician survey, respondents reported moderate rates of adoption of AI tools and generally positive attitudes toward AI's impact on efficiency and care quality. However, respondents also reported important concerns about job displacement, Health Insurance Portability and Accountability Act regulation compliance, and potential biases. Larger studies are needed to more fully understand emergency physician views on AI.
PMCID:12796722
PMID: 41536575
ISSN: 2688-1152
CID: 5986452
Understanding and Addressing Bias in Artificial Intelligence Systems: A Primer for the Emergency Medicine Physician
Abbott, Ethan E; Rehman, Tehreem; Rosania, Anthony; Lum, Donald L; Taylor, Todd B; Kirk, A J; Taylor, R Andrew; Baker, Eileen F; Rabin, Elaine; Padela, Aasim; Genes, Nicholas; Srivastava, Atul; Sangal, Rohit B; Apakama, Donald; ,
Artificial intelligence (AI) tools and technologies are increasingly being integrated into emergency medicine (EM) practice, not only offering potential benefits such as improved efficiency, better patient experience, and increased safety, but also resulting in potential risks including exacerbation of biases. These biases, inadvertently embedded in AI algorithms or training data, can adversely affect clinical decision making for diverse patient populations. Bias is a universal human attribute, subject to introduction into any human interaction. The risk with AI is magnification of, or even normalization of, patterns of biases across the health care ecosystem within tools that in time may be considered authoritative. This article, the work of members of the American College of Emergency Physicians (ACEP) AI Task Force, aims to equip emergency physicians (EPs) with a practical framework for understanding, identifying, and addressing bias in clinical and operational AI tools encountered in the emergency department (ED). For this publication, we have defined bias as a systematic flaw in a decision-making process that results in unfair or unintended outcomes that can be inadvertently embedded in AI algorithms or training data. This can result in adverse effects on clinical decision making for diverse patient populations. We begin by reviewing common sources of AI bias relevant to EM, including data, algorithmic, measurement, and human-interaction factors, and then, we discuss the potential pitfalls. Following this, we use illustrative examples from EM practice (eg, triage tools, risk stratification, and medical devices) to demonstrate how bias can manifest. We subsequently discuss the evolving regulatory landscape, structured assessment frameworks (including predeployment, continuous monitoring, and postdeployment steps), key principles (like sociotechnical perspectives and stakeholder engagement), and specific tools. Finally, this review outlines the EP's vital role in mitigation of AI-related biases through advocacy, local validation, clinical feedback, demanding transparency, and maintaining clinical judgment over automation.
PMCID:12797052
PMID: 41536573
ISSN: 2688-1152
CID: 5986442
Clinical decision making during supervised endotracheal intubations in academic emergency medicine
Offenbacher, Joseph; Kim, Jung G; Louie, Kenway; Patel, Savan; Genes, Nicholas; Smith, Silas W; Nikolla, Dhimitri A; Carlson, Jestin N; Gulati, Rajneesh; Sinha, Shreya; Sagalowsky, Selin T; Boatright, Dowin H; Glimcher, Paul
BACKGROUND:Endotracheal intubation in the emergency department (ED) is a critical and time-sensitive procedure requiring both technical skills and cognitive-based reasoning. Evidence on supervised resident-attending dyads with differing years of seniority on decision making during clinical encounters with endotracheal intubations is nascent. OBJECTIVE:To investigate the intersection of postgraduate years in clinical practice between resident and attending supervisor dyads and its associations for clinician choice of laryngoscopy technique and paralytic agent during ED intubations. METHODS:We conducted a retrospective analysis of intubations performed at a multi-site, urban, academic emergency medicine training program, analyzing institutional airway registry data from 2013 to 2023. Using a standardized predictor that accounted for similarity in years of clinical experience within a dyad between a resident and their supervising attending, we performed adjusted mixed-effects logistic regression examining the association of this dyad on two primary outcomes in endotracheal intubation decision making. Our primary outcome measures were the selection of a laryngoscopy technique (either DL or VL), and of a paralytic agent (either short-acting or long-acting) analyzed as categorical variables with a linear mixed effects model, using a binomial response distribution. RESULTS:We examined 2969 intubations for choice of laryngoscopy technique (n = 1117, 37.6 %) and paralytic agent (n = 967, 32.6 %). Higher adjusted odds (aOR) were associated with resident choice of DL over VL when years of experience between residents and supervising attendings were more concordant (aOR 3.05, 95 % CI: 1.1-8.2). Choice of paralytic agent was not associated with differing years of experience. CONCLUSION/CONCLUSIONS:Concordant years of experience between residents and their attendings were associated with technical skill-based laryngoscopy technique choice but not for cognitive-based reasoning in paralytic agent choice among ED intubations, suggesting supervising attending's years in clinical practice may influence decision making during time-sensitive procedures.
PMID: 41197425
ISSN: 1532-8171
CID: 5960122
Clinical Integration of Local AI Assistants Into Emergency Department Discharge Documentation
Silberlust, Jared; Genes, Nicholas
PMID: 41118169
ISSN: 2574-3805
CID: 5956772
Generative AI Summaries to Facilitate ED Handoff
Genes, Nicholas; Simon, Gregory; Koziatek, Christian; Kim, Jung G; Woo, Kar-Mun; Dahn, Cassidy; Chan, Leland; Wiesenfeld, Batia
Background Emergency Department (ED) handoff to inpatient teams is a potential source of error. Generative Artificial Intelligence (AI) has shown promise in succinctly summarizing large quantities of clinical data and may help improve Emergency Department (ED) handoff. Objectives Our objectives were to: 1) evaluate the accuracy, clinical utility, and safety of AI-generated ED-to-inpatient handoff summaries; 2) identify patient and visit characteristics influencing summary effectiveness; and 3) characterize potential error patterns to inform implementation strategies. Methods This exploratory study evaluated AI-generated handoff summaries at an urban academic ED (February-April 2024). A HIPAA-compliant GPT-4 model generated summaries aligned with the IPASS framework; ED providers assessed summary accuracy, usefulness, and safety through on-shift surveys. Results Among 50 cases, median quality and usefulness scores were 4/5 (SE = 0.13). Safety concerns arose in 6% of cases, with issues including data omissions and mischaracterizations. Consultation status significantly affected usefulness scores (p < 0.05). Omissions of relevant medications, laboratory results, and other essential detailss were noted (n=6), and EM clinicians disagreed with some AI characterizations of patient stability, vitals and workup (n=8). The most common response was positive impressions of the technology incorporated into the handoff process (n=11). Conclusions This exploratory provider-in-the-loop model demonstrated clinical acceptability and highlighted areas for refinement. Future studies should incorporate recipient perspectives and examine clinical outcomes to scale and optimize AI implementation.
PMID: 40795949
ISSN: 1869-0327
CID: 5907202
Real-World Clinical Impact of High-Sensitivity Troponin for Chest Pain Evaluation in the Emergency Department
Martin, Jacob A; Zhang, Robert S; Rhee, Aaron J; Saxena, Archana; Akindutire, Olumide; Maqsood, M Haisum; Genes, Nicholas; Gollogly, Nathan; Smilowitz, Nathaniel R; Quinones-Camacho, Adriana
BACKGROUND:High-sensitivity cardiac troponin (hs-cTnI) assays can quantify troponin concentrations with low limits of detection, potentially expediting and enhancing myocardial infarction diagnoses. This study investigates the real-world impact of hs-cTnI implementation on operational metrics and downstream cardiac services in patients presenting to the emergency department with chest pain. METHODS AND RESULTS/RESULTS:[lt] 0.001) during the index encounter. CONCLUSION/CONCLUSIONS:Implementation of the hs-cTnI assay was associated with reduced hospital admissions, shorter length of stay, and decreases in most downstream cardiac testing.
PMID: 40240953
ISSN: 2047-9980
CID: 5828482
Resident clinical dashboards to support precision education in emergency medicine
Moser, Joe-Ann S; Genes, Nicholas; Hekman, Daniel J; Krzyzaniak, Sara M; Layng, Timothy A; Miller, Danielle; Rider, Ashley C; Sagalowsky, Selin T; Smith, Moira E; Schnapp, Benjamin H
INTRODUCTION/UNASSIGNED:With the move toward competency-based medical education (CBME), data from the electronic health record (EHR) for informed self-improvement may be valuable as a part of programmatic assessment. Personalized dashboards are one way to view these clinical data. The purpose of this concept paper is to summarize the current state of clinical dashboards as they can be utilized by emergency medicine (EM) residency programs. METHODS/UNASSIGNED:The author group consisted of EM physicians from multiple institutions with medical education and informatics backgrounds and was identified by querying faculty presenting on resident clinical dashboards at the 2024 Society for Academic Emergency Medicine conference. Additional authors were identified by members of the initial group. Best practice literature was referenced; if none was available, group consensus was used. CATEGORIES OF METRICS/UNASSIGNED:Clinical exposures as well as efficiency, quality, documentation, and diversity metrics may be included in a resident dashboard. Resident dashboard metrics should focus on resident-sensitive measures rather than those primarily affected by attendings or systems-based factors. CONSIDERATIONS FOR IMPLEMENTATION/UNASSIGNED:Implementation of these dashboards requires the technical expertise to turn EHR data into actionable data, a process called EHR phenotyping. The dashboard can be housed directly in the EHR or on a separate platform. Dashboard developers should consider how their implementation plan will affect how often dashboard data will be refreshed and how to best display the data for ease of understanding. IMPLICATIONS FOR EDUCATION & TRAINING/UNASSIGNED:Dashboards can provide objective data to residents, residency leadership and clinical competency committees as they identify areas of strength, growth areas, and set specific and actionable goals. The success of resident dashboards is reliant on resident buy-in and creating a culture of psychological safety through thoughtful implementation, coaching, and regular feedback. . CONCLUSION/UNASSIGNED:Personalized clinical dashboards can play a crucial role in programmatic assessment within CBME, helping EM residents focus their efforts as they advance and refine their skills during training.
PMCID:12038736
PMID: 40308868
ISSN: 2472-5390
CID: 5834032
Palliative Care Initiated in the Emergency Department: A Cluster Randomized Clinical Trial
Grudzen, Corita R; Siman, Nina; Cuthel, Allison M; Adeyemi, Oluwaseun; Yamarik, Rebecca Liddicoat; Goldfeld, Keith S; ,; Abella, Benjamin S; Bellolio, Fernanda; Bourenane, Sorayah; Brody, Abraham A; Cameron-Comasco, Lauren; Chodosh, Joshua; Cooper, Julie J; Deutsch, Ashley L; Elie, Marie Carmelle; Elsayem, Ahmed; Fernandez, Rosemarie; Fleischer-Black, Jessica; Gang, Mauren; Genes, Nicholas; Goett, Rebecca; Heaton, Heather; Hill, Jacob; Horwitz, Leora; Isaacs, Eric; Jubanyik, Karen; Lamba, Sangeeta; Lawrence, Katharine; Lin, Michelle; Loprinzi-Brauer, Caitlin; Madsen, Troy; Miller, Joseph; Modrek, Ada; Otero, Ronny; Ouchi, Kei; Richardson, Christopher; Richardson, Lynne D; Ryan, Matthew; Schoenfeld, Elizabeth; Shaw, Matthew; Shreves, Ashley; Southerland, Lauren T; Tan, Audrey; Uspal, Julie; Venkat, Arvind; Walker, Laura; Wittman, Ian; Zimny, Erin
IMPORTANCE/UNASSIGNED:The emergency department (ED) offers an opportunity to initiate palliative care for older adults with serious, life-limiting illness. OBJECTIVE/UNASSIGNED:To assess the effect of a multicomponent intervention to initiate palliative care in the ED on hospital admission, subsequent health care use, and survival in older adults with serious, life-limiting illness. DESIGN, SETTING, AND PARTICIPANTS/UNASSIGNED:Cluster randomized, stepped-wedge, clinical trial including patients aged 66 years or older who visited 1 of 29 EDs across the US between May 1, 2018, and December 31, 2022, had 12 months of prior Medicare enrollment, and a Gagne comorbidity score greater than 6, representing a risk of short-term mortality greater than 30%. Nursing home patients were excluded. INTERVENTION/UNASSIGNED:A multicomponent intervention (the Primary Palliative Care for Emergency Medicine intervention) included (1) evidence-based multidisciplinary education; (2) simulation-based workshops on serious illness communication; (3) clinical decision support; and (4) audit and feedback for ED clinical staff. MAIN OUTCOME AND MEASURES/UNASSIGNED:The primary outcome was hospital admission. The secondary outcomes included subsequent health care use and survival at 6 months. RESULTS/UNASSIGNED:There were 98 922 initial ED visits during the study period (median age, 77 years [IQR, 71-84 years]; 50% were female; 13% were Black and 78% were White; and the median Gagne comorbidity score was 8 [IQR, 7-10]). The rate of hospital admission was 64.4% during the preintervention period vs 61.3% during the postintervention period (absolute difference, -3.1% [95% CI, -3.7% to -2.5%]; adjusted odds ratio [OR], 1.03 [95% CI, 0.93 to 1.14]). There was no difference in the secondary outcomes before vs after the intervention. The rate of admission to an intensive care unit was 7.8% during the preintervention period vs 6.7% during the postintervention period (adjusted OR, 0.98 [95% CI, 0.83 to 1.15]). The rate of at least 1 revisit to the ED was 34.2% during the preintervention period vs 32.2% during the postintervention period (adjusted OR, 1.00 [95% CI, 0.91 to 1.09]). The rate of hospice use was 17.7% during the preintervention period vs 17.2% during the postintervention period (adjusted OR, 1.04 [95% CI, 0.93 to 1.16]). The rate of home health use was 42.0% during the preintervention period vs 38.1% during the postintervention period (adjusted OR, 1.01 [95% CI, 0.92 to 1.10]). The rate of at least 1 hospital readmission was 41.0% during the preintervention period vs 36.6% during the postintervention period (adjusted OR, 1.01 [95% CI, 0.92 to 1.10]). The rate of death was 28.1% during the preintervention period vs 28.7% during the postintervention period (adjusted OR, 1.07 [95% CI, 0.98 to 1.18]). CONCLUSIONS AND RELEVANCE/UNASSIGNED:This multicomponent intervention to initiate palliative care in the ED did not have an effect on hospital admission, subsequent health care use, or short-term mortality in older adults with serious, life-limiting illness. TRIAL REGISTRATION/UNASSIGNED:ClinicalTrials.gov Identifier: NCT03424109.
PMID: 39813042
ISSN: 1538-3598
CID: 5776882
Addressing Note Bloat: Solutions for Effective Clinical Documentation
Genes, Nicholas; Sills, Joseph; Heaton, Heather A; Shy, Bradley D; Scofi, Jean
Clinical documentation in the United States has grown longer and more difficult to read, a phenomenon described as "note bloat." This issue is especially pronounced in emergency medicine, where high diagnostic uncertainty and brief evaluations demand focused, efficient chart review to inform decision-making. Note bloat arises from multiple factors: efforts to enhance billing, mitigate malpractice risk, and leverage electronic health record tools that improve speed and completeness. We discuss best practices based on available evidence and expert opinion to improve note clarity and concision. Recent E/M coding reforms aim to streamline documentation by prioritizing medical decision-making over details of historical and physical examination, though implementation varies. New technologies such as generative artificial intelligence present opportunities and challenges for documentation practices. Addressing note bloat will require ongoing effort from clinical leadership, electronic health record vendors, and professional organizations.
PMCID:11852943
PMID: 40012671
ISSN: 2688-1152
CID: 5801152
Evaluating Large Language Models in extracting cognitive exam dates and scores
Zhang, Hao; Jethani, Neil; Jones, Simon; Genes, Nicholas; Major, Vincent J; Jaffe, Ian S; Cardillo, Anthony B; Heilenbach, Noah; Ali, Nadia Fazal; Bonanni, Luke J; Clayburn, Andrew J; Khera, Zain; Sadler, Erica C; Prasad, Jaideep; Schlacter, Jamie; Liu, Kevin; Silva, Benjamin; Montgomery, Sophie; Kim, Eric J; Lester, Jacob; Hill, Theodore M; Avoricani, Alba; Chervonski, Ethan; Davydov, James; Small, William; Chakravartty, Eesha; Grover, Himanshu; Dodson, John A; Brody, Abraham A; Aphinyanaphongs, Yindalon; Masurkar, Arjun; Razavian, Narges
Ensuring reliability of Large Language Models (LLMs) in clinical tasks is crucial. Our study assesses two state-of-the-art LLMs (ChatGPT and LlaMA-2) for extracting clinical information, focusing on cognitive tests like MMSE and CDR. Our data consisted of 135,307 clinical notes (Jan 12th, 2010 to May 24th, 2023) mentioning MMSE, CDR, or MoCA. After applying inclusion criteria 34,465 notes remained, of which 765 underwent ChatGPT (GPT-4) and LlaMA-2, and 22 experts reviewed the responses. ChatGPT successfully extracted MMSE and CDR instances with dates from 742 notes. We used 20 notes for fine-tuning and training the reviewers. The remaining 722 were assigned to reviewers, with 309 each assigned to two reviewers simultaneously. Inter-rater-agreement (Fleiss' Kappa), precision, recall, true/false negative rates, and accuracy were calculated. Our study follows TRIPOD reporting guidelines for model validation. For MMSE information extraction, ChatGPT (vs. LlaMA-2) achieved accuracy of 83% (vs. 66.4%), sensitivity of 89.7% (vs. 69.9%), true-negative rates of 96% (vs 60.0%), and precision of 82.7% (vs 62.2%). For CDR the results were lower overall, with accuracy of 87.1% (vs. 74.5%), sensitivity of 84.3% (vs. 39.7%), true-negative rates of 99.8% (98.4%), and precision of 48.3% (vs. 16.1%). We qualitatively evaluated the MMSE errors of ChatGPT and LlaMA-2 on double-reviewed notes. LlaMA-2 errors included 27 cases of total hallucination, 19 cases of reporting other scores instead of MMSE, 25 missed scores, and 23 cases of reporting only the wrong date. In comparison, ChatGPT's errors included only 3 cases of total hallucination, 17 cases of wrong test reported instead of MMSE, and 19 cases of reporting a wrong date. In this diagnostic/prognostic study of ChatGPT and LlaMA-2 for extracting cognitive exam dates and scores from clinical notes, ChatGPT exhibited high accuracy, with better performance compared to LlaMA-2. The use of LLMs could benefit dementia research and clinical care, by identifying eligible patients for treatments initialization or clinical trial enrollments. Rigorous evaluation of LLMs is crucial to understanding their capabilities and limitations.
PMCID:11634005
PMID: 39661652
ISSN: 2767-3170
CID: 5762692