Try a new search

Format these results:

Searched for:

in-biosketch:yes

person:sodicd01

Total Results:

234


First-Order Spatial Encoding Simulations for Improved Accuracy in the Presence of Strong B0 and Gradient Field Variations

Tibrewala, Radhika; Collins, Christopher M; Mallett, Michael; Vom Endt, Axel; Sodickson, Daniel K; Assländer, Jakob
PURPOSE/OBJECTIVE: METHODS:Like many other MRI simulators, ours discretizes magnetic fields in space. However, we extend the MR signal simulation at each grid point from the 0th-order approximation, which assumes piecewise constant fields, to a 1st-order approximation, which assumes piecewise linear fields. We solve the signal equation by analytically integrating over each grid cube, assuming linear field variations, and then summing over all cubes. We provide analytical integrals for several pulse sequences. RESULTS:The 1st-order approximation captures strongly varying fields and associated intravoxel dephasing more accurately, avoiding severe "ringing" artifacts present in the usual 0th-order simulations. This enables simulations on a much coarser grid, facilitating computational feasibility. CONCLUSION/CONCLUSIONS:The first-order simulator enables the evaluation of unconventional scanner designs with strongly varying magnetic fields.
PMID: 41145956
ISSN: 1522-2594
CID: 5961042

Leveraging Representation Learning for Bi-parametric Prostate MRI to Disambiguate PI-RADS 3 and Improve Biopsy Decision Strategies

Umapathy, Lavanya; Johnson, Patricia M; Dutt, Tarun; Tong, Angela; Chopra, Sumit; Sodickson, Daniel K; Chandarana, Hersh
OBJECTIVES/OBJECTIVE:Despite its high negative predictive value (NPV) for clinically significant prostate cancer (csPCa), MRI suffers from a substantial number of false positives, especially for intermediate-risk cases. In this work, we determine whether a deep learning model trained with PI-RADS-guided representation learning can disambiguate the PI-RADS 3 classification, detect csPCa from bi-parametric prostate MR images, and avoid unnecessary benign biopsies. MATERIALS AND METHODS/METHODS:This study included 28,263 MR examinations and radiology reports from 21,938 men imaged for known or suspected prostate cancer between 2015 and 2023 at our institution (21 imaging locations with 34 readers), with 6352 subsequent biopsies. We trained a deep learning model, a representation learner (RL), to learn how radiologists interpret conventionally acquired T2-weighted and diffusion-weighted MR images, using exams in which the radiologists are confident in their risk assessments (PI-RADS 1 and 2 for the absence of csPCa vs. PI-RADS 4 and 5 for the presence of csPCa, n=21,465). We then trained biopsy-decision models to detect csPCa (Gleason score ≥7) using these learned image representations, and compared them to the performance of radiologists, and of models trained on other clinical variables (age, prostate volume, PSA, and PSA density) for treatment-naïve test cohorts consisting of only PI-RADS 3 (n=253, csPCa=103) and all PI-RADS (n=531, csPCa=300) cases. RESULTS:On the 2 test cohorts (PI-RADS-3-only, all-PI-RADS), RL-based biopsy-decision models consistently yielded higher AUCs in detecting csPCa (AUC=0.73 [0.66, 0.79], 0.88 [0.85, 0.91]) compared with radiologists (equivocal, AUC=0.79 [0.75, 0.83]) and the clinical model (AUCs=0.69 [0.62, 0.75], 0.78 [0.74, 0.82]). In the PIRADS-3-only cohort, all of whom would be biopsied using our institution's standard of care, the RL decision model avoided 41% (62/150) of benign biopsies compared with the clinical model (26%, P<0.001), and improved biopsy yield by 10% compared with the PI-RADS ≥3 decision strategy (0.50 vs. 0.40). Furthermore, on the all-PI-RADS cohort, RL decision model avoided 27% of additional benign biopsies (138/231) compared to radiologists (33%, P<0.001) with comparable sensitivity (93% vs. 92%), higher NPV (0.87 vs. 0.77), and biopsy yield (0.75 vs. 0.64). The combination of clinical and RL decision models further avoided benign biopsies (46% in PI-RADS-3-only and 62% in all-PI-RADS) while improving NPV (0.82, 0.88) and biopsy yields (0.52, 0.76) across the 2 test cohorts. CONCLUSIONS:Our PI-RADS-guided deep learning RL model learns summary representations from bi-parametric prostate MR images that can provide additional information to disambiguate intermediate-risk PI-RADS 3 assessments. The resulting RL-based biopsy decision models also outperformed radiologists in avoiding benign biopsies while maintaining comparable sensitivity to csPCa for the all-PI-RADS cohort. Such AI models can easily be integrated into clinical practice to supplement radiologists' reads in general and improve biopsy yield for any equivocal decisions.
PMID: 40586610
ISSN: 1536-0210
CID: 5887552

Prehospital real-time AI for trauma mortality prediction: a multi-institutional and multi-national validation study

Oh, Na-Eun; Oh, Thomas Young-Chul; Hsu, Jeremy; Kim, Do Wan; Yu, Byungchul; Cho, Jayun; Seok, Junepill; Lee, Jin Young; Jang, Woocheol; Kim, Jina; Femia, Robert J; Testa, Paul A; Yon, Dong Keon; Sodickson, Daniel K; Kang, Wu Seong; Lee, Jinseok
Early identification of high-risk trauma patients in the prehospital setting is crucial for optimizing resource allocation and improving survival. We developed and externally validated a real-time AI model predicting emergency room mortality using 21 prehospital variables. Model development and internal validation utilized the Korean Trauma Data Bank (KTDB; 204,189 patients), and external validation included four South Korean trauma centers (8,358 patients) and one Australian Level 1 center (3,578 patients). Our Prehospital-AI model, an ensemble of XGBoost, LightGBM, and random forest, achieved an AUROC of 0.923 (sensitivity: 0.780, specificity: 0.880) on the test set, outperforming the shock index (AUROC: 0.712). External validation yielded AUROCs of 0.925-0.956 across South Korean centers and 0.895 in the Australian center. Here we show that the Prehospital-AI model enables accurate, real-time risk assessment in the prehospital setting, outperforming traditional triage tools and improving trauma system efficiency. Nonetheless, additional multinational studies are warranted to further evaluate its generalizability across diverse trauma care systems.
PMID: 41501064
ISSN: 2041-1723
CID: 5981072

Artificial Intelligence in Prostate MRI: Addressing Current Limitations Through Emerging Technologies

Johnson, Patricia M; Umapathy, Lavanya; Gigax, Bradley; Rossi, Juan Kochen; Tong, Angela; Bruno, Mary; Sodickson, Daniel K; Nayan, Madhur; Chandarana, Hersh
Prostate MRI has transformed lesion detection and risk stratification in prostate cancer, but its impact is constrained by the high cost of the exam, variability in interpretation, and limited scalability. False negatives, false positives, and moderate inter-reader agreement undermine reliability, while long acquisition times restrict throughput. Artificial intelligence (AI) offers potential solutions to address many of the limitations of prostate MRI in the clinical management pathway. Machine learning-based triage can refine patient selection to optimize resources. Deep learning reconstruction enables accelerated acquisition while preserving diagnostic quality, with multiple FDA-cleared products now in clinical use. Ongoing development of automated quality assessment and artifact correction aims to improve reliability by reducing nondiagnostic exams. In image interpretation, AI models for lesion detection and clinically significant prostate cancer prediction achieve performance comparable to radiologists, and the PI-CAI international reader study has provided the strongest evidence to date of non-inferiority at scale. More recent work extends MRI-derived features into prognostic modeling of recurrence, metastasis, and functional outcomes. This review synthesizes progress across five domains-triage, accelerated acquisition and reconstruction, image quality assurance, diagnosis, and prognosis-highlighting the level of evidence, validation status, and barriers to adoption. While acquisition and reconstruction are furthest along, with FDA-cleared tools and prospective evaluations, triage, quality control, and prognosis remain earlier in development. Ensuring equitable performance across populations, incorporating uncertainty estimation, and conducting prospective workflow trials will be essential to move from promising prototypes to routine practice. Ultimately, AI could accelerate the adoption of prostate MRI toward a scalable platform for earlier detection and population-level prostate cancer management. EVIDENCE LEVEL: N/A TECHNICAL EFFICACY: 3.
PMID: 41348934
ISSN: 1522-2586
CID: 5975352

Overcoming MRI accessibility barriers in cancer imaging with cutting-edge solutions [Editorial]

Chandarana, Hersh; Sodickson, Daniel K
PMCID:12599087
PMID: 41214743
ISSN: 1470-7330
CID: 5966572

Multisession Longitudinal Dynamic MRI Incorporating Patient-Specific Prior Image Information Across Time

Chen, Jingjia; Chandarana, Hersh; Sodickson, Daniel K; Feng, Li
Serial Magnetic Resonance Imaging (MRI) exams are often performed in clinical practice, offering shared anatomical and motion information across imaging sessions. However, existing reconstruction methods process each session independently without leveraging this valuable longitudinal information. In this work, we propose a novel concept of longitudinal dynamic MRI, which incorporates patient-specific prior images to exploit temporal correlations across sessions. This framework enables progressive acceleration of data acquisition and reduction of scan time as more imaging sessions become available. The concept is demonstrated using the 4D Golden-angle RAdial Sparse Parallel (GRASP) MRI, a state-of-the-art dynamic imaging technique. Longitudinal reconstruction is performed by concatenating multi-session time-resolved 4D GRASP datasets into an extended dynamic series, followed by a low-rank subspace-based reconstruction algorithm. A series of experiments were conducted to evaluate the feasibility and performance of the proposed method. Results show that longitudinal 4D GRASP reconstruction consistently outperforms standard single-session reconstruction in image quality, while preserving inter-session variations. The approach demonstrated robustness to changes in anatomy, imaging intervals, and body contour, highlighting its potential for improving imaging efficiency and consistency in longitudinal MRI applications. More generally, this work suggests a new context-aware imaging paradigm in which the more we see a patient, the faster we can image.
PMCID:12310133
PMID: 40740507
ISSN: 2331-8422
CID: 5981862

T1 Relaxation-Enhanced Steady-State Acquisition With Radial k-Space Sampling: A Novel Family of Pulse Sequences for Motion-Robust Volumetric T1-Weighted MRI With Improved Lesion Conspicuity

Zi, Ruoxun; Edelman, Robert R; Maier, Christoph; Keerthivasan, Mahesh; Lattanzi, Riccardo; Sodickson, Daniel K; Block, Kai Tobias
OBJECTIVES/OBJECTIVE:Magnetization-prepared rapid gradient-echo (MP-RAGE) sequences are routinely acquired for brain exams, providing high conspicuity for enhancing lesions. Vessels, however, also appear bright, which can complicate the detection of small lesions. T1RESS (T1 relaxation-enhanced steady-state) sequences have been proposed as an alternative to MP-RAGE, offering improved lesion conspicuity and suppression of blood vessels. This work aims to evaluate the performance of radial T1RESS variants for motion-robust contrast-enhanced brain MRI. MATERIALS AND METHODS/METHODS:Radial stack-of-stars sampling was implemented for steady-state free-precession-based rapid T1RESS acquisition with saturation recovery preparation. Three variants were developed using a balanced steady-state free-precession readout (bT1RESS), an unbalanced fast imaging steady precession (FISP) readout (uT1RESS-FISP), and an unbalanced reversed FISP readout (uT1RESS-PSIF). Image contrast was evaluated in numerical simulations and phantom experiments. The motion robustness of radial T1RESS was demonstrated with a motion phantom. Four patients and six healthy volunteers were scanned at 3 T and 0.55 T. Extensions were developed combining T1RESS with GRASP for dynamic imaging, with GRAPPA for accelerated scans, and with Dixon for fat/water separation. RESULTS:In simulations and phantom scans, uT1RESS-FISP provided higher signal intensity for regions with lower T1 values (<500 ms) compared with MP-RAGE. In motion experiments, radial uT1RESS-FISP showed fewer artifacts than MP-RAGE and Cartesian uT1RESS-FISP. In patients, both unbalanced uT1RESS variants provided higher lesion conspicuity than MP-RAGE. Blood vessels appeared bright with MP-RAGE, gray with uT1RESS-FISP, and dark with uT1RESS-PSIF. At 0.55 T, bT1RESS provided high signal-to-noise ratio T1-weighted images without banding artifacts. Lastly, dynamic T1RESS images with a temporal resolution of 10.14 seconds/frame were generated using the GRASP algorithm. CONCLUSIONS:Radial T1RESS sequences offer improved lesion conspicuity and motion robustness and enable dynamic imaging for contrast-enhanced brain MRI. Both uT1RESS variants showed higher tumor-to-brain contrast than MP-RAGE and may find application as alternative techniques for imaging uncooperative patients with small brain lesions.
PMID: 40184541
ISSN: 1536-0210
CID: 5819432

Prostate Cancer Risk Stratification and Scan Tailoring Using Deep Learning on Abbreviated Prostate MRI

Johnson, Patricia M; Dutt, Tarun; Ginocchio, Luke A; Saimbhi, Amanpreet Singh; Umapathy, Lavanya; Block, Kai Tobias; Sodickson, Daniel K; Chopra, Sumit; Tong, Angela; Chandarana, Hersh
BACKGROUND:MRI plays a critical role in prostate cancer (PCa) detection and management. Bi-parametric MRI (bpMRI) offers a faster, contrast-free alternative to multi-parametric MRI (mpMRI). Routine use of mpMRI for all patients may not be necessary, and a tailored imaging approach (bpMRI or mpMRI) based on individual risk might optimize resource utilization. PURPOSE/OBJECTIVE:To develop and evaluate a deep learning (DL) model for classifying clinically significant PCa (csPCa) using bpMRI and to assess its potential for optimizing MRI protocol selection by recommending the additional sequences of mpMRI only when beneficial. STUDY TYPE/METHODS:Retrospective and prospective. POPULATION/METHODS:The DL model was trained and validated on 26,129 prostate MRI studies. A retrospective cohort of 151 patients (mean age 65 ± 8) with ground-truth verification from biopsy, prostatectomy, or long-term follow-up, alongside a prospective cohort of 142 treatment-naïve patients (mean age 65 ± 9) undergoing bpMRI, was evaluated. FIELD STRENGTH/SEQUENCE/UNASSIGNED:3 T, Turbo-spin echo T2-weighted imaging (T2WI) and single shot EPI diffusion-weighted imaging (DWI). ASSESSMENT/RESULTS:The DL model, based on a 3D ResNet-50 architecture, classified csPCa using PI-RADS ≥ 3 and Gleason ≥ 7 as outcome measures. The model was evaluated on a prospective cohort labeled by consensus of three radiologists and a retrospective cohort with ground truth verification based on biopsy or long-term follow-up. Real-time inference was tested on an automated MRI workflow, providing classification results directly at the scanner. STATISTICAL TESTS/METHODS:AUROC with 95% confidence intervals (CI) was used to evaluate model performance. RESULTS:In the prospective cohort, the model achieved an AUC of 0.83 (95% CI: 0.77-0.89) for PI-RADS ≥ 3 classification, with 93% sensitivity and 54% specificity. In the retrospective cohort, the model achieved an AUC of 0.86 (95% CI: 0.80-0.91) for Gleason ≥ 7 classification, with 93% sensitivity and 62% specificity. Real-time implementation demonstrated a processing latency of 14-16 s for protocol recommendations. DATA CONCLUSION/CONCLUSIONS:The proposed DL model identifies csPCa using bpMRI and integrates it into clinical workflows. EVIDENCE LEVEL/METHODS:1. TECHNICAL EFFICACY/UNASSIGNED:Stage 2.
PMID: 40259798
ISSN: 1522-2586
CID: 5830062

Morphological Brain Analysis Using Ultra Low-Field MRI

Hsu, Peter; Marchetto, Elisa; Sodickson, Daniel K; Johnson, Patricia M; Veraart, Jelle
Ultra low-field (ULF) MRI is an accessible neuroimaging modality that can bridge healthcare disparities and advance population-level brain health research. However, the inherently low signal-to-noise ratio of ULF-MRI often necessitates reductions in spatial resolution and, combined with the field-dependency of MRI contrast, challenges the accurate extraction of clinically relevant brain morphology. We evaluate the current state of ULF-MRI brain volumetry utilizing techniques for enhancing spatial resolution and leveraging recent advancements in brain segmentation. This is based on the agreement between ULF and corresponding high-field (HF) MRI brain volumes, and test-retest repeatability for multiple ULF scans. In this study, we find that accurate brain volumes can be measured from ULF-MRIs when combining orthogonal imaging directions for T2-weighted images to form a higher resolution image volume. We also demonstrate that not all orthogonal imaging directions contribute equally to volumetric accuracy and provide a recommended scan protocol given the constraints of the current technology.
PMCID:12207323
PMID: 40586128
ISSN: 1097-0193
CID: 5887542

Multimodal generative AI for interpreting 3D medical images and videos

Lee, Jung-Oh; Zhou, Hong-Yu; Berzin, Tyler M; Sodickson, Daniel K; Rajpurkar, Pranav
This perspective proposes adapting video-text generative AI to 3D medical imaging (CT/MRI) and medical videos (endoscopy/laparoscopy) by treating 3D images as videos. The approach leverages modern video models to analyze multiple sequences simultaneously and provide real-time AI assistance during procedures. The paper examines medical imaging's unique characteristics (synergistic information, metadata, and world model), outlines applications in automated reporting, case retrieval, and education, and addresses challenges of limited datasets, benchmarks, and specialized training.
PMCID:12075794
PMID: 40360694
ISSN: 2398-6352
CID: 5844212