-Author name in bold denotes the presenting author
-Asterisk * with author name denotes a Non-ASH member
Clinically Relevant Abstract denotes an abstract that is clinically relevant.

PhD Trainee denotes that this is a recommended PHD Trainee Session.

Ticketed Session denotes that this is a ticketed session.

161 Optimized Multi-Class VTE-BERT Large Language Model for Prediction of Cancer Associated Thrombosis Phenotype

Program: Oral and Poster Abstracts
Type: Oral
Session: 901. Health Services and Quality Improvement: Non-Malignant Conditions Excluding Hemoglobinopathies: Optimizing Classical Hematology Care
Hematology Disease Topics & Pathways:
Research, Bleeding and Clotting, Artificial intelligence (AI), Adult, Translational Research, Epidemiology, Clinical Research, Thromboembolism, Diseases, Technology and Procedures, Study Population, Human, Natural language processing
Saturday, December 7, 2024: 1:00 PM

Ang Li, MD, MS1, Omid Jafari, PhD1*, Shengling Ma, MD, PhD2, Arash Maghsoudi3*, Barbara D Lam, MD4, Justine Ryu, MD5, Jun Yang Jiang, MD6*, Mrinal Ranjan, MPH1*, Emily Zhou7*, Raka Bandyo, MS8*, Bo Peng9*, Christopher I Amos1*, Nathanael R. Fillmore, PhD10,11*, Abiodun O. Oluyomi12* and Jennifer La11*

1Baylor College of Medicine, Houston, TX
2Section of Hematology-Oncology, Baylor College of Medicine, HOUSTON, TX
3Health Services Research Department of Medicine, Baylor College of Medicine, Houston, TX
4Division of Hematology, Department of Medicine, Beth Israel Deaconess Medical Center, Boston, MA
5Section of Hematology, Yale University School of Medicine, New Haven, CT
6Section of Hematology-Oncology, Department of Medicine, Baylor College of Medicine, Houston, TX
7McGovern Medical School, University of Texas Health Science Center, Houston, TX
8Harris Health System, Houston, TX
9Baylor College of Medicine, houston, TX
10Harvard Medical School, Boston, MA
11VA Boston Healthcare System, Boston, MA
12Section of Epidemiology and Population Sciences, Department of Medicine, Baylor College of Medicine, Houston, TX

Introduction: Advances in artificial intelligence and large language models (LLMs) have led to significant improvement in the natural language processing (NLP) algorithms. However, the best NLP models remain those finetuned to accomplish specific tasks. We previously showed an LLM-based NLP to identify overall venous thromboembolism (VTE) from unstructured clinical notes and radiology reports of cancer patients performed with 93% precision and 93% recall at a single site (Blood 2023;142(Supplement 1):1267). In the current work, we update the LLM to differentiate locations of the acute VTE (pulmonary embolism [PE], lower extremity deep vein thrombosis [LE-DVT], upper extremity DVT [UE-DVT]) and validate its performance in two independent and diverse patient populations.

Methods: In derivation/internal validation (Harris Health System [HHS]) and external validation (Veterans Affairs [VA]) cohorts, we identified clinical progress notes, discharge summaries, and radiology reports from patients with active cancer receiving systemic therapy. We then preprocessed the notes to identify sections with high clinical value (e.g. history of present illness, assessment, plan, impression, hospital course) and isolated sentences containing VTE keywords. VTE was defined as a newly diagnosed and acute PE, LE-DVT, or UE-DVT. Unusual thrombosis like splanchnic vein (mostly tumor thrombi) or ambiguous events (chronic VTE or clinically suspected without radiologic confirmation) were classified as negative. To determine the gold standard, multiple medically trained and blinded annotators reviewed patient notes in our NLP web interface.

The derivation cohort consisted of 700 patients from HHS with gold standard multi-class VTE labels (~1,000 positive and ~15,000 negative sentences). We finetuned the Bio_ClinicalBERT transformer model with a learning rate of 2e-5 and max 30 epochs (tuned to positive label training metrics) on AWS EC2 g4dn.12xlarge server. We then applied the finetuned model (now called VTE-BERT) to unstructured clinical notes from two previously labeled independent datasets of 458 cancer patients at HHS (internal validation; 97 confirmed VTE) and 764 cancer patients at VA (external validation; 489 confirmed VTE). No additional transferred or federated learning was performed. Accuracy, positive predictive value (PPV or precision) and sensitivity (recall) was estimated.

Results: In the HHS validation cohort, VTE-BERT achieved accuracy, precision, and recall of 96%, 90%, and 94%, respectively (97 true positives, 12 false positives, 6 false negatives in 458 patients). In the VA external validation cohort, the same model achieved accuracy, precision, and recall of 91%, 92%, and 94% (462 true positives, 42 false positives, 27 false negatives in 764 patients). The model successfully predicted most detractors like tumor thrombi and septic thrombi as negative. False positives were driven by a combination of historical VTE events mistakenly labeled as new and arterial thrombosis events. When excluding patients with recent VTE diagnosis (as predicted by positive VTE-BERT before index date), the precision improved further to 98% in HHS and 93% in VA. Finally, the updated model was able to accurately differentiate among the different types of VTE events.

Conclusion: The updated multi-class VTE-BERT LLM performed well in two different healthcare systems with vastly different clinical note types and formatting, demonstrating the generalizability of a well-trained LLM transformer. This represents one of the first efforts to apply a fine-tuned LLM NLP algorithm on raw clinical notes from an external independent dataset. While the model training took 1+ year effort in human annotation and model tuning, its application was straightforward. With the ease of access to unstructured clinical notes in most healthcare systems that utilize electronic health records, the updated VTE-BERT model, along with our annotation pipeline designed for physicians, can greatly reduce the cost and time associated with annotation and improve thrombosis research.

Disclosures: La: Merck: Research Funding.

*signifies non-member of ASH