Augmenting machine learning algorithms to predict mortality using patient-reported outcomes in oncology.

Authors

Ravi Parikh

Ravi Bharat Parikh

University of Pennsylvania, Philadelphia, PA

Ravi Bharat Parikh , Jill Schnall , Manqing Liu , Peter Edward Gabriel , Corey Chivers , Michael Draugelis , Will Ferrell , Caryn Lerman , Justin E. Bekelman , Jinbo Chen

Organizations

University of Pennsylvania, Philadelphia, PA, USC Norris Comprehensive Cancer Center, Los Angeles, CA

Research Funding

U.S. National Institutes of Health
U.S. National Institutes of Health, Other Foundation

Background: Machine learning (ML) algorithms based on electronic health record (EHR) data have been shown to accurately predict mortality risk among patients with cancer, with areas under the curve (AUC) generally greater than 0.80. While patient-reported outcomes (PROs) may also predict mortality among patients with cancer, it is unclear whether routinely-collected PROs improve the predictive performance of EHR-based ML algorithms. Methods: This cohort study included 8600 patients with cancer who had an outpatient encounter at one of 18 medical oncology practices in a large academic health system between July 1st, 2019 and January 1st, 2020. 4692 (54.9%) patients completed assessments of symptoms, performance status, and quality of life from the PRO version of the Common Terminology Criteria for Adverse Events and the Patient-Reported Outcomes Measurement Information System Global v.1.2 scales. We hypothesized that ML models predicting 180-day all-cause mortality based on EHR + PRO data would improve AUC compared to ML models based on EHR data alone. We assessed univariate and adjusted associations between each PRO and 180-day mortality. To train the EHR-only model, we fit a Least Absolute Shrinkage and Selection Operator (LASSO) regression using 192 EHR demographic, comorbidity, and laboratory variables. To train the EHR + PRO model, we used a two-phase approach to fit a model using EHR data for all patients and PRO data for those who completed assessments. To test our hypothesis, we compared the bootstrapped AUC, area under the precision-recall curve (AUPRC), and sensitivity at a 20% risk threshold for both models. Results: 464 (5.4%) patients died within 180 days of the encounter. Decreased quality of life, functional status, and appetite were associated with greater 180-day mortality (Table). Compared to the EHR-only model, the EHR + PRO model significantly improved AUC (0.86 [95% CI 0.85-0.86] vs. 0.80 [95% CI 0.80-0.81]), AUPRC (0.40 [95% CI 0.37-0.42] vs. 0.30 [95% CI 0.28-0.32]), and sensitivity (0.45 [95% CI 0.42-0.48] vs. 0.33 [95% CI 0.30-0.35]). Conclusions: Routinely collected PROs augment EHR-based ML mortality risk algorithms. ML algorithms based on EHR and PRO data may facilitate earlier supportive care for patients with cancer. Association of PROs with 180-day mortality.

PRO
Univariable, Odds ratio [95% CI]
Adjusted for ML mortality risk, Odds ratio [95% CI]
Functional status
2.13 [1.90, 2.39]
1.52 [1.33, 1.73]
Anxiety
1.31 [1.15, 1.48]
1.21 [1.05, 1.39]
Constipation
1.43 [1.26, 1.63]
1.17 [1.01, 1.35]
Decreased appetite
1.89 [1.69. 2.12]
1.33 [1.17, 1.52]
Diarrhea
1.13 [1.00, 1.27]
0.97 [0.85, 1.11]
Fatigue
1.79 [1.61, 2.00]
1.38 [1.22, 1.56]
Nausea
1.56 [1.39, 1.75]
1.23 [1.07, 1.39]
Sadness
1.38 [1.21, 1.57]
1.23 [1.06, 1.42]
Dyspnea
1.59 [1.42, 1.77]
1.26 [1.11, 1.43]
Quality of life
1.97 [1.74, 2.24]
1.44 [1.25, 1.65]

Disclaimer

This material on this page is ©2024 American Society of Clinical Oncology, all rights reserved. Licensing available upon request. For more information, please contact licensing@asco.org

Abstract Details

Meeting

2021 ASCO Annual Meeting

Session Type

Poster Discussion Session

Session Title

Care Delivery and Regulatory Policy

Track

Care Delivery and Quality Care

Sub Track

Clinical Informatics/Advanced Algorithms/Machine Learning

Citation

J Clin Oncol 39, 2021 (suppl 15; abstr 1510)

DOI

10.1200/JCO.2021.39.15_suppl.1510

Abstract #

1510

Abstract Disclosures

Similar Abstracts