Deep learning uncertainty quantification for clinical text classification

Alina Peluso, Ioana Danciu, Hong Jun Yoon, Jamaludin Mohd Yusof, Tanmoy Bhattacharya, Adam Spannaus, Noah Schaefferkoetter, Eric B. Durbin, Xiao Cheng Wu, Antoinette Stroup, Jennifer Doherty, Stephen Schwartz, Charles Wiggins, Linda Coyle, Lynne Penberthy, Georgia D. Tourassi, Shang Gao

Research output: Contribution to journalArticlepeer-review

2 Scopus citations

Abstract

Introduction: Machine learning algorithms are expected to work side-by-side with humans in decision-making pipelines. Thus, the ability of classifiers to make reliable decisions is of paramount importance. Deep neural networks (DNNs) represent the state-of-the-art models to address real-world classification. Although the strength of activation in DNNs is often correlated with the network's confidence, in-depth analyses are needed to establish whether they are well calibrated. Method: In this paper, we demonstrate the use of DNN-based classification tools to benefit cancer registries by automating information extraction of disease at diagnosis and at surgery from electronic text pathology reports from the US National Cancer Institute (NCI) Surveillance, Epidemiology, and End Results (SEER) population-based cancer registries. In particular, we introduce multiple methods for selective classification to achieve a target level of accuracy on multiple classification tasks while minimizing the rejection amount—that is, the number of electronic pathology reports for which the model's predictions are unreliable. We evaluate the proposed methods by comparing our approach with the current in-house deep learning-based abstaining classifier. Results: Overall, all the proposed selective classification methods effectively allow for achieving the targeted level of accuracy or higher in a trade-off analysis aimed to minimize the rejection rate. On in-distribution validation and holdout test data, with all the proposed methods, we achieve on all tasks the required target level of accuracy with a lower rejection rate than the deep abstaining classifier (DAC). Interpreting the results for the out-of-distribution test data is more complex; nevertheless, in this case as well, the rejection rate from the best among the proposed methods achieving 97% accuracy or higher is lower than the rejection rate based on the DAC. Conclusions: We show that although both approaches can flag those samples that should be manually reviewed and labeled by human annotators, the newly proposed methods retain a larger fraction and do so without retraining—thus offering a reduced computational cost compared with the in-house deep learning-based abstaining classifier.

Original languageEnglish
Article number104576
JournalJournal of Biomedical Informatics
Volume149
DOIs
StatePublished - Jan 2024

Funding

This work was supported in part by the Joint Design of Advanced Computing Solutions for Cancer program established by Department of Energy and the National Cancer Institute of the National Institutes of Health. This work was performed under the auspices of DOE by Argonne National Laboratory under Contract DE-AC02-06-CH11357, Lawrence Livermore National Laboratory under Contract DE-AC52-07NA27344, Los Alamos National Laboratory under Contract DE-AC5206NA25396, and Oak Ridge National Laboratory under Contract DE-AC05-00OR22725. KCR data were collected with funding from NCI’s SEER Program ( HHSN261201800013I ), the CDC’s National Program of Cancer Registries (NPCR) ( U58DP 00003907 ), and the Commonwealth of Kentucky . LTR data were collected using funding from NCI’s SEER program ( HHSN261201800007I ), the CDC’s NPCR ( NU58DP 006332-02-00 ), and the State of Louisiana . NJSCR data were collected using funding from NCI’s SEER program ( HHSN261201300021I ), the CDC’s NPCR ( NU58DP 006279-02-00 ), as well as the State of New Jersey and the Rutgers Cancer Institute of New Jersey . NMTR’s participation in this project was supported by Contract HHSN261201800014I , Task Order HHSN26100001 from the NCI’s SEER program . The Cancer Surveillance System is supported by the NCI’s SEER program ( HHSN261291800004I ) with additional funds provided by the Fred Hutchinson Cancer Center and the State of Washington . UCR is funded by the NCI’s SEER program ( HHSN261201800016I ) and the CDC’s NPCR ( NU58DP0063200 ) with additional support from the University of Utah and Huntsman Cancer Foundation . Notice: This manuscript has been authored in part by UT-Battelle LLC under contract DE-AC05-00OR22725 with the US Department of Energy (DOE). The publisher acknowledges the US government license to provide public access under the DOE Public Access Plan ( http://energy.gov/downloads/doe-public-access-plan ). This work was supported in part by the Joint Design of Advanced Computing Solutions for Cancer program established by Department of Energy and the National Cancer Institute of the National Institutes of Health. This work was performed under the auspices of DOE by Argonne National Laboratory under Contract DE-AC02-06-CH11357, Lawrence Livermore National Laboratory under Contract DE-AC52-07NA27344, Los Alamos National Laboratory under Contract DE-AC5206NA25396, and Oak Ridge National Laboratory under Contract DE-AC05-00OR22725. KCR data were collected with funding from NCI's SEER Program (HHSN261201800013I), the CDC's National Program of Cancer Registries (NPCR) (U58DP 00003907), and the Commonwealth of Kentucky. LTR data were collected using funding from NCI's SEER program (HHSN261201800007I), the CDC's NPCR (NU58DP 006332-02-00), and the State of Louisiana. NJSCR data were collected using funding from NCI's SEER program (HHSN261201300021I), the CDC's NPCR (NU58DP 006279-02-00), as well as the State of New Jersey and the Rutgers Cancer Institute of New Jersey . NMTR's participation in this project was supported by Contract HHSN261201800014I, Task Order HHSN26100001 from the NCI's SEER program. The Cancer Surveillance System is supported by the NCI's SEER program (HHSN261291800004I) with additional funds provided by the Fred Hutchinson Cancer Center and the State of Washington. UCR is funded by the NCI's SEER program (HHSN261201800016I) and the CDC's NPCR (NU58DP0063200) with additional support from the University of Utah and Huntsman Cancer Foundation.

FundersFunder number
CDC's NPCR
CDC’s NPCRNU58DP 006332-02-00
University of Utah and Huntsman Cancer Foundation
National Institutes of Health
U.S. Department of Energy
National Cancer InstituteU58DP 00003907, HHSN261201800007I
Argonne National LaboratoryDE-AC02-06-CH11357
Lawrence Livermore National LaboratoryDE-AC52-07NA27344
Oak Ridge National LaboratoryDE-AC05-00OR22725
Rutgers Cancer Institute of New JerseyHHSN26100001, HHSN261201800014I, HHSN261291800004I
Los Alamos National LaboratoryDE-AC5206NA25396
UT-Battelle
State of New Jersey
State of LouisianaHHSN261201300021I, NU58DP 006279-02-00
State of WashingtonHHSN261201800016I, NU58DP0063200

    Keywords

    • Abstaining classifier
    • Accuracy
    • CNN
    • DNN
    • Deep learning
    • HiSAN
    • NCI SEER
    • Pathology reports
    • Selective classification
    • Text classification
    • Uncertainty quantification

    Fingerprint

    Dive into the research topics of 'Deep learning uncertainty quantification for clinical text classification'. Together they form a unique fingerprint.

    Cite this