Classification of clinically significant prostate cancer on multi-parametric MRI

A validation study comparing deep learning and radiomics

Journal Article (2022)
Author(s)

J.M. Castillo (Erasmus MC)

Muhammad Arif (Erasmus MC)

M.P.A. Starmans (Erasmus MC)

W.J. Niessen (TU Delft - ImPhys/Computational Imaging, TU Delft - ImPhys/Medical Imaging, Erasmus MC)

C.H. Bangma (Erasmus MC)

Ivo G. Schoots (Erasmus MC)

J.F. Veenland (Erasmus MC)

Department
Biomechanical Engineering
Copyright
© 2022 J.M. Castillo, M. Arif, M.P.A. Starmans, W.J. Niessen, C.H. Bangma, Ivo G. Schoots, J.F. Veenland
DOI related publication
https://doi.org/10.3390/cancers14010012
More Info
expand_more
Publication Year
2022
Language
English
Copyright
© 2022 J.M. Castillo, M. Arif, M.P.A. Starmans, W.J. Niessen, C.H. Bangma, Ivo G. Schoots, J.F. Veenland
Department
Biomechanical Engineering
Issue number
1
Volume number
14
Reuse Rights

Other than for strictly personal use, it is not permitted to download, forward or distribute the text or part of it, without the consent of the author(s) and/or copyright holder(s), unless the work is under an open content license such as Creative Commons.

Abstract

The computer-aided analysis of prostate multiparametric MRI (mpMRI) could improve significant-prostate-cancer (PCa) detection. Various deep-learning-and radiomics-based methods for significant-PCa segmentation or classification have been reported in the literature. To be able to assess the generalizability of the performance of these methods, using various external data sets is crucial. While both deep-learning and radiomics approaches have been compared based on the same data set of one center, the comparison of the performances of both approaches on various data sets from different centers and different scanners is lacking. The goal of this study was to compare the performance of a deep-learning model with the performance of a radiomics model for the significant-PCa diagnosis of the cohorts of various patients. We included the data from two consecutive patient cohorts from our own center (n = 371 patients), and two external sets of which one was a publicly available patient cohort (n = 195 patients) and the other contained data from patients from two hospitals (n = 79 patients). Using multiparametric MRI (mpMRI), the radiologist tumor delineations and pathology reports were collected for all patients. During training, one of our patient cohorts (n = 271 patients) was used for both the deep-learning-and radiomics-model development, and the three remaining cohorts (n = 374 patients) were kept as unseen test sets. The performances of the models were assessed in terms of their area under the receiver-operating-characteristic curve (AUC). Whereas the internal cross-validation showed a higher AUC for the deep-learning approach, the radiomics model obtained AUCs of 0.88, 0.91 and 0.65 on the independent test sets compared to AUCs of 0.70, 0.73 and 0.44 for the deep-learning model. Our radiomics model that was based on delineated regions resulted in a more accurate tool for significant-PCa classification in the three unseen test sets when compared to a fully automated deep-learning model.