Your browser doesn't support javascript.
loading
Explainable machine learning can outperform Cox regression predictions and provide insights in breast cancer survival.
Moncada-Torres, Arturo; van Maaren, Marissa C; Hendriks, Mathijs P; Siesling, Sabine; Geleijnse, Gijs.
Afiliação
  • Moncada-Torres A; Department of Research and Development, Netherlands Comprehensive Cancer Organization (IKNL), Zernikestraat 29, 5612 HZ, Eindhoven, The Netherlands. a.moncadatorres@iknl.nl.
  • van Maaren MC; Department of Research and Development, Netherlands Comprehensive Cancer Organization (IKNL), Zernikestraat 29, 5612 HZ, Eindhoven, The Netherlands.
  • Hendriks MP; Department of Health Technology and Services Research, University of Twente, Enschede, The Netherlands.
  • Siesling S; Department of Research and Development, Netherlands Comprehensive Cancer Organization (IKNL), Zernikestraat 29, 5612 HZ, Eindhoven, The Netherlands.
  • Geleijnse G; Department of Medical Oncology, Northwest Clinics, Alkmaar, The Netherlands.
Sci Rep ; 11(1): 6968, 2021 03 26.
Article em En | MEDLINE | ID: mdl-33772109
Cox Proportional Hazards (CPH) analysis is the standard for survival analysis in oncology. Recently, several machine learning (ML) techniques have been adapted for this task. Although they have shown to yield results at least as good as classical methods, they are often disregarded because of their lack of transparency and little to no explainability, which are key for their adoption in clinical settings. In this paper, we used data from the Netherlands Cancer Registry of 36,658 non-metastatic breast cancer patients to compare the performance of CPH with ML techniques (Random Survival Forests, Survival Support Vector Machines, and Extreme Gradient Boosting [XGB]) in predicting survival using the [Formula: see text]-index. We demonstrated that in our dataset, ML-based models can perform at least as good as the classical CPH regression ([Formula: see text]-index [Formula: see text]), and in the case of XGB even better ([Formula: see text]-index [Formula: see text]). Furthermore, we used Shapley Additive Explanation (SHAP) values to explain the models' predictions. We concluded that the difference in performance can be attributed to XGB's ability to model nonlinearities and complex interactions. We also investigated the impact of specific features on the models' predictions as well as their corresponding insights. Lastly, we showed that explainable ML can generate explicit knowledge of how models make their predictions, which is crucial in increasing the trust and adoption of innovative ML techniques in oncology and healthcare overall.
Assuntos

Texto completo: 1 Base de dados: MEDLINE Assunto principal: Neoplasias da Mama / Sistema de Registros / Medição de Risco / Máquina de Vetores de Suporte / Aprendizado de Máquina Idioma: En Ano de publicação: 2021 Tipo de documento: Article

Texto completo: 1 Base de dados: MEDLINE Assunto principal: Neoplasias da Mama / Sistema de Registros / Medição de Risco / Máquina de Vetores de Suporte / Aprendizado de Máquina Idioma: En Ano de publicação: 2021 Tipo de documento: Article