 Research
 Open access
 Published:
Statistical models versus machine learning for competing risks: development and validation of prognostic models
BMC Medical Research Methodology volume 23, Article number: 51 (2023)
Abstract
Background
In health research, several chronic diseases are susceptible to competing risks (CRs). Initially, statistical models (SM) were developed to estimate the cumulative incidence of an event in the presence of CRs. As recently there is a growing interest in applying machine learning (ML) for clinical prediction, these techniques have also been extended to model CRs but literature is limited. Here, our aim is to investigate the potential role of ML versus SM for CRs within noncomplex data (small/medium sample size, low dimensional setting).
Methods
A dataset with 3826 retrospectively collected patients with extremity softtissue sarcoma (eSTS) and nine predictors is used to evaluate modelpredictive performance in terms of discrimination and calibration. Two SM (causespecific Cox, FineGray) and three ML techniques are compared for CRs in a simple clinical setting. ML models include an original partial logistic artificial neural network for CRs (PLANNCR original), a PLANNCR with novel specifications in terms of architecture (PLANNCR extended), and a random survival forest for CRs (RSFCR). The clinical endpoint is the time in years between surgery and disease progression (event of interest) or death (competing event). Time points of interest are 2, 5, and 10 years.
Results
Based on the original eSTS data, 100 bootstrapped training datasets are drawn. Performance of the final models is assessed on validation data (left out samples) by employing as measures the Brier score and the Area Under the Curve (AUC) with CRs. Miscalibration (absolute accuracy error) is also estimated. Results show that the ML models are able to reach a comparable performance versus the SM at 2, 5, and 10 years regarding both Brier score and AUC (95% confidence intervals overlapped). However, the SM are frequently better calibrated.
Conclusions
Overall, ML techniques are less practical as they require substantial implementation time (data preprocessing, hyperparameter tuning, computational intensity), whereas regression methods can perform well without the additional workload of model training. As such, for noncomplex real life survival data, these techniques should only be applied complementary to SM as exploratory tools of model’s performance. More attention to model calibration is urgently needed.
Background
Survival analysis (also referred as timetoevent analysis) is used to estimate the lifespan of a particular population under study. Frequently, survival data are right censored; time to event is not observed for all patients due to followup interruption before experiencing the event of interest or time limitations (study termination). Competing risks (CRs) occur frequently in clinical applications of survival data [1,2,3,4]. In this type of data an individual may fail from one of several causes. A CR is an event whose occurrence precludes the occurrence of an event of interest (for instance death may preclude the occurrence of disease relapse) [5, 6]. In health research, CRs are unlikely to be independent as the biology suggests at least some dependence between events. In several chronic diseases attributable to aging and frailty such as cancer, chronic heart failure, or dementia, study populations are susceptible to CRs [7].
The most popular nonparametric approach to estimate survival in the presence of right censored timetoevent data is the KaplanMeier’s methodology (KM) [8]. However, in the presence of CRs, this methodology overestimates the probability of failure which might lead to overtreatment of patients [1, 5, 9]. Different statistical models (SM) have been developed to estimate the cumulative incidence (absolute risk) of an event in the presence of CRs such as the causespecific Cox model [10], and the FineGray subdistribution hazards regression model [11]. The former is a natural extension of the standard proportional hazards Cox model for the CRs setting where a Cox model is applied for each causespecific hazard. The latter models the effect of covariates directly on the cumulative incidence function (CIF) over time reporting on the subdistribution hazard ratio [9].
Nowadays, there is a growing interest in applying machine learning (ML) for prediction (diagnosis or prognosis) of clinical outcomes [12, 13] which has sparked a debate regarding the added value of ML techniques versus SM in the medical field. Criticism is attributed to ML prediction models. Despite no assumptions about the data structure are made, and being able to naturally incorporate interactions between predictive features, they are prone to overfitting of the training data and they lack extensive assessment of predictive accuracy (i.e., absence of calibration curves) [14, 15]. On the other hand, traditional regression methods are considered straightforward to use and harder to overfit. That being said, they do make certain (usually strong) assumptions such as the proportional hazards over time for the Cox model, and require manual prespecification of interaction terms.
Amongst ML techniques, artificial neural networks have been a common choice in healthcare. This trend is pertinent with the collection of large and complex patient information in electronic health records, and the rise of computational power [16]. Over the years, neural networks and other ML techniques have been developed for survival data. Wang et al. in 2019 provide a comprehensive survey of conventional and modern approaches for rightcensored timetoevent data [17]. The authors describe several ML techniques and suggest that neural networks are wellsuited to predict survival and estimate disease risk.
A common approach in the literature is the partial logistic artificial neural network (PLANN) of Biganzoli et al. (1998) [18]. For the purpose of implementation, time is specified in discrete nonoverlapping time intervals which are added as an input feature in a longitudinally transformed feedforward network with logistic activation, and entropy error function. The output layer estimates smoothed discrete hazards for each time interval. PLANN was extended by Lisboa et al. (2003) under a Bayesian regularisation framework which performs automatic relevance determination (PLANNARD) [19]. Recently, Kantidakis et al. in 2020 proposed extensions of PLANN in terms of architecture i.e., new hyperparameters, new activation functions, and time interval specification as multiple input features [20]. Next to survival neural networks (SNNs), another wellknown ML technique for clinical prediction of survival data is random survival forests (RSF, Ishwaran et al. 2008) [21]. RSF adapt Breiman’s random forest method by using a collection of survival trees [22].
ML approaches have also been employed for CRs, but the literature is limited. The PLANNCR approach was developed by Biganzoli et al. in 2006 for the joint modelling of discrete causespecific hazards [23]. This extends PLANN by using the time (in discrete time intervals) as an input feature in a longitudinally transformed network with multinomial error function and logistic  softmax activation functions for the hidden and the output layer (multiple output nodes), respectively. Later, Lisboa et al. (2009) implemented PLANNCR under a Bayesian regularisation framework (PLANNCRARD) [24]. Ishwaran et al. extended RSF for CRs (RSFCR) in 2014 to estimate the CIF of competing events [25].
For this work, a dataset with small/medium sample size and limited number of predictive features (lowdimensional setting) is analysed. This concerns a retrospectively collected cohort of 3826 patients with highgrade extremity softtissue sarcomas (eSTS) treated surgically with curative intent. Nine prognostic factors are used to develop and validate several clinical prediction models with CRs for ML techniques and SM. The clinical endpoint of the study is defined as the time in years between surgery and disease progression (as local recurrence or distant metastasis; event of interest) of eSTS, where death is a competing event. Time points of interest are 2, 5, and 10 years (5year horizon is of major clinical interest). Analyses were performed in R programming language version 4.1.2 [26].
The aims of this manuscript can be summarised as: (i) examination of extensions of PLANNCR method (PLANNCR extended) for the development and validation of prognostic clinical prediction models with competing events, (ii) systematic evaluation of modelpredictive performance for ML techniques (PLANNCR original, PLANNCR extended, RSFCR) and SM (causespecific Cox, FineGray) regarding discrimination and calibration, (iii) investigation of the potential role of ML in contrast to conventional regression methods for CRs in noncomplex eSTS data (small/medium sample size, low dimensional setting), (iv) practical utility of the methods for prediction.
The paper is organized as follows. In Section “Methods”, the eSTS data is presented. Further sections discuss basic concepts for CRs, the SM and the ML techniques, model training, and how the predictive performance was assessed. Section “Results” describes PLANNCR extended tuned with two measures, and compares the predictive performance of all methods in terms of discrimination and calibration. The manuscript ends with a “Discussion” about findings, limitations, and future perspectives of this work.
Methods
This section is divided into several subsections where the methodology used for this work is presented to the reader. To begin with, the clinical data is described. Next, the SM and the ML techniques are discussed. Two wellknown statistical models for CRs are employed: the causespecific Cox model [10], and the FineGray subdistribution hazards regression model [11], as well as two extensions of popular ML techniques for CRs: the RSFCR [25], and the PLANNCR [23] as originally developed or with some modifications. Afterwards, it is presented how the models were trained, and which performance measures were used to evaluate their predictive ability. More technical details are provided in the Supplementary material.
Dataset
Extremity softtissue sarcomas (eSTS) constitute a wide variety of histological subtypes with different sizes and grades that affect patients of any age group. Treatment protocols may differ between institutes and countries. Hence, important differences can be observed in the clinical course and prognosis of patients [27]. Over the years, several prognostic prediction models have been developed for overall survival and local recurrence [28,29,30].
For this project, a retrospectively collected cohort of 3826 patients with eSTS was used [29]. The dataset contained pseudoanonymised patients from Leiden University Medical Center (Leiden, the Netherlands), Royal Orthopaedic Hospital (Birmingham and Stanmore, UK), Netherlands Cancer Institute (Amsterdam, the Netherlands), Mount Sinai Hospital (Toronto, Canada), the Norwegian Radium Hospital (Oslo, Norway), Aarhus University Hospital (Aarhus, Denmark), Skåne University Hospital (Lund, Sweden), Medical University Graz (Graz, Austria), Royal Marsden Hospital (London, UK), Daniel den Hoed (Rotterdam, the Netherlands), Radboud University Medical Center (Nijmegen, the Netherlands), University Medical Center Groningen (Groningen, the Netherlands), Haukeland University Hospital (Bergen, Norway), Helios Klinikum BerlinBuch (Berlin, Germany), MedUni Vienna (Vienna, Austria), Vienna General Hospital (Vienna, Austria). In addition, eSTS patients from EORTC 62931 randomised controlled trial were included [31]. Data from the centers was collected between January 2000 and December 2014. Patients from the EORTC trial were recruited between February 1995 and December 2003.
Patients were selected from the sarcoma registry of each hospital based on histological diagnosis. Those initially treated without curative intent, showed local recurrence or distant metastasis at baseline, had Kaposi’s sarcoma or rhabdomyosarcoma (pediatric form), tumor was present in their abdomen, thorax, head or neck, or were treated with isolated limp perfusion as neoadjuvant treatment were excluded from the collection.
The dataset contained nine prognostic factors. Seven were categorical; gender (female or male), surgical margin (\(R_{0}\) for negative or \(R_{12}\) for positive with tumor cells in the inked surface of the resection margin), adjuvant chemotherapy (no or yes), tumor grade (II or III), tumor depth in relation to investing fascia (superficial or deep), radiotherapy (no, neoadjuvant or adjuvant), histological subtype (myxofibrosarcoma, synovial sarcoma, malignant fibrous histiocytoma / undifferentiated pleomorphic sarcoma / (pleomorphic) soft tissue sarcomas nototherwisespecified, leiomyosarcoma, liposarcoma or other). Two were continuous; age at baseline (in years) and tumor size by the largest diameter measured at pathological examination (in centimetres).
Median followup survival time is 5.98 years estimated by reverse KaplanMeier (25% quartile: 3.94 years, 75% quartile: 8.80 years, range: 0.01 to 16.85 years) [8]. The endpoint of interest is defined as the time in years between surgery and disease progression (local recurrence or distant metastasis) of eSTS, with death as competing event; 1773 patients were alive/censored at the end of followup (46.34%), 1554 had disease progression (40.62%), and 499 died without local recurrence/distant metastasis (13.04%).
The dataset contained 3.70% missing data overall for the nine variables, with 2514 complete cases (65.71%). More specifically, there were missing values (0.9711%) for all variables; 11.00% for tumor depth (421/3826), 8.21% for histological subtype (314/3826), 7.40% for surgical margin (283/3826), 4.36% for adjuvant chemotherapy (167/3826), 4.05% for tumor size (155/3826), 3.53% for gender (135/3826), 2.61% for radiotherapy (100/3826), 1.99% for tumor grade (76/3826), and 0.97% for age (37/3826), in decreasing order, respectively.
A simple imputation was used to avoid discarding observations from nearly complete records. The missForest algorithm was applied to reconstruct any missing values, which is the most exhaustive/accurate random forest algorithm for missing data [32]. This is a nonparametric imputation method that does not make any a priory assumptions regarding the data structure. A random forest with 1000 trees (for model stability) was built for each variable with missing information, testing all possible variable combinations as responses. Table 1 provides patient demographics of the final dataset (demographics of the original dataset are provided in Table S1 of Additional file 1).
Basic concepts for competing risks
Typically for survival data, if several types of events occur, a model describing progression for each of the CRs is needed. The observable data is represented by the time of failure T, the cause of failure D (\(D \in 1, \cdots , k\), \(k \ge 1\); here k = 2), and a covariate vector \(\mathbf {Z}\). Usually there is one type of event that is of interest (i.e., disease progression as local recurrence or distant metastasis) whereas the other events could prevent it from occurring (here competing event is death).
Following Putter et al. (2007) [1], a fundamental concept in modelling CRs is the causespecific hazard function which denotes the hazard of failing from a given cause in the presence of CRs:
Then, the cumulative causespecific hazard can be specified as
and the survival function (probability of not having failed from any cause at time t) can be written as
The cumulative incidence function (CIF) of cause k is defined as \(I_{k}(t) = Prob(T \le t, D = k)\), the probability of failing from cause k before time t. This can be linked to the causespecific hazards through the expression:
This is also called the subdistribution function based on the fact that the cumulative probability to fail from cause k cannot reach one, and therefore, it is not a proper probability distribution.
Regression models for competing risks
Causespecific Cox model
Regression on causespecific hazards is an extension of the popular Cox proportional hazards model for CRs [10, 33]. The causespecific hazard of cause k of a subject with covariate vector \(\mathbf {Z}\) is modelled as
where \(\lambda _{k, 0}(t)\) is the causespecific hazard, and the vector \(\varvec{\beta }_{k}\) represents the effects of covariates on cause k. Patients who move to another state other than k are censored at their transition time.
Fine and Gray model
In 1999, Fine and Gray introduced a subdistribution hazards model, which can directly regress on CIF [11]:
For the causespecific Cox model, the risk set (number of patients at risk) decreases at each time point where there is a failure of another cause. On the other hand, for Fine and Gray’s model, individuals who fail from another cause remain in the risk set. The subdistribution hazards are then modelled assuming proportional hazards:
Similar to the standard Cox model, the partial likelihood approach is used to estimate the parameters.
Machine learning techniques for competing risks
Random survival forests
Random survival forests for competing risks (RSFCR) [25] are an extension of the RSF framework [21, 22] for CRs with right censored data proposed by Ishwaran et al. in 2014. It is a fully nonparametric ensemble tree approach for the estimation of the CIF for competing events (CIF and causespecific hazard function are related as shown in equation (4)). RSFCR can directly model nonlinear effects and interactions to perform accurate prediction without making any prior assumptions about the underlying data.
The algorithm of RSFCR is based on recursive binary partitioning while injecting randomness in two ways: (a) drawing B bootstrap samples from the learning data, and (b) growing a single CRs tree for each bootstrap sample by randomly selecting a subset of candidate variables at each node (region of the tree). A CR splitting rule is maximised to split each parent node into daughter nodes using the selected variables. The authors propose two splitting rules: either an eventspecific or a combination of eventspecific splitting rules across the k events. Here, the eventspecific splitting rule was applied because disease progression was of major interest (weighted logrank splitting, technical details in [25]). Then each tree is grown to full size under the constraint that terminal nodes (the ends of each tree) should have at least one unique case. In the terminal nodes, the KaplanMeier [8] and the AalenJohansen [34] methodologies are used to estimate the eventfree survival function and the causespecific CIF, respectively. Finally, the ensemble estimates are calculated averaging each estimator over the B grown trees. More technical details are provided in Additional file 2.
Partial logistic artificial neural networks
In 2006, Biganzoli et al. extended the partial logistic artificial neural network to competing risks (PLANNCR) for the joint modelling of discrete causespecific hazards [18, 23]. PLANNCR is a feedforward network comprised of a group of units called nodes (or neurons) in each layer. It has an input layer that picks up the signals and passes them to a single hidden layer after the application of an activation (also called transformation) function. An activation function modulates the degree of nonlinearity transferred from the input features to the hidden layer. Connections between the artificial neurons of different layers are called edges  each having a weight. Weights are adjusted through training increasing or decreasing the strength of each connection [35]. Signals are transmitted towards the output layer, which provides a smoothed estimation of discrete conditional event probabilities (in multiple output nodes; each for an event), with another activation function.
For the purpose of implementation, survival times are discretized into a set of \(l = 1, \cdots , L\) disjoint intervals \(A_{l} = (\tau _{l1}, \tau _{l}]\), where \(0=\tau _{0}<\tau _{1}<\cdots <\tau _{L}\) is a set of predefined time points (usually years). For the \(l^{th}\) interval, observed times are grouped on a single point \(\tau _{l}\). Data has to be transformed into a longitudinal format where the time variable (in intervals) is added as part of the input features next to the prognostic features. Subjects are repeated for the number of intervals observed on the training data, and for all time intervals on the test data. PLANNCR can model nonlinear, nonproportional, and nonadditive effects between the prognostic factors on the causespecific hazards. Here, without loss of generality, each subject was repeated for 1 up to 11 time intervals denoting years since surgery. The last interval included survival times longer than 10 years (subsequent intervals were not of interest).
In the CRs model, the response vector has \(R + 1\) variables, with \(r = 1, \cdots ,R\) the possible causes of interest (here \(R = 2\)). Let \(\mathbf {z}_{k} = (\tau _{l}, \mathbf {x}_{k}\)) be defined by two components: the covariate vector \(\mathbf {x}_{k}\) (\(k = 1, 2, \cdots , p\)) and the time interval \(\tau _{l}\). The joint dependence of the discrete causespecific hazards is modelled as:
where \(h = 1, \cdots , H\) nodes in the hidden layer, \(\varvec{\beta }\) the vector of estimated weights for the inputhidden (\(\beta _{01}, \cdots , \beta _{0H}, \beta _{1}, \cdots , \beta _{H}\)), hiddenoutput layers (\(\beta _{0}, \beta _{1}^{a}, \cdots , \beta _{R}^{a}\)), and \(\alpha _{h}\) the sigmoid (logistic) activation function for the hidden layer \(\alpha _{h}(\mathbf {z}_{k}, \varvec{\beta }_{h}) = \frac{\exp (\beta _{0h} + \beta _{h}^T \mathbf {z}_{k})}{1 + \exp (\beta _{0h} + \beta _{h}^T\mathbf {z}_{k})}\).
Activation function for the output layer is the softmax providing the discrete causespecific hazards:
for \(l = 1, \cdots , L\) intervals, and \(r = 1, \cdots ,R\) causes of interest. Since PLANNCR has a different output node for each CR (1 + R output nodes in total), it is an extension of standard neural networks for multiple classification resorting to the multinomial likelihood. For the rest of this paper, this will be called PLANNCR original [23].
Similar extensions to the specification of the PLANNCR are provided as in Kantidakis et al. (PLANN extended, 2020) [20]. More specifically, PLANNCR extended is tuned investigating two new activation functions for the hidden layer: (1) the rectified linear unit (ReLU) a common activation function, \(\alpha _{h}(\mathbf {z}_{k}, \varvec{\beta }_{h}) = \max (0, \beta _{0h} + \beta _{h}^T \mathbf {z}_{k})\), or (2) the hyperbolic tangent (tanh), \(\alpha _{h}(\mathbf {z}_{k}, \varvec{\beta }_{h}) = \frac{1  \exp (2(\beta _{0h} + \beta _{h}^T \mathbf {z}_{k}))}{1 + \exp (2(\beta _{0h} + \beta _{h}^T \mathbf {z}_{k}))}\). Each time a neural network is fitted with one of these activation functions for the hidden layer or with the sigmoid (logistic) activation function (as in PLANNCR original). Note that the activation function for the output layer is necessarily the softmax to provide smoothed discrete hazard estimation. New hyperparameters are specified in a stateoftheart R library [36]. In contrast with Kantidakis et al. (2020), the L nonoverlapping intervals are specified in one time variable (instead of L separate variables) to not inflate the number of input features. Moreover, networks with two hidden layers are not tested here due to the danger for overfitting (smallmedium sample size, small number of predictors). More technical details for PLANNCR original and PLANNCR extended are provided in Additional file 2.
Model training
Figure 1 shows how model training was performed. Based on the original eSTS data, 100 bootstrapped training datasets were drawn with 3826 patients each (sampling with replacement, \(\approx\) 63.2% of the original data). These datasets were randomly split into two complementary parts to tune the hyperparameters of the ML models using grid search (\(\frac{3}{4}\) to train the models and \(\frac{1}{4}\) to test their performance, same parts for all methods). Performance of the final models was assessed on the validation data, which were the left out samples (outofbag, \(\approx\) 36.8% of the data). Outofbag error estimates are almost identical to Nfold crossvalidation [37]. For the standard regression approaches, models were built on each complete training dataset (consisted of 3826 patients) using the nine covariates. Their predictive performance was evaluated on the respective validation dataset. Complex functional form dependencies (nonlinear, nonadditive, timedependent effects) were not investigated. All analyses were performed in R programming language version 4.1.2 [26]. Packages used in the implementation and tuning parameters for the ML techniques are provided in Additional file 2.
Predictive performance assessment
Predictive performance of the methods was assessed in terms of discrimination and calibration on each validation dataset. The Area Under the Curve (AUC) and Brier score with CRs were used. Miscalibration (absolute accuracy error) was also estimated. These evaluation measures were employed since they are model agnostic  they can be applied to any model to evaluate its predictive performance. Other measures such as the Akaike Information Criterion (AIC) or the Bayesian Information Criterion (BIC) were not selected, since they cannot be (easily) calculated for comparison of the different SM and ML techniques applied here.
Following Blanche et al. [38], we present the dynamic version of the measures with CRs (see also [39]). Let \(\pi _{i}(\cdot ,\cdot )\) be a subjecti specific prediction process (\(i = 1, 2, \cdots , n\) independent and identically distributed subjects) for all landmark times s (times at which predictions are made) and prediction horizon t. Without loss of generality, we set \(\pi _{i} (s, t) = 0\) for all subjects i who are no longer at risk at time s, and focus on prediction of event \(D = 1\) (main event investigated). A dynamic AUC at landmark time s for a prediction horizon t can be defined as
where \(\Delta _{i} (s, t)\) = \(\mathbbm {1}_{s<T_{i}\le s+t, D_{i} = 1}\), \(\Delta _{i} (s, t) = 1\) when subject i experiences the main event of interest within time interval \((s, s + t]\) (case), and \(\Delta _{i} (s, t) = 0\) when subject i experiences a competing event within the time interval or is eventfree at \(s+t\) (control) [40].
Dynamic AUC with CRs is a measure of discrimination. It typically ranges from 0.5 to 1 (the higher the better). A good predictive accuracy is provided by a model that usually gives higher predicted risks of event for subjects who experience the event of interest compared to subjects who did not experience the event of interest.
A more complete predictive accuracy measure with CRs is the Brier score. The dynamic expected Brier score can be written as
This expression can be expanded based on Graaf et al. 1999 [41] taking the following form
where \(H(s) = \{ \mathbf {X}, Y(s), T>s \}\) the information at time s used to compute the prediction of \(\pi (s, t)\). The first term in (12) measures calibration  how close the predictions are to \(\mathbb{E} [\Delta (s, t)  H(s)]\), the “true” underlying risk of event in \((s, s+t]\) given H(s). In addition, the second term depends on the discrimination ability of H(s). Thus, Brier score is a measure of both calibration and discrimination. Typically, it ranges from 0 to 0.25 (lower values mean smaller prediction error).
When censored data are present, the indicator \(\Delta _{i}(s, t)\) is unknown (cannot be computed) for all subjects i censored within interval \((s, s+t]\). Therefore, the Inverse Probability of Censoring Weighting (IPCW) technique has to be applied for the estimation of both dynamic AUC and Brier score for CRs. For details see [38]. Here, the landmark time was set to \(s = 0\) (baseline) for all analyses as all prognostic factors were time fixed.
Last, the predictive ability of the methods was evaluated based on their miscalibration on each validation dataset (see Fig. 1). Model calibration refers to the agreement between observed and predicted outcomes, in this case agreement between observed and predicted cumulative incidence event probabilities for a cause \(D = k\) at time \(t = t_{0}\) [42, 43]. For each SM and ML model, the predicted cumulative incidence event probabilities are estimated on a validation dataset, and the data is split into \(m = 4\) equally sized groups based on the quantiles of the predicted event probabilities. Quantiles were selected instead of (for instance) deciles to avoid any computational issues. Then, the observed cumulative incidence probabilities are calculated for each group. Miscalibration is defined as the mean squared error (MSE) of the difference between the observed and the predicted cumulative probabilities of failure from a specific cause \(D = k\) at time horizon \(t = t_{0}\)
with \(I_{k}^{(m)}(t_{0})\) and \(\hat{I_{k}}^{(m)}(t_{0})\) the observed and predicted cumulative event probability for group m, respectively.
Results
In this section, results for the eSTS data are presented. The following models are compared in terms of predictive performance: (1) Causespecific Cox, (2) FineGray, (3) PLANNCR original, (4) PLANNCR extended, (5) RSFCR. Each model is assessed on 100 validation datasets (see Fig. 1). More results about the comparison between the methods are provided in Additional file 3.
PLANNCR tuned with Brier score or AUC at 5 years
The hyperparameters selected for PLANNCR original and PLANNCR extended are provided in section 1 of Additional file 3. The most effective combinations are reported separately based on the Brier score / AUC at 5 years (5year horizon was of major clinical interest).
For PLANN original, both performance measures selected the same values for the 2 hyperparameters (size and decay). On the other hand, separate hyperparameters were selected for PLANNCR extended on a 5D space (nodesize, dropout rate, learning rate, momentum, weak class weight). The technical details can be found in Additional file 2. From the 3 activation functions tested for the hidden layer (“sigmoid”, “relu”, “tanh”), the “sigmoid” provided the best performance on the training data for both Brier score and AUC. A weak class weight of 1 was selected (no adjustment for disease progression or death).
The performance of the tuned PLANNCR extended was compared for disease progression (event of interest). Results are presented in Table 2. PLANNCR extended tuned with Brier score at 5 years had a better performance in terms of Brier score and miscalibration at 2, 5, or 10 years. However, PLANNCR extended tuned with AUC at 5 years had a better performance regarding AUC at 5 and 10 years. These results were expected as Brier score is a more complete measure taking into account both discrimination and calibration. For the rest of the results presented below, optimal combinations for Brier score at 5 years were selected for PLANNCR extended.
Predictive performance comparison
In this section, the five methods are compared on the 100 validation datasets for different predictive performance measures: (1) Brier scores, (ii) AUC, (iii) miscalibration at 2, 5, and 10 years, respectively, for disease progression (local recurrence or distant metastasis). Optimal hyperparameters and additional plots for the event of interest (disease progression) and the competing event (death) are included in sections 1 and 2 of Additional file 3.
Brier score  AUC
Figure 2 shows the Brier score (lower values better) and AUC (higher values better) at 2, 5 and 10 years since surgery for all methods regarding disease progression.
For the timedependent Brier score, the causespecific Cox model had in general the best performance followed by the FineGray model and RSFCR at 2 years, and the PLANNCR extended and FineGray at 5 and 10 years. PLANNCR original had slightly the worst performance at these time points. 95% confidence intervals (CI) based on the percentile method for 100 validation datasets using the outofbag data overlapped. PLANNCR extended had marginally larger 95% CI at 2 years and RSFCR at 10 years. Regarding AUC at 2, 5, and 10 years, the causespecific Cox model and the PLANNCR extended had the best performance (very close to each other) followed by FineGray model, RSFCR and PLANNCR original in decreasing order of performance. The 95% confidence intervals were very similar for the methods, except for PLANNCR original which had much wider intervals at all times. This means that its discrimination ability (AUC) was not consistent (fluctuated) in the validation datasets.
Figure S1 in Additional file 3 provides the same plot with PLANNCR extended tuned with AUC at 5 years. The predictive ability decreased in terms of Brier score but slightly increased regarding AUC at 5 and 10 years (see also Table 2). Figures S3 and S5 in Additional file 3 illustrate the prognostic ability (Brier score, AUC) of all models for death (the competing event). The SM (causespecific Cox and FineGray) had the lowest Brier score followed by the RSFCR. PLANNCR models had worse performance and larger CI than the rest at 2 years. PLANNCR original continued to have larger CIs at 5 and 10 years, whereas PLANNCR extended had narrower CIs at 5 and 10 years (more consistent performance). For AUC, the causespecific Cox model and the PLANNCR extended had the highest values followed by the FineGray model and the RSFCR. PLANNCR original the lowest performance and the largest 95% CI.
Miscalibration
The five models were investigated in terms of miscalibration (definition in section “Predictive performance assessment”) at 2, 5, and 10 years. Results are depicted in Fig. 3 with boxplots. The SM (causespecific Cox model, FineGray) had by far the lowest miscalibration error at 2 years for disease progression (cause 1). The SM and then the PLANNCR original had the lowest miscalibration at 5 years (the SM and PLANNCR extended at 10 years). PLANNCR extended had the highest miscalibration error at 2 years, the second highest at 5 years and the lowest at 10 years (next to causespecific Cox model for this time point). The RSFCR had the worst calibration at 5 and 10 years for the cumulative incidence of the event of interest.
The miscalibration plot for PLANNCR extended tuned with AUC at 5 years is available in Additional file 3 (Fig. S2). PLANNCR extended is less well calibrated compared to Fig. 3. This result was expected since in the Supplementary figure the model was tuned for discrimination only (AUC at 5 years), whereas in Fig. 3 it was tuned taking into account both discrimination and calibration (Brier score at 5 years). Figures S4 and S6 show the miscalibration error for all five methods for the competing event (death). The causespecific Cox model and the Finegray model had the lowest miscalibration error. RSFCR show a similar miscalibration error for death at 2 and 5 year and slightly worse error at 10 years. The two neural networks had the highest miscalibration error at any time point (distinct from the other three models). A tentative explanation of the higher PLANNCR miscalibration for the competing event is that it arises from heavier regularisation of the predicted death probabilities (for a given time point) resulting in a smaller spread of the predictions there. A solution to improve the calibration could be to tune the performance of PLANNCR (e.g. Brier score at 5 years) for the competing event. However, as disease progression was of major interest here, PLANNCR original and extended were both tuned for disease progression.
Discussion
To the best of our knowledge, this is the first study which compared SM with ML techniques for CRs in softtissue sarcoma. A total of 3826 retrospectively collected patients were analysed with highgrade eSTS based on nine prognostic factors (small/medium sample size, lowdimensional setting). The SM (causespecific Cox, FineGray) and the RSFCR used exact times to event whereas the neural networks (PLANNCR original, PLANNCR extended) required a data preparation into a long format where the exact time points were turned into L separate time intervals (years). The five methods predicted the cumulative incidence of disease progression (event of interest) and death (competing event) since the date of surgery.
The results showed that the ML models have similar performance to the SM in terms of Brier score and AUC at 2, 5, and 10 years for disease progression and death (95% confidence intervals overlapped). Predictive ability of PLANNCR extended was usually better than RSFCR and PLANNCR original especially for AUC. This means that PLANNCR extended had the ability to discriminate better between low and high risk groups of patients. Nevertheless, the SM were frequently better calibrated than the three ML techniques. Miscalibration of PLANNCR original and extended was more pronounced for the competing event. These findings are consistent with a simulation study of our group that compared the predictive performance of SNN (PLANN original and extensions) with Cox models for osteosarcoma data in a similar simple setting (250 or 1000 patients, five prognostic factors) [44]. Hence, more attention to model calibration (absolute predictive accuracy) is urgently needed for ML methods.
For this work, we sampled with replacement 100 times (bootstrapping) from the eSTS data to train the ML models. Then, the left out samples were used to internally validate all models’ performance and obtain empirical 95% CIs (see Fig. 1). This can be an advantageous approach when the sample size is limited because it avoids decreasing the number of patients for model development / validation. However, it comes with a cost as this procedure is repeated multiple times and is therefore computationally expensive. The performance of all models was assessed with two timedependent measures: Brier score (discrimination and calibration) and AUC (discrimination) at 2, 5, and 10 years, respectively. We chose the timedependent AUC over the adaptation of Harrell’s concordance index to the CRs setting [45, 46]  a global performance measure for discrimination  since the latter is not a proper measure for the evaluation of tyear predicted risks (see [47]).
Two regression models for CRs were applied for the comparison with ML techniques; the causespecific hazard regression Cox and the FineGray. The causespecific Cox model might be better suited for addressing etiological questions, whereas the FineGray for estimating the clinical prognosis of patients  which was the aim here [3, 5, 48]. Nonetheless, both SM were employed for a more comprehensive approach, providing similar results, and outperforming the ML models in calibration. Complex functional dependencies such as nonlinear and nonadditive effects were not investigated, which shows how effective the SM can be in simple settings (with small/medium sample size and limited number of predictors) despite they assume additivity of effects and proportionality of hazards over time. On the other hand, ML methods may be very flexible (no a priori modelling assumptions), but usually require (very) large datasets to ensure small overfitting of their developed clinical prediction models [49, 50].
Other MLdriven models have been recently proposed for survival analysis with CRs and their prognostic ability was compared with typical benchmarks such as the causespecific Cox, FineGray, and RSFCR. In 2017, Alaa and van der Schaar [51] proposed a nonparametric Bayesian model to jointly assess a patient’s risk of multiple competing adverse events. The patient’s causespecific survival times are modelled as a function of the covariates using deep multitask Gaussian processes. Bellot and van der Schaar [52] developed in 2018 a treebased Bayesian mixture model for CRs. They constructed a hierarchical Bayesian mixture model through multivariate random survival forests and evaluated the importance of variables for each cause. Recently, a deep neural network (multiple hidden layers) was employed by Nagpal et al. called deep survival machines [53]. This is a parametric methodology to jointly learn a common deep nonlinear representation of the input features. This network estimates separately the event distribution for each CR. Note that for this project, we only specified shallow neural networks (1 hidden layer) to avoid excessive danger of overfitting in this simple setting.
Focusing on the practical utility, the two SM have the advantage compared to three ML techniques examined. The latter require a substantial implementation time for data preprocessing, tuning of the parameters, and are computationally more intensive to run (in terms of hours here). At the same time model optimisation of PLANNCR is a delicate task which requires robust numerical methods and skillful use, else the network might converge in suboptimal minima in the error function [35]. From the three ML techniques, PLANNCR extended demanded more time and effort for training because of the larger number of tuning parameters (five versus two for PLANNCR original and RSFCR). On the contrary, the causespecific Cox and FineGray models do not require any hyperparameter tuning and offer a fast implementation.
Nowadays, the employment of ML is overhyped in some contexts of medicine due to the increased interest in applying modern techniques to create prediction models. Therefore, it is necessary to report prediction models powered by artificial intelligence completely and transparently to allow critical appraisal, reproducibility of the modelling steps and results by a wider audience, and to avoid research waste [14, 15, 54]. In general, a traditional regression approach may still provide more accurate predicted survival probabilities and prognostic performance compared to a stateoftheart ML model, especially in noncomplex medical settings (lowmedium sample size, small number of predictors). In this instance, application of ML algorithms should only be motivated for exploration of the collected data.
In the future, it might be useful to compare the predictive ability of the causespecific proportional hazard Cox model with the PLANNCR original / extended for timedependent variables. The first method allows the inclusion of timedependent covariates in standard software, and the second can naturally incorporate timedependent covariates due to the essential data transformation into a long format for each patient. Moreover, the FineGray and RSFCR can be extended to provide dynamic predictions with timedependent covariates for CRs by creating a landmark dataset at a set of landmark time points \(t_{LM}\) [55]. Last but not least, it would be interesting to compare the SM and ML techniques regarding interpretation. Overall, SM offer a more straightforward interpretation via causespecific hazard ratios, while PLANNCR can provide the shape of the causespecific hazard function over time and covariates, and RSFCR the variable importance. Further research is needed on a common metric to directly compare all methods.
Conclusions
In this article, we discussed ML alternatives (PLANNCR original, PLANNCR extended, RSFCR) to SM (causespecific Cox model, FineGray) to build prognostic models for survival analysis with CRs in eSTS data with small/medium sample size and limited number of predictors (simple setting). Methods were compared in terms of discrimination and calibration. ML models reached an equivalent performance in terms of suitable predictive performance measures at 2, 5, or 10 years since surgery (95% confidence intervals overlapped), but the conventional regression models were generally better calibrated. Hence, more attention to calibration is needed. Modern MLdriven techniques are less practical as they require substantial implementation time (data preprocessing, hyperparameter tuning, computational intensity), whereas regression models are straightforward to use and can perform well without the additional workload of model training. Overall, complete and transparent reporting of all methods is required to allow critical appraisal, reproducibility, and avoid research waste. In our opinion, for noncomplex real life data such as this, ML techniques should only be employed complementary to SM as exploratory tools of model’s performance.
Availability of data and materials
The clinical data used for this research project is private. The Rcode developed to perform this analysis is provided in the following GitHub repository https://github.com/GKantidakis/SMvsMLforCRs. The reader will also find a zip file with Rcodes, which is a comprehensive example of this analysis in publicly available R data for Follicular Cell Lymphoma (data “follic”). The analysis plots of the “follic” data (n = 541, p = 4), which illustrate the same methodologies within noncomplex data, support the findings of the eSTS data.
Abbreviations
 AUC:

Area Under the Curve
 CI:

Confidence Interval
 CIF:

Cumulative Incidence Function
 CRs:

Competing risks
 eSTS:

Extremity softtissue sarcomas
 IPCW:

Inverse Probability of Censoring Weighting
 KM:

KaplanMeier
 ML:

Machine learning
 MSE:

Mean squared error
 PLANN:

Partial logistic artificial neural network
 PLANNARD:

Partial logistic artificial neural network  automatic relevance determination
 PLANNCR:

Partial logistic artificial neural network for competing risks
 PLANNCRARD:

Partial logistic artificial neural network for competing risks  automatic relevance determination
 ReLU:

Rectified linear unit
 RSF:

Random survival forests
 RSFCR:

Random survival forests for competing risks
 SM:

Statistical models
 SNNs:

Survival neural networks
References
Putter H, Fiocco M, Geskus RB. Tutorial in biostatistics: competing risks and multistate models. Stat Med. 2007;26(11):2389–430. https://doi.org/10.1002/SIM.2712.
Varadhan R, Weiss CO, Segal JB, Wu AW, Scharfstein D, Boyd C. Evaluating health outcomes in the presence of competing risks: A review of statistical methods and clinical applications. Med Care. 2010;48(6 SUPPL.):96–105. https://doi.org/10.1097/MLR.0b013e3181d99107.
Geskus RB. Data Analysis with Competing Risks and Intermediate States.1st ed. Boca Raton: Chapman and Hall/CRC; 2015.
Zhang Z, Cortese G, Combescure C, Marshall R, Lim M, et al. Overview of model validation for survival regression model with competing risks using melanoma study data. Ann Transl Med. 2018;6(16):325. https://doi.org/10.21037/atm.2018.07.38.
Austin PC, Lee DS, Fine JP. Introduction to the Analysis of Survival Data in the Presence of Competing Risks. Circulation. 2016;133(6):601–9. https://doi.org/10.1161/CIRCULATIONAHA.115.017719.
Austin PC, Fine JP. Accounting for competing risks in randomized controlled trials: a review and recommendations for improvement. Stat Med. 2017;36(8):1203–9. https://doi.org/10.1002/sim.7215.
Koller MT, Raatz H, Steyerberg W, Wolbers M. Competing risks and the clinical community : irrelevance or ignorance? Stat Med. 2012;31(11–12):1089–97. https://doi.org/10.1002/sim.4384.
Kaplan EL, Meier P. Nonparametric Estimation from Incomplete Observations. J Am Stat Assoc. 1958;53(282):457–81. https://doi.org/10.2307/2281868.
Zhang Z. Survival analysis in the presence of competing risks. Ann Transl Med. 2016;5(3). https://doi.org/10.21037/atm.2016.08.62.
Cox DR. Regression Models and LifeTables. J R Stat Soc Ser B (Methodol). 1972;34(2):187–220.http://www.jstor.org/stable/2985181.
Fine JP, Gray RJ. A Proportional Hazards Model for the Subdistribution of a Competing Risk. J Am Stat Assoc. 1999;94(446):496–509. https://doi.org/10.1080/01621459.1999.10474144.
Kourou K, Exarchos TP, Exarchos KP, Karamouzis MV, Fotiadis DI. Machine learning applications in cancer prognosis and prediction. Comput Struct Biotechnol J. 2015;13:8–17. https://doi.org/10.1016/j.csbj.2014.11.005.
SideyGibbons JAM, SideyGibbons CJ. Machine learning in medicine: a practical introduction. BMC Med Res Methodol. 2019;19(1):1–18. https://doi.org/10.1186/s1287401906814.
Collins GS, Reitsma JB, Altman DG, Moons KGM. Transparent reporting of a multivariable prediction model for individual prognosis or diagnosis (TRIPOD): The TRIPOD Statement. BMC Med. 2015;13(1). http://www.biomedcentral.com/17417015/13/1. https://doi.org/10.1186/s129160140241z.
Collins GS, Collins KGM. Reporting of artificial intelligence prediction models. Lancet. 2019;393(10181):1577–9. https://doi.org/10.1016/S01406736(19)300376.
Shahid N, Rappon T, Berta W. Applications of artificial neural networks in health care organizational decisionmaking: A scoping review. PLoS ONE. 2019;14(2):e0212356. https://doi.org/10.1371/journal.pone.0212356.
Wang P, Li Y, Reddy CK. Machine learning for survival analysis: A survey. ACM Comput Surv. 2019;51(6):1–36. https://doi.org/10.1145/3214306.
Biganzoli E, Boracchi P, Mariani L, Marubini E. Feed forward neural networks for the analysis of censored survival data: a partial logistic regression approach. Stat Med. 1998;17(10):1169–86. https://doi.org/10.1002/(SICI)10970258(19980530)17:10<1169::AIDSIM796>3.0.CO;2D.
Lisboa PJG, Wong H, Harris P, Swindell R. A Bayesian neural network approach for modelling censored data with an application to prognosis after surgery for breast cancer. Artif Intell Med. 2003;28(1):1–25. https://doi.org/10.1016/S09333657(03)000332.
Kantidakis G, Putter H, Lancia C, de Boer J, Braat AE, Fiocco M. Survival prediction models since liver transplantation  comparisons between Cox models and machine learning techniques. BMC Med Res Methodol. 2020;20(1):1–14. https://doi.org/10.1186/s12874020011531.
Ishwaran H, Kogalur UB, Blackstone EH, Lauer MS. Random survival forests. Ann Appl Stat. 2008;2(3):841–60. https://doi.org/10.1214/08AOAS169.
Breiman L. Random Forests. Mach Learn. 2001;45(1):5–32. https://doi.org/10.1023/A:1010933404324.
Biganzoli E, Boracchi P, Ambrogi F, Marubini E. Artificial neural network for the joint modelling of discrete causespecific hazards. Artif Intell Med. 2006;37(2):119–30. https://doi.org/10.1016/j.artmed.2006.01.004.
Lisboa PJG, Etchells TA, Jarman IH, Arsene CTC, Aung MSH, Eleuteri A, et al. Partial logistic artificial neural network for competing risks regularized with automatic relevance determination. IEEE Trans Neural Netw. 2009;20(9):1403–16. https://doi.org/10.1109/TNN.2009.2023654.
Ishwaran H, Gerds TA, Kogalur UB, Moore RD, Gange SJ, Lau BM. Random survival forests for competing risks. Biostatistics. 2014;15(4):757–73. https://doi.org/10.1093/biostatistics/kxu010.
Team RC. R: A Language and Environment for Statistical Computing. R Foundation for Statistical Computing. 2014. http://www.rproject.org/.
RuetenBudde AJ, van Praag VM, van de Sande MAJ, Fiocco M, Aston W, Bonenkamp H, et al. External validation and adaptation of a dynamic prediction model for patients with highgrade extremity soft tissue sarcoma. J Surg Oncol. 2021;123(4):1050–6. https://doi.org/10.1002/jso.26337.
Mariani L, Miceli R, Kattan MW, Brennan MF, Colecchia M, Fiore M, et al. Validation and adaptation of a nomogram for predicting the survival of patients with extremity soft tissue sarcoma using a threegrade system. Cancer. 2005;103(2):402–408. https://pubmed.ncbi.nlm.nih.gov/15578681/. https://doi.org/10.1002/CNCR.20778.
van Praag VM, RuetenBudde AJ, Jeys LM, Laitinen M, Pollock R, Aston W, et al. A prediction model for treatment decisions in highgrade extremity softtissue sarcomas: Personalised sarcoma care (PERSARC). Eur J Cancer. 2017;83:313–23. https://doi.org/10.1016/j.ejca.2017.06.032.
Callegaro D, Miceli R, Bonvalot S, Ferguson P, Strauss DC, Levy A, et al. Impact of perioperative chemotherapy and radiotherapy in patients with primary extremity soft tissue sarcoma: retrospective analysis across major histological subtypes and major reference centres. Eur J Cancer. 2018;105:19–27. https://doi.org/10.1016/j.ejca.2018.09.028.
Woll PJ, Reichardt P, Le Cesne A, Bonvalot S, Azzarelli A, Hoekstra HJ, et al. Adjuvant chemotherapy with doxorubicin, ifosfamide, and lenograstim for resected softtissue sarcoma (EORTC 62931): A multicentre randomised controlled trial. Lancet Oncol. 2012;13(10):1045–54. https://doi.org/10.1016/S14702045(12)703467.
Stekhoven DJ, Bühlmann P. MissforestNonparametric missing value imputation for mixedtype data. Bioinformatics. 2012;28(1):112–8. https://doi.org/10.1093/bioinformatics/btr597.
Holt JD. Competing risk analyses with special reference to matched pair experiments. Biometrika. 1978;65(1):159–65. https://doi.org/10.1093/BIOMET/65.1.159.
Aalen OO, Johansen S. An Empirical Transition Matrix for NonHomogeneous Markov Chains Based on Censored Observations. Scand J Stat. 1978;5(3):141–150. https://www.jstor.org/stable/4615704.
Bishop CM. Pattern recognition and machine learning. New York: Springer; 2006.
Chollet F. keras. R Foundation for Statistical Computing. 2015. https://github.com/kerasteam/keras.
Hastie T, Tibshirani R, Friedman J. The Elements of Statistical Learning: Data Mining, Inference, and Prediction. 2nd ed. Springer Series in Statistics. Springer; 2009. https://doi.org/10.1007/9780387848587.
Blanche P, ProustLima C, Loubère L, Berr C, Dartigues JF, JacqminGadda H. Quantifying and comparing dynamic predictive accuracy of joint models for longitudinal marker and timetoevent in presence of censoring and competing risks. Biometrics. 2015;71(1):102–13. https://doi.org/10.1111/biom.12232.
Schoop R, Beyersmann J, Schumacher M, Binder H. Quantifying the predictive accuracy of timetoevent models in the presence of competing risks. Biom J. 2011;53(1):88–112. https://doi.org/10.1002/bimj.201000073.
Blanche P, Dartigues JF, JacqminGadda H. Estimating and comparing timedependent areas under receiver operating characteristic curves for censored event times with competing risks. Stat Med. 2013;32(30):5381–97. https://doi.org/10.1002/sim.5958.
Graf E, Schmoor C, Sauerbrei W, Schumacher M. Assessment and comparison of prognostic classification schemes for survival data. Stat Med. 1999;18(1718):2529–2545. http://www.ncbi.nlm.nih.gov/pubmed/10474158. https://doi.org/10.1002/(SICI)10970258(19990915/30)18:17/18<2529::AIDSIM274>3.0.CO;25.
Steyerberg EW, Vickers AJ, Cook NR, Gerds T, Gonen M, Obuchowski N, et al. Assessing the performance of prediction models: A framework for some traditional and novel measures. Epidemiology. 2010;21(1):128–138. https://pubmed.ncbi.nlm.nih.gov/20010215/. https://doi.org/10.1097/EDE.0b013e3181c30fb2.
Gerds TA, Andersen PK, Kattan MW. Calibration plots for risk prediction models in the presence of competing risks. Stat Med. 2014;33(18):3191–203. https://doi.org/10.1002/sim.6152.
Kantidakis G, Biganzoli E, Putter H, Fiocco M. A Simulation Study to Compare the Predictive Performance of Survival Neural Networks with Cox Models for Clinical Trial Data. Comput Math Methods Med. 2021;2021:1–15. https://doi.org/10.1155/2021/2160322.
Wolbers M, Koller MT, Witteman JCM, Steyerberg EW. Prognostic models with competing risks methods and application to coronary risk prediction. Epidemiology. 2009;20(4):555–61. https://doi.org/10.1097/EDE.0b013e3181a39056.
Wolbers M, Blanche P, Koller MT, Witteman JCM, Gerds TA. Concordance for prognostic models with competing risks. Biostatistics. 2014;15(3):526–39. https://doi.org/10.1093/biostatistics/kxt059.
Blanche P, Kattan MW, Gerds TA. The cindex is not proper for the evaluation of tyear predicted risks. Biostatistics. 2019;20(2):347–57. https://doi.org/10.1093/biostatistics/kxy006.
Tullio A, Magli A, Moretti E, Valent F. Why we should take care of the competing risk bias in survival analysis: A phase II trial on the toxicity profile of radiotherapy for prostate cancer. Rep Pract Oncol Radiother. 2019;24(6):511–519. https://doi.org/10.1016/j.rpor.2019.08.001.
Van Der Ploeg T, Austin PC, Steyerberg EW. Modern modelling techniques are data hungry: A simulation study for predicting dichotomous endpoints. BMC Med Res Methodol. 2014;14(1):1–13. https://doi.org/10.1186/1471228814137.
Riley RD, Ensor J, Snell KIE, Harrell FE, Martin GP, Reitsma JB, et al. Calculating the sample size required for developing a clinical prediction model. BMJ. 2020;368(March):1–12. https://doi.org/10.1136/bmj.m441.
Alaa AM, Van Der Schaar M. Deep multitask Gaussian processes for survival analysis with competing risks. Adv Neural Inf Process Syst. 2017;2326–2334. http://medianetlab.ee.ucla.edu/papers/AlaaDeepCompetingRisk.pdf.
Bellot A, van der Schaar M. Treebased Bayesian mixture model for competing risks. Int Conf Artif Intell Stat PMLR 2018. 2018;910–918. http://proceedings.mlr.press/v84/bellot18a/bellot18a.pdf.
Nagpal C, Li X, Dubrawski A. Deep Survival Machines: Fully Parametric Survival Regression and Representation Learning for Censored Data with Competing Risks. IEEE J Biomed Health Inf. 2021;25(8):3163–75. https://doi.org/10.1109/JBHI.2021.3052441.
Dhiman P, Ma J, Navarro CA, Speich B, Bullock G, Damen JA, et al. Reporting of prognostic clinical prediction models based on machine learning methods in oncology needs to be improved. J Clin Epidemiol. 2021;138:60–72. https://doi.org/10.1016/j.jclinepi.2021.06.024.
Nicolaie MA, van Houwelingen JC, de Witte TM, Putter H. Dynamic prediction by landmarking in competing risks. Stat Med. 2013;32(12):2031–47. https://doi.org/10.1002/sim.5665.
Acknowledgements
This publication was supported by a donation from the Kom Op Tegen Kanker (Stand up to Cancer), the Flemish cancer society from Belgium. The authors would like to thank the members of the PERsonalised SARcoma Care (PERSARC) Study Group for the provision of the data used in this paper.
PERSARC Study Group
Ibtissam Acem, Will Aston, Han Bonenkamp, Ingrid M E Desar, Peter C Ferguson, Marta Fiocco, Hans Gelderblom, Anthony M Griffin, Dirk J Grünhagen, Rick L Haas, Andrew J Hayes, Lee M Jeys, Johnny Keller, Minna K Laitinen, Andreas Leithner, Katja MarettyKongstad, Rob Pollock, Anja RuetenBudde, Myles Smith, Maria A Smolle, Emelie Styring, Joanna Szkandera, PerUlf Tunn, Jos A van der Hage, Robert J van Ginkel, Winan J van Houdt, Veroniek van Praag, Michiel van de Sande, Kees Verhoef, Madeleine Willegger, Reinard Windhager, Jay S Wunder, Olga Zaikova.
Funding
Georgios Kantidakis’s work as a Fellow at EORTC Headquarters was supported by a grant from the EORTC Soft Tissue and Bone Sarcoma Group and Leiden University Medical Center (LUMC) Department of Medical Oncology. The funding sources had no role in the design of the study and collection, analysis, and interpretation of data or preparation of the manuscript.
Author information
Authors and Affiliations
Contributions
MF wrote a proposal to enquire the data. GK, and MF conceived and designed the study. GK carried out the statistical analysis. GK, HP, SL, and MF interpreted the results. GK drafted the manuscript, and HP, SL, and MF critically revised it. All authors read and approved the final version.
Corresponding author
Ethics declarations
Ethics approval and consent to participate
Ethical approval for this study was waived by the institutional review board of Leiden University Medical Center Department of Orthopaedics because clinical data was collected from medical records and were pseudoanonymised. All participants provided written informed consent in the original studies. Research was performed in accordance with the Declaration of Helsinki.
Consent for publication
Not applicable.
Competing interests
The authors declare no competing interests.
Additional information
Publisher’s Note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Supplementary Information
Rights and permissions
Open Access This article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons licence, and indicate if changes were made. The images or other third party material in this article are included in the article's Creative Commons licence, unless indicated otherwise in a credit line to the material. If material is not included in the article's Creative Commons licence and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this licence, visit http://creativecommons.org/licenses/by/4.0/. The Creative Commons Public Domain Dedication waiver (http://creativecommons.org/publicdomain/zero/1.0/) applies to the data made available in this article, unless otherwise stated in a credit line to the data.
About this article
Cite this article
Kantidakis, G., Putter, H., Litière, S. et al. Statistical models versus machine learning for competing risks: development and validation of prognostic models. BMC Med Res Methodol 23, 51 (2023). https://doi.org/10.1186/s1287402301866z
Received:
Accepted:
Published:
DOI: https://doi.org/10.1186/s1287402301866z