 Research article
 Open Access
 Published:
Accounting for parameter uncertainty in the definition of parametric distributions used to describe individual patient variation in health economic models
BMC Medical Research Methodology volume 17, Article number: 170 (2017)
Abstract
Background
Parametric distributions based on individual patient data can be used to represent both stochastic and parameter uncertainty. Although general guidance is available on how parameter uncertainty should be accounted for in probabilistic sensitivity analysis, there is no comprehensive guidance on reflecting parameter uncertainty in the (correlated) parameters of distributions used to represent stochastic uncertainty in patientlevel models. This study aims to provide this guidance by proposing appropriate methods and illustrating the impact of this uncertainty on modeling outcomes.
Methods
Two approaches, 1) using nonparametric bootstrapping and 2) using multivariate Normal distributions, were applied in a simulation and case study. The approaches were compared based on pointestimates and distributions of timetoevent and health economic outcomes. To assess sample size impact on the uncertainty in these outcomes, sample size was varied in the simulation study and subgroup analyses were performed for the casestudy.
Results
Accounting for parameter uncertainty in distributions that reflect stochastic uncertainty substantially increased the uncertainty surrounding health economic outcomes, illustrated by larger confidence ellipses surrounding the costeffectiveness pointestimates and different costeffectiveness acceptability curves. Although both approaches performed similar for larger sample sizes (i.e. n = 500), the second approach was more sensitive to extreme values for small sample sizes (i.e. n = 25), yielding infeasible modeling outcomes.
Conclusions
Modelers should be aware that parameter uncertainty in distributions used to describe stochastic uncertainty needs to be reflected in probabilistic sensitivity analysis, as it could substantially impact the total amount of uncertainty surrounding health economic outcomes. If feasible, the bootstrap approach is recommended to account for this uncertainty.
Background
Clinical decisionmaking is aiming towards patientspecific and preferencesensitive treatment, based on multiple biomarkers for treatment targeting and monitoring patients’ response to treatment [1]. Consequently, there is an increasing need for corresponding patientlevel models to accurately represent clinical practice when estimating the health economic impact of novel healthcare interventions [2, 3]. To facilitate decision making, such models should adequately reflect all types of uncertainty in the synthesized evidence used for analysis [4]. This is particularly relevant in patientlevel modeling studies in which reflecting patient heterogeneity may effectively increase uncertainty, for example by relatively low sample sizes in defined subgroups or by an increasing number of parameters that need to be estimated to account for patient characteristics in individualized predictions.
Uncertainty in evidence can be disaggregated into stochastic uncertainty (i.e. patientlevel variation or firstorder uncertainty) and parameter uncertainty (i.e. secondorder uncertainty) [4]. This can be illustrated using a timetoevent parameter, e.g. the timetoprogression after surgery. If, for a certain patient group, this parameter is defined by a mean estimate (Table 1 – Box A), the parameter uncertainty in this mean estimate needs to be accounted for, which can be done using a parametric distribution in the probabilistic sensitivity analysis (PSA) (Table 1 – Box B) [4]. For example, this parameter uncertainty could be reflected by defining a Normal distribution for the mean timetoprogression based on the estimated mean and standard error, by applying the Central Limit Theorem [5]. Additionally, it is possible to account for stochastic uncertainty by using parametric distributions to describe individual patient variation, such as typically used in patientlevel statetransition models [6] and discrete event simulation (DES) models [7]. For example, a Weibull distribution [8] can be used in a DES model to derive and assign patientspecific timetoprogression values (Table 1 – Box C), rather than assigning the estimated mean timetoprogression to all patients (Table 1 – Box A). As for the parameter defined by the mean estimate in Box A of Table 1, the uncertainty in the parameters defining this Weibull distributions needs to be accounted for [4]. However, there is currently no clear guidance on how to reflect parameter uncertainty (i.e. secondorder uncertainty) in these parametric distributions used to describe stochastic uncertainty (i.e. patientlevel variation or fistorder uncertainty) in PSA (Table 1 – Box D).
Reflecting parameter uncertainty in parametric distributions is challenging because these distributions are typically defined by multiple parameters, e.g. the shape and scale parameter of the Weibull distribution, with their values preferably estimated from individual patient data (IPD). When estimated from IPD, these distributions’ parameters are correlated for most of the parametric distributions commonly used in health economic analyses, such as Gamma, Beta, LogNormal, and Weibull distributions [8]. Consequently, it is incorrect to estimate the values of, for example, the shape and scale parameters independently, and to define independent (i.e. separate) distributions for each of these parameters’ uncertainty in PSA. Therefore, guidance is needed towards approaches that maintain the correlation in the parameters of distributions reflecting stochastic uncertainty when accounting for the parameter uncertainty these parameters introduce.
The lack of specific guidance is illustrated by recently published modeling studies that do not fully utilize IPD while performing PSA, for example, by adjusting predicted values (stochastic uncertainty) with a random percentage [9], by excluding correlation between parametric distributions’ parameters [10], or altogether not reporting (including) parameter uncertainty in parametric distributions [11]. As a consequence of inadequately reflecting all types of uncertainty in the synthesized evidence used for analysis, suboptimal resource allocation and research prioritization decisions may be made due to biased outcomes of PSA, overestimated confidence in the corresponding expected values of the PSA, and ensuing biased estimation of the value of collecting additional evidence to better inform decision making [12].
The objective of this study is to provide explicit guidance for health economic modelers on how parameter uncertainty in parametric distributions used to describe stochastic uncertainty (i.e. patientlevel variation) can be considered in PSA. In order to do so, two alternative solutions are illustrated and compared: 1) a nonparametric bootstrapping approach and 2) an approach using multivariate Normal distributions. Both approaches are used in a DES simulation study and in a DES case study in which stochastic uncertainty in timetoevent data is described by parametric distributions. Additionally, the potential impact of increased parameter uncertainty due to subgroup stratification on health economic outcomes is illustrated.
Methods
Consider the scenario in which timetoevent observations from a clinical study are used to describe stochastic uncertainty by fitting a Gamma distribution [8], i.e. the distributions’ (correlated) shape and rate parameter values are estimated from the IPD. Figure 1a shows an example of the estimated values of these parameters, including the 95% confidence ellipse, representing the sets (i.e. combinations) of parameter values. If uncertainty in the parameter estimates (i.e. parameter uncertainty) is ignored by using the parameters’ pointestimates in each run of the PSA, individual timetoevent values would be repeatedly drawn based on the same mean density curve of Fig. 1b in all runs of the PSA. However, if this parameter uncertainty is reflected by using different sets of correlated parameter estimates for each run of the PSA, a variety of distributions is simulated, which is illustrated by the 95% confidence interval (surface) in Fig. 1b. As this variety of distributions reflects uncertainty in the individual timetoevent values as well as overall (mean) timetoevent values, this uncertainty can be seen as parameter uncertainty (in this case of the ‘hyperparameters’, i.e. the parameters defining the timetoevent distribution). Both proposed approaches, which are introduced subsequently, can be used to generate these correlated sets of distributions’ parameter values based on IPD. A different set of generated parameters values can be used in the PSA, incorporating one set of correlated values in each Monte Carlo sample.
Nonparametric bootstrapping (Bootstrap approach)
Nonparametric bootstrapping is a statistical technique that can be used to construct an approximate sampling distribution of a statistic of interest, without the need for assumptions regarding the distribution of this statistic [13]. It has been applied in health economics, for example, to construct confidence intervals for the incremental net benefit in economic evaluations alongside clinical trials [5, 14]. In the Bootstrap approach, the distributions’ parameters are repeatedly estimated based on different bootstrap samples of the original dataset, which are obtained by resampling the original dataset with replacement, such that the size of the bootstrap sample equals the size of the original dataset [13]. A detailed discussion of the Bootstrap approach is provided in Additional file 1. Briefly, reflecting parameter uncertainty with this approach consists of the following four steps:

(1)
Generate a feasible* bootstrap sample of the original dataset, by resampling this dataset with replacement, such that the sample size of the bootstrap sample equals that of the original dataset.

(2)
Fit the prespecified distribution(s)** to the bootstrap sample and record the estimated parameter values.

(3)
Repeat (1) and (2) r times, where r equals the required number of PSA runs.

(4)
Perform the PSA, using a different set of parameter values to define the distribution(s) for each PSA run.
* The definition of feasible bootstrap samples may vary between studies. Please see Additional file 1 for a more indepth discussion.
** Note that if multiple distributions are fitted in step (2), all distributions need to be fitted on the same bootstrap sample to preserve correlation among all distributions and other parameters used to describe variables in the dataset.
Multivariate normal distributions (MVNorm approach)
The MVNorm approach assumes the distributions’ parameters to be Normal distributed, which is valid for sufficiently large sample sizes according to the Central Limit Theorem [5], as suggested by Briggs et al. for regression models in general [15]. Given a distribution’s parameter estimates and their variancecovariance matrix, a multivariate Normal distribution can be defined and used to draw correlated sets of parameter values. A detailed discussion of the MVNorm approach is provided in Additional file 1. Briefly, reflecting parameter uncertainty with this approach consists of the following four steps:

(1)
Fit the prespecified distribution to the original dataset and record the estimated parameter values and (calculate) the variancecovariance matrix.

(2)
Define a multivariate Normal distribution from the parameters’ estimates and their variancecovariance matrix according to (1).

(3)
Draw r feasible* sets of parameter values from the defined distribution (2), where r equals the required number of PSA runs.

(4)
Perform the PSA, using a different set of parameter values to define the distribution(s) for each PSA run.
* The drawn sets of parameter values need to be assessed for their feasibility, i.e. whether the parameter values are appropriate for the prespecified distributions. Please see Additional file 1 for a more indepth discussion.
Simulation study
A simulation study was performed to assess potential differences in the performance of both approaches and compare them to the scenario in which parameter uncertainty in the timetoevent distributions was not considered. This simulation study was performed in R Statistical Software version 3.3.2 [16] and used a basic health economic DES model to compare two treatment strategies in terms of health economic outcomes. This health economic DES model was structured according to a basic three state disease progression model, i.e. healthy, progressed, and death, and included two competing risks for patients in the healthy state, i.e. progression and death (Fig. 2). Timetoevent data was simulated using Weibull distributions, separately for the intervention and control patient populations, which differed in terms of survival and treatment costs in the progressed state. The exact definitions of the timetoevent distributions and cost and effect parameters are provided in Additional file 1.
The simulation study was carried out for several sample sizes, i.e. 500, 100, 50, and 25 patients, performing 2500 simulation runs (i.e. hypothetical trials) of 5000 PSA runs, each including 5000 patients per treatment strategy, for each sample size (Fig. 3). Weibull distributions were used to describe stochastic uncertainty in timetoevent data to avoid potential bias due to mismatching distributions, as Weibull distributions were also used to simulate the hypothetical populations. Distributions were fitted using the fitdist function of the fitdistrplus package [17] in R Statistical Software [16]. Stochastic uncertainty in the model’s cost and effect parameters was deliberately not considered to represent the common scenario in which IPD is not available for all model parameters, and PSA samples for these parameters were generated according to the distributions as defined in Additional file 1. Random draws from a multivariate Normal distribution were performed using the mvrnorm function of the MASS package [18].
Estimations of the “true” mean value and distribution of the parameters and health economic outcomes were obtained by analyzing the model for 2500 different samples from the simulated populations, with corresponding sample sizes. This resembles the scenario in which the population values, i.e. values based on observing the total population, would be approximated in practice by performing 2500 clinical studies. These “true” results were used to compare the generated sets of distributions’ parameter values and the health and economic outcomes using either of both approaches. The distributions of the timetoevent distributions’ parameter values and health economic outcomes were compared based on their relative entropy, i.e. the KullbackLeibler divergence [19], using the KLdiv function of the flexmix package [20,21,22]. The relative entropy is a measure of the difference between two probability distributions, for which lower values indicate a better match of distributions. Additionally, the impact of considering the parameter uncertainty in the timetoevent distributions’ parameters was illustrated in an incremental costeffectiveness plane for one hypothetical trial, i.e. one random run out of 2500 simulation runs, and in mean costeffectiveness acceptability curves (CEACs), including a 95% confidence interval, based on all simulation runs.
Case study
To illustrate how parameter uncertainty in timetoevent distributions’ parameter estimates could impact health economic outcomes in practice, a case study was performed based on the randomized phase 3 CAIRO3 study (NCT00442637) that was carried out by the Dutch Colorectal Cancer Group [23]. A total of 558 metastatic colorectal patients with stable disease or better after six cycles of capecitabine, oxaliplatin, and bevacizumab (CAPOXB) induction therapy were randomized to either receive capecitabine and bevacizumab maintenance treatment or observation until progression of disease. CAPOXB treatment was to be reintroduced upon progression on either maintenance or observation, and continued until second progression (PFS2) the primary endpoint of the study.
A previous developed DES model was adapted to use the sets of distributions’ parameter values generated by the Bootstrap and MVNorm approach in the PSA (Additional file 1). The model was developed and validated in AnyLogic multimethod simulation software [24] according to good research practices guidelines [7, 25, 26], and structured according to the same health states as the statetransition model used for the original evaluation of the CAIRO3 study: postinduction, reinduction, salvage, and death [27] (Fig. 4). Eventspecific probabilities and Weibull distributions were used to describe timetoevent data and handle the competing risks of disease progression and death in the postinduction and reinduction state [28]. Parameters of the distributions used to reflect parameter uncertainty in nontimetoevent parameters, e.g. costs and utilities, were deliberately defined exactly as in the original evaluation of the CAIRO3 study for all subgroups [27], so that observed differences in outcomes could be designated specifically to the uncertainty in timetoevent distributions’ parameter estimates. The health economic outcomes for all case study analyses were presented in incremental costeffectiveness planes and CEACs based on 10,000 PSA runs of 10,000 patients per treatment strategy.
Clinical relevant subgroups were defined to resemble the personalized context described in the introduction, illustrating the impact of parameter uncertainty in timetoevent distributions’ parameter estimates on health economic outcomes for different sample sizes. Patients were stratified according to their treatment response (stable disease (SD) versus complete or partial response (CR/PR)) and stage of disease (synchronous versus metachronous), which resulted in a total of 8 subgroups with sample sizes ranging from 50 to 410 (Table 2). As illustrated in Table 2, this stratification created subgroups in which events were observed only once, or not all, which prohibits fitting of a Weibull distribution, and the probability of that event occurring was therefore set to zero.
Results
Simulation study
The potential impact of considering parameter uncertainty in timetoevent distributions’ parameter estimates in PSA is illustrated for several sample sizes in Fig. 5, which shows results of one single run of the simulation study. Ignoring parameter uncertainty in the timetoevent distributions’ parameter estimates leads to an underestimation of the uncertainty surrounding costeffectiveness outcomes. This is illustrated by the smaller confidence ellipse for this scenario (longdashed black ellipse) compared to other scenarios in which uncertainty in the timetoevent distributions’ parameter estimates is accounted for using one of the proposed approaches (dashed gray and dotted lightgray ellipses). The indicated effect is already observed for rather large sample sizes (i.e. n = 500) and increases as sample size decreases, illustrated by the larger distance between the longdashed black confidence ellipse and the dashed gray and dotted lightgray confidence ellipses for smaller sample sizes. Although the Bootstrap (dashed gray ellipse) and MVNorm approach (dotted lightgray ellipse) yield incremental costeffectiveness pointestimates similar to the “real” value, both approaches slightly overestimate the magnitude of the uncertainty for sample sizes of n = 100 and smaller, demonstrated by smaller confidence ellipses for the “real” uncertainty (solid black ellipse). For very small sample sizes (i.e. n = 25), the MVNorm approach generates unrealistic parameter values, e.g. indicating a mean survival far beyond lifeexpectancy, leading to extreme and unrealistic health economic outcomes, which results in an unrealistic large confidence ellipse (not presented). However, since Fig. 5 represents only one run of the simulation study, these results may not be representative for the approaches in general.
To assess the performance of the approaches in general, 2500 of these comparisons were performed in the simulation study. Results show that both approaches yield comparable mean parameter estimates and standard errors (Additional file 1). Although the MVNorm approach seems to perform slightly better for small sample sizes (n = 25) in terms of mean parameter estimates, this approach too often yields extreme and unrealistic outcomes in the health economic simulation. Considering the KullbackLeibler divergence, both approaches perform similar, though on average the MVNorm approach seems to represent the “estimated true” distributions slightly better for very small sample sizes (i.e. n = 25) (Additional file 1). Also the CEACs presented in Fig. 6 show that the results for both approaches are similar (light gray and gray lines). However, compared to the results of the strategy in which the uncertainty in timetoevent distributions’ parameter estimates is not considered (black line), both approaches yield different mean CEACs (solid lines) with different confidence intervals (dashed lines), illustrating the potential health economic impact of ignoring this uncertainty.
Case study
Incremental costeffectiveness planes for the cohort analysis and selected subgroup analyses of the case study are presented in Fig. 7. CEACs and incremental costeffectiveness planes for the cohort analysis and all subgroups analyses are available in Additional file 1. The costeffectiveness pointestimate for the cohort (Subgroup 0) is not affected by considering the parameter uncertainty in the timetoevent distributions’ parameter estimates in PSA, as the points in the corresponding incremental costeffectiveness plane overlap. However, there is a substantial increase in the amount of uncertainty surrounding this pointestimate when the uncertainty in the distributions’ parameter estimates is accounted for, illustrated by the distance between the longdashed black confidence ellipse and the dashed gray and dotted lightgray confidence ellipses. The potential impact of this increase in uncertainty is illustrated by the results for Subgroup 3. Without considering the parameter uncertainty in the distributions’ parameter estimates, the results indicate that health loss due to maintenance treatment (experimental strategy) is unlikely in this subgroup, as the corresponding (dashed black) confidence ellipse is entirely located right of the vertical axis. However, the dashed gray and dotted lightgray confidence ellipses, including this uncertainty, show there actually is a nonzero probability of health loss, as these are partly located left of the vertical axis.
As the cohort is stratified, and sample size decreases, the impact of considering parameter uncertainty in the timetoevent distributions’ parameter estimates in PSA increases substantially, which is illustrated by the results for Subgroup 4 and Subgroup 6. These results show that the increase in uncertainty due to stratification may become so large that (further) subgroup stratification might not be informative. Additionally, the results for Subgroup 6 illustrate that for small sample sizes (i.e. n = 50) the pointestimates of the costeffectiveness outcomes themselves may also be affected by the uncertainty in the distributions’ parameters, which is illustrated by the nonoverlapping points in the corresponding incremental costeffectiveness plot.
No meaningful differences in the point estimates and the magnitude of uncertainty surrounding these estimates between the Bootstrap and MVNorm approach are observed, as their pointestimates and confidence ellipse overlap to a great extent. However, as also observed in the simulation study, the use of the MVNorm approach occasionally results in extreme and unlikely parameter values, which were to be excluded from the simulation.
Discussion
As demonstrated in this paper, parameter uncertainty in parametric distributions used to describe stochastic uncertainty (i.e. patientlevel variation) should be explicitly accounted for in PSA by modelers, as it does impact incremental costeffectiveness pointestimates and CEACs. If this type of uncertainty is ignored, suboptimal resource allocations or research prioritization decisions may be made, due to an underestimation of the total uncertainty surrounding health economic outcomes. This is particularly relevant in a personalized treatment context in which patient stratification obviously leads to an increase in uncertainty on the level of subgroups compared to the level of the full patient group due to decreasing sample sizes in subgroups. This increase in uncertainty should of course be reflected in PSA, and although it is likely that many experienced modelers already do so, clear guidance on appropriate methods was not yet available. Besides timetoevent distributions, this notion also applies to other types of distributions that are used to reflect stochastic uncertainty, e.g. Gamma distributions to describe patientlevel costs. Although DES was used in both the simulation and case study, these findings apply to any patientlevel modeling method used to reflect stochastic uncertainty, e.g. microsimulation statetransition models. Reflecting all parametric uncertainty in this way does require drawing values from a different distribution in each PSA sample, providing some additional work for modelers. However, the required effort is minimal and an online tool, including tutorials, has been made available alongside this paper to easily analyze individual patient data for implementation into patientlevel models [29].
When accounting for parameter uncertainty in distributions’ parameter estimates, the Bootstrap approach has some advantages over the MVNorm approach. The Bootstrap approach seems more robust for smaller sample sizes, does not require any assumptions regarding underlying distributions, and preserves the correlation in the whole dataset. Although not distinctively illustrated in this study, the latter also concerns the correlation with other, nontimetoevent, related parameters, such as utilities and costs, and can be considered a major advantage. The Bootstrap approach does, however, require the definition of a feasible bootstrap sample. Especially in case of scarce events, which are inevitable in personalized medicine, this can be challenging. Moreover, the sets of parameters values probably need to be generated outside the software environment used for the simulation and need to be imported into this environment for performing the PSA.
Although the MVNorm approach is better capable of handling scarce events in theory, mainly because it does not require the definition of a feasible bootstrap sample, it has several severe downsides. The MVNorm approach requires the definition of feasible values for the distributions’ parameter estimates and the assumption that these estimates follow a Normal distribution, which is not appropriate for insufficiently large sample sizes. For example, the shape and rate parameter of the Gamma distribution may be skewed, depending on the IPD on which their values are estimated. Additionally, many of the commonly used parametric distributions are defined for positive parameter values only, whereas Normal distributions are defined for any real number, including negatives. Furthermore, although this approach is better capable of handling scarce events in theory, it is likely to yield extreme parameter values for smaller sample sizes due to increasing standard errors, a scenario in which the use of the MVNorm approach is advised against.
Given the advantages and disadvantages of both approaches, the Bootstrap approach is recommended over the MVNorm approach if constructing enough feasible bootstrap samples is possible. However, if modelers feel the need to use the MVNorm approach, for example because multivariate Normal distributions are supported in the used software environment and the Bootstrap approach is not, they should carefully check whether 1) the sample size is sufficient, i.e. no implausible and extreme parameter values are observed, and 2) whether correlation between defined distributions to reflect patientlevel variation is low or negligible and the additional uncertainty introduced by using independent multivariate Normal distributions (e.g. for costs, utilities, timetoevents) is therefore limited.
Several choices in the study design may have influenced the final results. Both the simulation study and case study focus on the impact of parameter uncertainty specifically in timetoevent distributions, because primary outcomes in clinical studies are often related to timetoevents, e.g. overall survival, and these distributions characterize DES. Additionally, accounting for the parameter uncertainty in nontimetoevent related distributions may further increase the amount of uncertainty surrounding the health economic outcomes, which will contribute to the conclusion that parameter uncertainty in distributions used to describe stochastic uncertainty (i.e. patientlevel variation) needs to be accounted for. Furthermore, Weibull distributions are used to describe the IPD, which is a design choice and therefore introduces structural uncertainty. This choice is not expected to meaningfully influence the outcomes, because the fitted distributions match the data well. Moreover, other distributions’ parameters may have a stronger correlation, which would further stress the need to use one of the proposed approaches to appropriately account for the uncertainty in these parameters.
Further research may be directed towards evaluating the impact of using different types of distributions to describe stochastic uncertainty on health economic outcomes, the uncertainty surrounding these outcomes, and the performance of both approaches. Additionally, sensitivity analyses other than PSA, such as deterministic sensitivity analysis and structural sensitivity analysis, might be considered. Furthermore, additional guidance is desirable on how uncertainty in IPD can be appropriately combined with uncertainty in aggregated data on population level, e.g. a reported mean estimate and standard error from literature, for sensitivity analyses in patientlevel models.
Conclusions
With an increasing need for patientlevel models to accurately represent clinical practice, modelers should be aware that the parameter uncertainty in parametric distributions used to describe stochastic uncertainty (i.e. patientlevel variation) should be accounted for in PSA performed in health economic modeling studies. This type of uncertainty could have a substantial impact on the total amount of uncertainty surrounding the health economic outcomes and may influence healthcare decisionmaking. At least two approaches are available to account for the parameter uncertainty in parametric distributions used to describe stochastic uncertainty. If feasible, the Bootstrap approach is recommended to account for this type of uncertainty.
Abbreviations
 Bootstrap:

The bootstrap approach
 CAPOXB:

Capecitabine, oxaliplatin, and bevacizumab
 CEAC:

Costeffectiveness acceptability curve
 CR:

Complete response
 DES:

Discrete event simulation
 IPD:

Individual patient data
 MVNorm:

The multivariate Normal distributions approach
 PFS2:

Time until second progression
 PR:

Partial response
 PSA:

Probabilistic sensitivity analysis
 SD:

Stable disease
References
 1.
Barbieri CE, Chinnaiyan AM, Lerner SP, Swanton C, Rubin MA. The emergence of precision urologic oncology: a collaborative review on biomarkerdriven therapeutics. Eur Urol. 2017;71(2):237–46.
 2.
Annemans L, Redekop K, Payne K. Current methodological issues in the economic assessment of personalized medicine. Value Health. 2013;16(6 Suppl):S20–6.
 3.
Degeling K, Koffijberg H, IJzerman MJ. A systematic review and checklist presenting the main challenges for health economic modeling in personalized medicine: towards implementing patientlevel models. Expert Rev Pharmacoecon Outcomes Res. 2017;17(1):17–25.
 4.
Briggs AH, Weinstein MC, Fenwick EAL, Karnon J, Sculpher MJ, Paltiel AD. Model parameter estimation and uncertainty analysis: a report of the ISPORSMDM modeling good research practices task force working group–6. Med Decis Mak. 2012;32(5):722–32.
 5.
Nixon RM, Wonderling D, Grieve RD. Nonparametric methods for costeffectiveness analysis: the central limit theorem and the bootstrap compared. Health Econ. 2010;19(3):316–33.
 6.
Siebert U, Alagoz O, Bayoumi AM, Jahn B, Owens DK, Cohen DJ, et al. Statetransition modeling: a report of the ISPORSMDM modeling good research practices task Force3. Value Health. 2012;15(6):812–20.
 7.
Karnon J, Stahl J, Brennan A, Caro JJ, Mar J, Möller J. Modeling using discrete event simulation: a report of the ISPORSMDM modeling good research practices task Force4. Value Health. 2012;15(6):821–7.
 8.
Law AM. Simulation Modeling and Analysis. 4th ed. Singapore: McGrawHill Higher Education; 2007.
 9.
Chen T, Yu D, Cornelius V, et al. Potential health impact and costeffectiveness of drug therapy for prehypertension. Int J Cardiol. 2017;240:403–8.
 10.
Montgomery SM, Maruszczak MJ, Slater D, et al. A discrete event simulation to model the costutility of fingolimod and natalizumab in rapidly evolving severe relapsingremitting multiple sclerosis in the UK. J Med Econ. 2017;20:474–82.
 11.
Parikh RC, Du XL, Robert MO, et al. Costeffectiveness of treatment sequences of chemotherapies and targeted biologics for elderly metastatic colorectal cancer patients. J Manag Care Spec Pharm. 2017;23:64–73.
 12.
Claxton K. Exploring uncertainty in costeffectiveness analysis. PharmacoEconomics. 2008;26(9):781–98.
 13.
Davison AC, Hinkley DV. Bootstrap methods and their application. 1st ed. Cambridge: Cambridge University Press; 1997.
 14.
Carpenter J, Bithell J. Bootstrap confidence intervals: when, which, what? A practical guide for medical statisticians. Stat Med. 2000;19(9):1141–64.
 15.
Briggs AH, Claxton K, Sculpher MJ. Decision modeling for health economic evaluation. Oxford: Oxford University Press; 2006.
 16.
R Core Team. R: a language and environment for statistical computing. Vienna: R Foundation for Statistical Computing; 2016. https://www.rproject.org/. Accessed 22 Feb 2017
 17.
DelignetteMuller ML, Dutang C. Fitdistrplus: an R package for fitting distributions. J Stat Softw. 2015;64(4):1–34.
 18.
Venables WN, Ripley BD. Modern applied statistics with S. 4th ed. New York: Springer New York; 2002.
 19.
Cover TM, Thomas JA. Elements of information theory. 2nd ed. Hoboken: Wiley; 2006.
 20.
Grün B, Leisch F. Fitting finite mixtures of generalized linear regressions in R. Comput Stat Data Anal. 2007;51(11):5247–52.
 21.
Grün B, Leisch F. FlexMix version 2: finite mixtures with concomitant variables and varying and constant parameters. J Stat Softw. 2008;28(4):35.
 22.
Leisch F. FlexMix: a general framework for finite mixture models and latent class regression in R. J Stat Softw. 2004;11(8):1–18.
 23.
Simkens LHJ, van Tinteren H, May A, ten Tije AJ, Creemers GJM, Loosveld OJL, et al. Maintenance treatment with capecitabine and bevacizumab in metastatic colorectal cancer (CAIRO3): a phase 3 randomised controlled trial of the Dutch colorectal cancer group. Lancet. 2015;385(9980):1843–52.
 24.
The AnyLogic Company. AnyLogic multimethod simulation software. Available from: http://www.anylogic.com/. Accessed 8 Mar 2016.
 25.
Eddy DM, Hollingworth W, Caro JJ, Tsevat J, McDonald KM, Wong JB. Model transparency and validation: a report of the ISPORSMDM modeling good research practices task Force7. Value Health. 2012;15(6):843–50.
 26.
Vemer P, Corro Ramos I, van Voorn GAK, Al MJ, Feenstra TL. AdViSHE: a validationassessment tool of healtheconomic models for decision makers and model users. PharmacoEconomics. 2016;34:349–61.
 27.
Franken MD, van Rooijen EM, May AM, Koffijberg H, van Tinteren H, Mol L, ten Tije AJ, Creemers GJ, van der Velden AMT, Tanis BC, Uylde Groot CA, Punt CJA, Koopman M, van Oijen MGH. Costeffectiveness of capecitabine and bevacizumab maintenance treatment after firstline induction treatment in metastatic colorectal cancer. Eur J Cancer. 2017;75:204–12.
 28.
Barton P, Jobanputra P, Wilson J, Bryan S, Burls A. The use of modelling to evaluate new drugs for patients with a chronic condition: the case of antibodies against tumour necrosis factor in rheumatoid arthritis. Health Technol Assess. 2004;8(11):104.
 29.
iDist: Individual patient data analysis tool. https://personex.nl/tools. Accessed 4 July 2017.
Acknowledgements
The authors acknowledge prof. dr. C.J.A. Punt from the Academic Medical Center of the University of Amsterdam, who was the principal investigator of the CAIRO3 trial together with prof. dr. M. Koopman, for providing the dataset of the trial as a case study.
Funding
No funding was received for performing this research.
Availability of data and materials
An online tool to easily analyze individual patient data for implementation into patientlevel models is available at: https://personex.nl/tools. The data of the CAIRO3 trial is not publicly available.
Author information
Affiliations
Contributions
KD contributed to the research design, performed the simulation study and case study, and drafted the manuscript. MIJ contributed to the research design, supervised the statistical methodology, and provided critical revisions and discussion. MK contributed to the research design of the case study, supervised the case study, and provided critical revisions and discussion. HK contributed to the research design, supervised the simulation study and statistical methodology, and drafted the manuscript. All authors read and approved the final manuscript.
Corresponding author
Ethics declarations
Ethics approval and consent to participate
The CAIRO3 study protocol, which includes secondary data analyses such as performed in this study, was approved by the Committee on HumanRelated Research ArnhemNijmegen in the Netherlands. Written informed consent was not required for this study separately, as written informed consent was obtained from all participants in the CAIRO3 study, which includes secondary data analyses such as performed in this study.
Competing interests
The authors declare that they have no competing interests.
Publisher’s Note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Additional file
Additional file 1:
Includes: 1.1) Detailed discussion of the Bootstrap approach, 1.2) Detailed discussion of the MVNorm approach, 1.3) Description of the simulation study model, 1.4) Description of the case study model, 1.5) Parameter estimates for the simulation study, 1.6) KullbackLeibler Divergence for the simulation study, 1.7) Incremental costeffectiveness planes for all case study analyses, 1.8) CEACs for all case study analyses, and 1.9) References. (PDF 722 kb)
Rights and permissions
Open Access This article is distributed under the terms of the Creative Commons Attribution 4.0 International License (http://creativecommons.org/licenses/by/4.0/), which permits unrestricted use, distribution, and reproduction in any medium, provided you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons license, and indicate if changes were made. The Creative Commons Public Domain Dedication waiver (http://creativecommons.org/publicdomain/zero/1.0/) applies to the data made available in this article, unless otherwise stated.
About this article
Cite this article
Degeling, K., IJzerman, M.J., Koopman, M. et al. Accounting for parameter uncertainty in the definition of parametric distributions used to describe individual patient variation in health economic models. BMC Med Res Methodol 17, 170 (2017). https://doi.org/10.1186/s128740170437y
Received:
Accepted:
Published:
DOI: https://doi.org/10.1186/s128740170437y
Keywords
 Patientlevel variation
 Stochastic uncertainty
 Parameter uncertainty
 Statetransition modeling
 Discrete event simulation
 Personalized medicine
 Individual patient data