 Research
 Open Access
 Published:
Bayesian additional evidence for decision making under small sample uncertainty
BMC Medical Research Methodology volume 21, Article number: 221 (2021)
Abstract
Background
Statistical inference based on small datasets, commonly found in precision oncology, is subject to low power and high uncertainty. In these settings, drawing strong conclusions about future research utility is difficult when using standard inferential measures. It is therefore important to better quantify the uncertainty associated with both significant and nonsignificant results based on small sample sizes.
Methods
We developed a new method, Bayesian Additional Evidence (BAE), that determines (1) how much additional supportive evidence is needed for a nonsignificant result to reach Bayesian posterior credibility, or (2) how much additional opposing evidence is needed to render a significant result noncredible. Although based in Bayesian analysis, a prior distribution is not needed; instead, the tipping point output is compared to reasonable effect ranges to draw conclusions. We demonstrate our approach in a comparative effectiveness analysis comparing two treatments in a real world biomarkerdefined cohort, and provide guidelines for how to apply BAE in practice.
Results
Our initial comparative effectiveness analysis results in a hazard ratio of 0.31 with 95% confidence interval (0.09, 1.1). Applying BAE to this result yields a tipping point of 0.54; thus, an observed hazard ratio of 0.54 or smaller in a replication study would result in posterior credibility for the treatment association. Given that effect sizes in this range are not extreme, and that supportive evidence exists from a similar published study, we conclude that this problem is worthy of further research.
Conclusions
Our proposed method provides a useful framework for interpreting analytic results from small datasets. This can assist researchers in deciding how to interpret and continue their investigations based on an initial analysis that has high uncertainty. Although we illustrated its use in estimating parameters based on timetoevent outcomes, BAE easily applies to any normallydistributed estimator, such as those used for analyzing binary or continuous outcomes.
Background
In scientific research, statistical inference is crucial to drawing robust conclusions from data. This is often done through testing a parameter estimate for “statistical significance”, using pvalues and confidence intervals. These quantities are strongly dependent on sample size; in precision oncology, datasets for rare diseases or biomarkerdefined cohorts are often small. This leads to difficulties in deriving insight from analytic results by using standard statistical inference tools.
The primary issue with small sample sizes is that they lead to a lack of statistical power in analyses, meaning that the probability of declaring a true effect or association as statistically significant is small. Due to wellknown publication bias, and conflation of “absence of evidence” with “evidence of absence”, nonsignificant findings are often not reported or published at all [1]. As such, there is no opportunity to learn from the analysis conducted. Even if reported, such findings are usually qualified as “trending towards” or “approaching” significance, which is an arbitrary designation; it does not inform how likely the hypothesis of interest is, or whether future research is worthwhile.
Even a statistically significant result may have high uncertainty, and be unconvincing on its own if derived from a small dataset. This is particularly germane given recent attention on the “replicability crisis” in science [2]. In this scenario, it would be important to ensure the finding is not spurious. Standard analyses do not directly determine how likely the result would be to hold in future studies. Hence, there is a need for statistical tools that make it easier to derive utility from small sample datasets. In this area, Gelman and Carlin (2014) introduced the concept of Type S (sign) and M (magnitude) errors, which quantify the probability of an estimate being in the wrong direction and the expected factor by which its magnitude is exaggerated, conditional on being statistically significant [3]. Segal (2021) also derived confidence intervals for the probability that a replication study would yield estimates more extreme than a certain value (such as the statistical significance threshold) [4]. These methods improve upon standard inference towards the goal of replicable scientific results.
In this article, we introduce a new method, Bayesian Additional Evidence (BAE), for better quantifying the uncertainty of statistical inference output. Our goal is to aid researchers to better interpret results from analyses of small datasets and make decisions about the value of further pursuing research questions. BAE is based on the Bayesian analysis framework, but does not require explicitly setting a prior distribution, and is similar to the recently proposed Analysis of Credibility (AnCred) approach [5]. To implement the BAE approach, we illustrate how to “invert” Bayesian posterior computations to (1) assess the robustness of a significant result, and (2) determine the evidence gap given a nonsignificant result. Specifically, given an observed parameter estimate and standard error, BAE computes the range of parameter estimates that would need to be observed in a followup study in order to make a certain conclusion either in favor or against the hypothesis of interest.
The rest of this paper is structured as follows: in Section 2, we describe the Bayesian normalnormal model, on which our method is based. We then introduce BAE, and illustrate how to use it for decision making given significant or nonsignificant inferential results. In Section 3, we apply our method to a comparative effectiveness analysis of two treatments for a biomarkerdefined cohort from an oncology electronic health recordderived deidentified database, and report the results. We conclude with a brief discussion and conclusion in Section 4 and 5.
Methods
For the purposes of illustration, we will assume that the goal of the analysis is to estimate β_{true}, the log hazard ratio comparing two treatment arms, adjusting for relevant covariates. We further assume that this parameter is estimated through a standard Cox proportional hazards model, as is common practice in clinical research. However, the methods detailed are applicable to any estimand with a normallydistributed estimator. We also define a statistically significant result as one where the 95% confidence interval excludes the null value, though other significance levels may be used.
The methods we describe are based around Bayesian analysis, and the concept of incorporating prior information to improve precision in estimation. Specifically, we use the concept of “inverting” Bayes’ Theorem, or computing priors that would result in specific posterior distributions of interest.
Bayesian normalnormal model
A Bayesian analysis computes a posterior distribution for β_{true} (which is treated as a random variable), based on the distribution of the observed estimator and a prespecified prior distribution for the true parameter. For our methods, we consider a normally distributed estimator, and a normal prior distribution; this is known as the Bayesian normalnormal model.
We begin with the asymptotic normal distribution of the Cox model estimator \(\hat{\beta}\) (Andersen and Gill, 1982) [6], which is:
where n is the sample size, and σ is the standard deviation of the estimator.
Then, we can assume a normal prior for β_{true}: ~ N(µ, s )
By conjugacy of the normalnormal model, we then have a closed form for the posterior distribution:
where
Based on the posterior, we can calculate a 95% Bayesian credible interval as μ_{P} ± 1.96 s_{P}. The posterior mean can be interpreted as a weighted average of the prior and observed means, based on how much confidence we have in both quantities. Therefore, if n is low (implying more uncertainty in the observed data estimator) and s is also low (implying high confidence in the prior), then the posterior mean will be pulled towards μ. Conversely, as n increases, the posterior mean will tend towards \(\hat{\beta}\).
The posterior precision, which is defined as the reciprocal of the variance, is equal to the sum of the prior and observed precision. Therefore, the posterior variance will always be less than (or equal) to the observed variance, and so a Bayesian credible interval will necessarily be tighter than the corresponding frequentist confidence interval (or the same width).
Note that this model assumes that the standard deviation σ is known, which is often not true in practice. In the Cox model setting, σ is a function of β_{true}, which is also unknown. For the implementation of our method described below, we use the estimate \(\hat{\sigma}\left(\hat{\beta}\right)\) as a plugin, which should provide a reasonable estimate. A fully Bayesian analysis would define an additional prior distribution for σ.
Bayesian additional evidence (BAE)
The goal of our proposed method, leveraging the Bayesian framework, is to answer one of two potential questions:

1.
Given a nonsignificant frequentist inferential result, how much additional supportive evidence is needed to result in Bayesian posterior credibility?

2.
Given a significant frequentist inferential result, how much additional opposing evidence is needed to render the result noncredible in the posterior?
For illustration, we start by considering the first question. Suppose we have computed a normally distributed test statistic, with a 95% confidence interval that includes the null value. BAE then searches over prior distributions, which represent potential future data, to determine which posterior results provide credible evidence against the null hypothesis. Specifically, we fix the prior standard deviation and search over prior means. The direction of where to search depends on the hypothesis (i.e., is a parameter value greater than or less than the null value of substantive interest?) and is specified by the analyst.
The output of the BAE method is then the “tipping point” or least extreme prior mean μ_{∗} that results in a posterior credible interval that includes the null value. Therefore, all prior means that are more extreme than μ_{∗} will result in posterior credible intervals that exclude the null value, yielding 95% credibility. This is depicted visually in Fig. 1.
In other words, this method quantifies what type of additional result is needed to have sufficient evidence for an effect. If the returned tipping point is too extreme, then that indicates there is not much evidence for an effect. However, if the tipping point is within a plausible range (based on scientific domain knowledge), then we can declare that the analysis is worthy of followup. Although requiring some knowledge of plausible effect sizes, this method does not assume a fully known prior distribution.
In the normalnormal posterior, the prior and the observed quantities are symmetric. Therefore even though we are computing the “prior” mean which leads to a particular posterior, we can think of our initial result as the true prior study in time, and consider the BAE output as the range of effect sizes that would need to be observed in a future study in order to achieve a posterior credible interval that excludes the null. We are essentially encoding a replication followup study which results in sufficiently credible evidence as our prior.
The interpretation of the BAE output is also dependent on the prior standard deviation used. For example, if we use the same estimated standard error from our frequentist analysis, we can interpret the result as “in a future study with the same level of precision as our current analysis, this is the range of estimates that would yield a posterior credible interval that excludes the null”. We can also decrease or increase the prior standard deviation to encode future studies having higher or lower precision. For example, assuming that σ stays constant as n changes implies that dividing the observed standard error by \(\sqrt{X}\) corresponds to the standard error that would be observed in a study with X times more subjects. Due to the dependency of σ with β_{true} in the Cox model setting, this does not hold exactly, but can be used as an approximate heuristic.
With respect to the second question, given a significant initial result, this method can be inverted to see the range of prior means which would make the result noncredible. Here, the output of BAE is then the least extreme prior mean μ_{∗} that results in a posterior credible interval that includes the null value. Therefore, all prior means that are larger in magnitude than μ_{∗} will either result in posterior credible intervals that include the null value yielding noncredibility, or provide credible evidence for an effect in the opposite direction.
Interpretations of BAE applied in either scenario can be found in Table 1 below. We assume that the standard frequentist estimator has been computed, with confidence interval and pvalue.
Finally, although the BAE tipping point does not have a closedform solution, implementation is straightforward, and only requires using a rootfinding algorithm. Example R code is available in the supplementary material. BAE can also be extended to more complex estimators and prior distributions, so long as the posterior can be computed within the tipping point search algorithm.
Study design and data sources
A study by Innocenti et al. (2019) aimed to identify genomic factors associated with overall survival (OS) in metastatic colorectal cancer (mCRC) patients treated with either fluorouracil and leucovorin plus oxaliplatin (FOLFOX) or irinotecan (FOLFIRI) chemotherapy and either bevacizumab or cetuximab in the first line (1 L) setting [7]. The authors analyzed data from primary tumor DNA for 843 patients from a larger phase III trial. While the original trial found no statistically significant difference in OS between the treatment arms, the authors reported a very strong clinical benefit of bevacizumab compared to cetuximab (HR = 0.13, 95% CI [0.06, 0.30]) among 37 patients who were known to have microsatellite instabilityhigh (MSIH) tumors. Although statistically significant, this result is based on a small sample size, with 21 patients receiving bevacizumab and 16 patients receiving cetuximab. The authors also note the potential for selection bias due to MSIH status not being available for all patients in the original trial. Therefore, it is of interest to attempt a replication of this result with a different dataset in order to confirm this finding.
We compared OS for these two regimens for relevant patients with mCRC from the nationwide Flatiron Health EHRderived deidentified database. This longitudinal database is comprised of deidentified patientlevel structured and unstructured data, curated via technologyenabled abstraction [8, 9]. During the study period, the deidentified data originated from approximately 280 US cancer clinics (~ 800 sites of care). The majority of patients in the database originate from community oncology settings; relative community/academic proportions may vary depending on study cohort. Survival analysis was conducted using a composite mortality variable that aggregates EHRderived data (structured and unstructured) with links to the SSDI and obituary data [10].
Specifically, we selected a cohort of patients with mCRC diagnosed between 2013 and 2020 who had microsatellite instability high (MSIH) tumors, and were treated in the 1 L setting with FOLFOX or FOLFIRI chemotherapy plus either bevacizumab or cetuximab (Supplemental Fig. 1). Followup began on the start date of 1 L treatment, and ended at the earliest of either date of death or last confirmed structured EHR activity (e.g., noncancelled medication orders, medication administrations, or clinic visits with vital signs measured). We excluded patients who had a gap of more than 90 days between their mCRC diagnosis date and their first confirmed structured EHR activity date in the Flatiron Health network. We also excluded patients whose date of death was prior to their recorded 1 L start or MSI test result date; such inconsistencies can occur with real world EHRderived data [10].
The conditional association between OS and 1 L treatment was assessed by fitting a Cox proportional hazards model, adjusting for age, sex, race (dichotomized to White or nonWhite), BRAF mutation status (present or absent before start of 1 L therapy), and KRAS or NRAS mutation status (present or absent before start of 1 L therapy). Risk set adjustment was applied in order to account for the delayed entry of patients who received their MSI test after the start of 1 L therapy. This analysis is as similar as possible to that conducted by Innocenti et al., (2019) though we were not able to adjust for tumor location, number of metastatic sites, and synchronous or metachronous metastases since those data were not part of the core Flatiron Health data model.
Institutional Review Board approval of the study protocol was obtained prior to study conduct, and included a waiver of informed consent.
Results
After applying the selection criteria for our study, there were 118 patients in the bevacizumab cohort and 7 patients in the cetuximab cohort. Table 2 provides a description of baseline patient characteristics. Note that due to the small size of the cetuximab arm, we expect there to be high uncertainty in the estimation of the hazard ratio comparing treatments, making this analysis relevant to our method.
Fitting a Cox model, we estimate that the adjusted hazard ratio of death for patients treated with 1 L chemotherapy plus bevacizumab compared to patients only treated with 1 L chemotherapy plus cetuximab is 0.42 (95% CI: 0.14, 1.23), with a pvalue of 0.11. Therefore, we cannot conclude that this association is statistically significant at the 5% significance level.
Although we observed a nonsignificant result, due to the small sample size in the chemotherapy plus cetuximab arm, it is important to quantify the uncertainty in the analysis beyond standard methods. We computed the Bayesian Additional Evidence from the frequentist regression analysis. Using the estimated standard error from this analysis as the prior standard deviation, the BAE tipping point is 0.52 on the hazard ratio scale. Therefore, given a replication study with the same level of precision, an observed hazard ratio of 0.52 or smaller would result in posterior credibility for the association of interest. This is depicted graphically in Fig. 2. Scientifically, such hazard ratios would not be considered extreme enough to be implausible. Recall that the similar analysis by Innocenti et al. (2019) reported a hazard ratio of 0.13 with 95% confidence interval (0.06, 0.30) [7]. Therefore, we can conclude that it is worth gathering more evidence for this research question.
The BAE output also tells us that using the published result of Innocenti et al. (2019) as a prior combined with our observed result in a Bayesian analysis would result in 95% posterior credibility. This is because the standard error associated with the prior result is less than that estimated by our analysis. Therefore, given the BAE tipping point of 0.45, a prior mean of 0.13 results in a Bayesian analysis that yields posterior credibility.
Discussion
It is difficult to extract useful statistical inference from small sample datasets. In this paper, we developed a new approach, for interpreting the evidence provided by these datasets. This is motivated by settings involving rare diseases or biomarkers, where conducting a wellpowered study is difficult. Although our method uses the Bayesian framework, it does not require an explicit prior distribution; only domain knowledge of reasonable effect sizes is needed. BAE thus allows for easy integration of prior knowledge with these analyses in order to inform the value of research questions. BAE helps to interpret highuncertainty results, which can then result in easier decisionmaking for researchers on how to conduct future studies. We illustrated this in a real world data example involving a small cohort. Here, a standard frequentist analysis yields a nonsignificant result, but additional uncertainty quantification using BAE shows that credible evidence is plausible with more data. This evidence gap is shown to be tractable given results from a previously published similar analysis.
Within the literature related to statistical analyses using small samples, a change in analytic method is often proposed. Examples include the use of Fisher’s exact test for contingency tables [11], the Firth bias correction for certain generalized linear models [12], and the generalized log rank test statistic for survival analysis [13]. These methods aim to correct inference when asymptotic results may not hold; moreover, BAE can be applied as a complement to analytic results from these methods. Despite this, these methods alone do not assist in making a decision based on an inferential result with high uncertainty, which would likely still occur. As previously discussed, Bayesian analyses present a potential solution. Given sufficiently strong domain knowledge that can be encoded as a prior distribution, we can reduce analytic uncertainty. Bayesian methods can also accommodate a wide variety of datagenerating distributions and analyses. However, selecting an appropriate prior is inherently subjective, and may be difficult in many situations due to a lack of published evidence.
Other work also attempts to solve this problem by improving the understanding of analytic results beyond a dichotomous significance threshold. For example, Blume et al, 2019 propose adapting pvalues to be based on interval (instead of point) null hypotheses, to estimate the fraction of datasupported hypotheses that are “scientifically null”, without requiring a significance threshold [14]. Similarly, Gannon et al, 2019 define a new type of hypothesis test that minimizes a linear combination of the false positive and false negative rates [15]. Within the classic hypothesis testing framework, Segal, 2021 provides confidence intervals for replication probabilities [16]. These approaches and others (see Wasserstein et al, 2019 for an overview) [17] may be useful in certain scenarios depending on the research goals and prior knowledge available.
Toward the goal of improving decisionmaking, BAE is closely related to the analysis of credibility (AnCred) approach developed by Matthews et al. (2018) [5], which takes the same general approach of performing an inverse Bayesian analysis to find a prior that yields a specific posterior. As with our method, a statistic based on this prior is then compared to plausible effect sizes to arrive at a decision. Although we find this inverse Bayesian approach to be appealing and a useful way to contextualize inferential results, AnCred is more difficult to interpret than BAE, since it provides intervals of prior effect sizes that are consistent with (non)credible evidence of effects. In the case of nonsignificant initial results, these intervals can be wide enough to effectively contain any effect size, which is unhelpful for decision making. On the other hand, BAE outputs a clear evidence threshold that would need to be observed in a followup study to make certain conclusions.
A limitation of our method is in using the plugin estimate \(\hat{\sigma}\left(\hat{\beta}\right)\) of the coefficient standard deviation in the future study. A fully Bayesian analysis would also model the relationship between σ and β_{true}. However, this would involve selection of appropriate priors and additional computational complexity, while our approximation is very straightforward and fast to use in practice. An R implementation of BAE is available in the online supplement.
The use of BAE is also similar to interim analyses of clinical trials, where posterior distributions are computed (based on a prespecified prior) in order to determine whether the trial should be stopped early due to clear efficacy or futility. It is important to note, however, that BAE should not be seen as a replacement for confirmatory hypothesis testing e.g. in regulatory settings. Rather, it should be used to inform the utility and design of future confirmatory studies. Although we illustrate our approach with a survival analysis of EHRderived data, it can also be applied to other analyses of datasets having high uncertainty, when using normallydistributed estimators.
Conclusions
We have developed a novel method to determine the amount of additional evidence needed for a nonsignificant result to reach Bayesian posterior credibility, or for a significant result to reach noncredibility. We believe BAE can be used to draw initial conclusions from small datasets, which can then be validated with followup confirmatory studies. This approach could mitigate underreporting of analytic results that are not statistically significant, but are clinically useful.
Availability of data and materials
The data that support the findings of this study were originated by Flatiron Health, Inc. In order to comply with legal requirements, and to preserve the deidentification status, these deidentified data may be made available upon request, and are subject to a license agreement with Flatiron Health; interested researchers should contact <DataAccess@flatiron.com> to determine licensing terms.
Abbreviations
 1 L:

First line
 BAE:

Bayesian Additional Evidence
 EHR:

Electronic health record
 MSIH:

Microsatellite instability high
 OS:

Overall survival
 mCRC:

Metastatic colorectal cancer
References
Begg CB, Berlin JA. Publication bias: A problem in interpreting medical data. Journal of the Royal Statistical Society. Series A (Statistics in Society). 1988;151(3):419–463. http://www.jstor.org/stable/2982993. doi: https://doi.org/10.2307/2982993.
Wasserstein RL, Schirm AL, Lazar NA. Moving to a world beyond “p < 0.05”. Am Stat. 2019;73:1–19. https://doi.org/10.1080/00031305.2019.1583913. https://doi.org/10.1080/00031305.2019.1583913.
Gelman A, Carlin J. Beyond power calculations: assessing type S (sign) and type M (magnitude) errors. Perspect Psychol Sci. 2014;9(6):641–51. https://doi.org/10.1177/1745691614551642.
Segal BD. Toward Replicability with confidence intervals for the exceedance probability. Am Stat. 2021;75(2):128–38. https://doi.org/10.1080/00031305.2019.1678521.
Matthews RAJ. Moving towards the post p < 0.05 era via the analysis of credibility. Am Stat. 2019;73:202–12. https://doi.org/10.1080/00031305.2018.1543136.
Andersen PK, Gill RD. Cox's regression model for counting processes: a large sample study. Ann Stat. 1982;10(4):1100–20 http://www.jstor.org/stable/2240714.
Innocenti F, Ou F, Qu X, et al. Mutational analysis of patients with colorectal cancer in CALGB/SWOG 80405 identifies new roles of microsatellite instability and tumor mutational burden for patient outcome. JCO. 2019;37(14):1217–27. https://doi.org/10.1200/JCO.18.01798.
Ma X, Long L, Moon S, Adamson BJS, Baxi SS. Comparison of population characteristics in realworld clinical oncology databases in the US: Flatiron Health, SEER, and NPCR. medRxiv. 2020. https://doi.org/10.1101/2020.03.16.20037143.
Birnbaum B, Nussbaum N, SeidlRathkopf K, et al. Modelassisted cohort selection with bias analysis for generating largescale cohorts from the EHR for oncology research. ArXiv. 2020. https://arxiv.org/abs/2001.09765.
Curtis MD, Griffith SD, Tucker M, et al. Development and validation of a highquality composite realworld mortality endpoint. Health Serv Res. 2018;53(6):4460–76. https://doi.org/10.1111/14756773.12872.
Upton, G. “Fisher's Exact Test.” Journal of the Royal Statistical Society. Series A (Statistics in Society), vol. 155, no. 3, 1992, pp. 395–402. JSTOR, www.jstor.org/stable/2982890. Accessed 3 Aug 2021.
Firth D. Bias reduction of maximum likelihood estimates. Biometrika. March 1993;80(1):27–38. https://doi.org/10.1093/biomet/80.1.27.
Mehrotra DV, Roth AJ. Relative risk estimation and inference using a generalized logrank statistic. Stat Med. 2001;20(14):2099–113. https://doi.org/10.1002/sim.854.
Blume JD, Greevy RA, Welty VF, Smith JR, Dupont WD. An introduction to secondgeneration pvalues. Am Stat. 2019;73:157–67. https://doi.org/10.1080/00031305.2018.1537893. https://doi.org/10.1080/00031305.2018.1537893.
Gannon MA. de Bragança Pereira, Carlos Alberto, Polpo a. blending bayesian and classical tools to define optimal samplesizedependent significance levels. Am Stat. 2019;73:213–22. https://doi.org/10.1080/00031305.2018.1518268.
Segal, BD. Toward Replicability With Confidence Intervals for the Exceedance Probability, The American Statistician, 75:2, 128–138, DOI: https://doi.org/10.1080/00031305.2019.1678521
Wasserstein, RL, Schirm, AL, Lazar, NA. Moving to a World Beyond “p<0.05”, The American Statistician, 73:sup1, 1–19, DOI: https://doi.org/10.1080/00031305.2019.1583913
Acknowledgements
Hannah Gilham for writing and editorial support.
Authors’ information (optional)
Not applicable.
Funding
This study was sponsored by Flatiron Health, Inc., which is an independent subsidiary of the Roche Group.
Author information
Affiliations
Contributions
Study concept and design: AS. Data collection, analysis and interpretation: All. Manuscript writing and review: All. The author(s) read and approved the final manuscript.
Corresponding author
Ethics declarations
Ethics approval and consent to participate
Institutional Review Board approval of the study protocol was obtained prior to study conduct from the WCG IRB, and included a waiver of informed consent, in accordance with relevant ethical guidelines (Declaration of Helsinki).
Consent for publication
Not applicable.
Competing interests
At the time of the study, AS, BS, JS, OH, MM, report employment in Flatiron Health, Inc., an independent subsidiary of Roche. AS, BS, JS, OH, MM report stock ownership in Roche. BS, JS report equity ownership in Flatiron Health, Inc.
Additional information
Publisher’s Note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Supplementary Information
Additional file 2.
(R 2 kb)
Rights and permissions
Open Access This article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons licence, and indicate if changes were made. The images or other third party material in this article are included in the article's Creative Commons licence, unless indicated otherwise in a credit line to the material. If material is not included in the article's Creative Commons licence and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this licence, visit http://creativecommons.org/licenses/by/4.0/. The Creative Commons Public Domain Dedication waiver (http://creativecommons.org/publicdomain/zero/1.0/) applies to the data made available in this article, unless otherwise stated in a credit line to the data.
About this article
Cite this article
Sondhi, A., Segal, B., Snider, J. et al. Bayesian additional evidence for decision making under small sample uncertainty. BMC Med Res Methodol 21, 221 (2021). https://doi.org/10.1186/s12874021014325
Received:
Accepted:
Published:
DOI: https://doi.org/10.1186/s12874021014325
Keywords
 Real world
 Real world evidence
 Small sample size
 Bayesian