Adjustment for reporting bias in network meta-analysis of antidepressant trials
© Trinquart et al.; licensee BioMed Central Ltd. 2012
Received: 12 June 2012
Accepted: 19 September 2012
Published: 27 September 2012
Network meta-analysis (NMA), a generalization of conventional MA, allows for assessing the relative effectiveness of multiple interventions. Reporting bias is a major threat to the validity of MA and NMA. Numerous methods are available to assess the robustness of MA results to reporting bias. We aimed to extend such methods to NMA.
We introduced 2 adjustment models for Bayesian NMA. First, we extended a meta-regression model that allows the effect size to depend on its standard error. Second, we used a selection model that estimates the propensity of trial results being published and in which trials with lower propensity are weighted up in the NMA model. Both models rely on the assumption that biases are exchangeable across the network. We applied the models to 2 networks of placebo-controlled trials of 12 antidepressants, with 74 trials in the US Food and Drug Administration (FDA) database but only 51 with published results. NMA and adjustment models were used to estimate the effects of the 12 drugs relative to placebo, the 66 effect sizes for all possible pair-wise comparisons between drugs, probabilities of being the best drug and ranking of drugs. We compared the results from the 2 adjustment models applied to published data and NMAs of published data and NMAs of FDA data, considered as representing the totality of the data.
Both adjustment models showed reduced estimated effects for the 12 drugs relative to the placebo as compared with NMA of published data. Pair-wise effect sizes between drugs, probabilities of being the best drug and ranking of drugs were modified. Estimated drug effects relative to the placebo from both adjustment models were corrected (i.e., similar to those from NMA of FDA data) for some drugs but not others, which resulted in differences in pair-wise effect sizes between drugs and ranking.
In this case study, adjustment models showed that NMA of published data was not robust to reporting bias and provided estimates closer to that of NMA of FDA data, although not optimal. The validity of such methods depends on the number of trials in the network and the assumption that conventional MAs in the network share a common mean bias mechanism.
KeywordsNetwork meta-analysis Publication bias Small-study effect
Network meta-analyses (NMAs) are increasingly being used to evaluate the best intervention among different existing interventions for a specific condition. The essence of the approach is that intervention A is compared with a comparator C, then intervention B with C, and adjusted indirect comparison allows for comparing A and B, despite the lack of any head-to-head randomized trial comparing A and B. An NMA, or multiple-treatments meta-analysis (MA), allows for synthesizing comparative evidence for multiple interventions by combining direct and indirect comparisons [1–3]. The purpose is to estimate effect sizes for all possible pair-wise comparisons of interventions, although some comparisons have no available trial.
Reporting bias is a major threat to the validity of results of conventional systematic reviews or MAs [4, 5]. Accounting for reporting biases in NMA is challenging, because unequal availability of findings across the network of evidence may jeopardize NMA validity [6, 7]. We previously empirically assessed the impact of reporting bias on the results of NMAs of antidepressant trials and showed that it may bias estimates of treatment efficacy .
Numerous methods have been used as sensitivity analyses to assess the robustness of conventional MAs to publication bias and related small-study effects [9–20]. Modeling methods include regression-based approaches and selection models. We extend these approaches to NMAs in the Bayesian framework.
First, we extended a meta-regression model of the effect size on its standard error, recently described for MAs [21, 22]. In this approach, the regression slope reflects the magnitude of the association of effect size and precision (ie, the “small-study effect”), and the intercept provides an adjusted pooled effect size (ie, the predicted effect size of a trial with infinite precision). Second, we introduced a selection model, which models the probability of a trial being selected and is taken into account with inverse weighting in the NMA. Both adjustment models rely on the assumption that biases are exchangeable across the network, ie, biases, if present, operate in a similar way in trials across the network. Third, we applied these adjustment models to datasets created from US Food and Drug Administration (FDA) reviews of antidepressant trials and from their matching publications. These datasets were shown to differ because of reporting bias . We compared the results of the adjustment models applied to published data and standard NMA for published and for FDA data, the latter considered the reference standard.
The standard model for NMA was formalized by Lu and Ades [2, 24, 25]. We assume that each trial i assessed treatments j and k among the T interventions in the network. Each trial provided an estimated intervention effect size y ijk of j over k and its variance v ijk . We assume that y ijk > 0 indicates superiority of j over k. Assuming normal likelihood and according to a random-effects model, y ijk ~ N(θ ijk , v ijk ) and , where θ ijk is the true effect underlying each randomized comparison between treatments j and k and is the mean of the random-effects effect sizes over randomized comparisons between treatments j and k. The model assumes homogeneous variance (ie, τ jk 2 = τ 2). This assumption can be relaxed [2, 26]. The model also assumes consistency between direct and indirect evidence: if we consider treatment b as the overall network baseline treatment, the treatment effects of j, k, etc. relative to treatment b, , , etc., are considered basic parameters, and the remaining contrasts, the functional parameters, are derived from the consistency equations for every j, k ≠ b.
for every j, k ≠ b
Figure A in Additional file 2 shows a graphical representation of the model. In the regression equation, θ ijk is the treatment effect adjusted for small-study effects underlying each randomized comparison between treatments j and k; β jk represents the potential small-study effect (ie, the slope associated with funnel plot asymmetry for the randomized comparisons between treatments j and k). The model assumes that these comparison-specific regression slopes follow a common normal distribution, with mean slope β and common between-slopes variance σ 2. This is equivalent to the assumption that comparison-specific small-study biases are exchangeable within the network. Since we assumed that y ijk > 0 indicates superiority of j over k, β > 0 would mean an overall tendency for a small-study effect (ie, treatment contrasts tend to be over-estimated in smaller trials). Finally, is equal to 1 if a small-study effect is expected to favor treatment j over k, equal to −1 if a small-study effect is expected to favor treatment k over j, and equal to 0 when one has no reason to believe that there is bias in either direction (e.g., for equally novel active vs. active treatment). In trials comparing active and inactive treatments (e.g., placebo, no intervention), we can reasonably expect the active treatment to be always favored by small-study bias.
for every j, k ≠ b
Figure B in Additional file 2 shows a graphical representation of the model. In the logistic regression equation, w i represents the propensity of the trial results to be published, β 0jk sets the overall probability of observing a randomized comparison between treatments j and k, and β 1jk controls how fast this probability evolves as the standard error increases. We expect β 1jk to be negative, so trial results yielding larger standard errors have lower propensity to be published. The model assumes exchangeability of the β 0jk and β 1jk coefficients within the network. By setting γ ijk = θ ijk /w i , we define a simple scheme that weights up trial results with lower propensity of being published so that they have a disproportionate influence in the NMA model. θ ijk is the treatment contrast corrected for the selection process underlying each randomized comparison between treatments j and k. Finally, is defined in the same way as in the preceding section.
We estimated 4 models: standard NMA model of published data, 2 adjustment models of published data and a standard NMA model of FDA data. In each case, model estimation involved Markov chain Monte Carlo methods with Gibbs sampling. Placebo was chosen as the overall baseline treatment to compare all other treatments. Consequently, the 12 effects of drugs relative to placebo are the basic parameters. For 2 treatments j and k, SMD jk > 0 indicate that j is superior to k. In both the meta-regression and selection models, we assumed that the active treatments would always be favored by small-study bias as compared to placebo; consequently, is always equal to 1.
In the standard NMA model, we defined prior distributions for the basic parameters and the common variance τ 2: and . In the meta-regression model, we further chose vague priors for the mean slope β and common between-slopes variance σ 2: and . In the selection model, we chose weakly informative priors for the central location and dispersion parameters (β 0, σ 0 2) and (β 1, σ 1 2). We considered p min and p max the probability of publication when the standard error takes its minimum and maximum values across the network of published data and specified beta priors for these probabilities . The latter was achieved indirectly by specifying prior guesses for the median and 5th or 95th percentile . For trials with standard error equal to the minimum observed value, we assumed that the chances of p min being < 50% were 5% and the chances of p min being < 80% were 50%. For trials with standard error equal to the maximum observed value, our guess was that the chances of p max being < 40% were 50% and the chances of p max being < 70% were 95%. We discuss these choices further in the Discussion. From this information, we determined Beta(7.52, 2.63) and Beta(3.56, 4.84) as prior distributions for p min and p max , respectively. Finally, we expressed β 0 and β 1 in terms of p min and p max and chose uniform distributions in the range (0,2) on the standard deviations σ 0 and σ 1. For each analysis, we constructed posterior distributions from 2 chains of 500,000 simulations, after convergence achieved from an initial 500,000 simulations for each (burn-in). Analysis involved use of WinBUGS v1.4.3 (Imperial College and MRC, London, UK) to estimate all Bayesian models and R v2.12.2 (R Development Core Team, Vienna, Austria) to summarize inferences and convergence. Codes are reported in the Additional file 1: Appendix 2.
We compared the results of the 2 adjustment models applied to published data and results of the standard NMA model applied to published data and the FDA data, the latter considered the reference standard. First, we compared posterior means and 95% credibility intervals for the 12 basic parameters and common variance, as well as for the 66 functional parameters (ie, all 12 × 11/2 = 66 possible pair-wise comparisons of the 12 drugs). Second, we compared the rankings of the competing treatments. We assessed the probability that each treatment was best, then second best and third best, etc. We plotted the cumulative probabilities and computed the surface under the cumulative ranking (SUCRA) line for each treatment . Third, to compare the different models applied to published data, we used the posterior mean of the residual deviance and the deviance information criteria .
In the meta-regression model applied to published data, the posterior mean slope β was 1.7 (95% credible interval −0.3–3.6), which suggests an overall tendency for a small-study effect in the network. The 12 regression slopes were similar, with posterior means ranging from 1.4 to 1.9. In the selection model applied to published data, the mean slope β 1 was −10.0 (−18.0 – -2.50), so trials yielding larger standard errors tended overall to have lower propensity to be published. In both models, all estimates were subject to large uncertainty (Additional file 1: Appendix 3).
Comparison of network meta-analysis (NMA)-based estimates between the 2 adjustment models applied to published data and the standard NMA model applied to US Food and Drug Administration (FDA) data and to published data
Standard NMA model
Standard NMA model
Θ PAR CR
Θ VEN XR
Comparison of fit and complexity between the 2 adjustment models and the standard NMA model, all applied to published data
Mean posterior residual deviance ()
Effective number of parameters (pD)
Deviance Information Criterion (DIC)
We extended two adjustment methods for reporting bias from MAs to NMAs. The first method combined NMA and meta-regression models, with effect sizes regressed against their precision. The second one combined the NMA model with a logistic selection model estimating the probability that a trial was published or selected in the network. The former method basically adjusts for funnel plot asymmetry or small study effects, which may arise from causes other than publication bias. The latter adjusts for publication bias (ie, the suppression of an entire trial depending on results). The two models borrow strength from other trials in the network with the assumption that biases operate in a similar way in trials across the domain.
In a specific network of placebo-controlled trials of antidepressants, based on data already described and published previously by Turner et al., comparing the results of adjustment models applied to published data and those of the standard NMA model applied to published data allowed for assessing the robustness of efficacy estimates and ranking to publication bias or related small-study effects. Both models showed a decrease in all basic parameters (ie, the 12 effect sizes of drugs relative to placebo). The 66 contrasts for all possible pair-wise comparisons between drugs, the probabilities of being the best drug and the ranking were modified as well. The NMA of published data was not robust to publication bias and related small-study effects.
This specific dataset offered the opportunity to perform NMAs on both published and FDA data. The latter may be considered "an unbiased (but not the complete) body of evidence" for placebo-controlled trials of antidepressants . The comparison of the results of the 2 models applied to published data and the standard NMA model applied to FDA data showed that the effect sizes of drugs relative to placebo were corrected for some but not all drugs. This observation led to differences in the 66 possible pair-wise comparisons between drugs, the probabilities of being the best drug and the ranking. It suggests that the 2 models should not be considered optimal; that is, the objective is not to produce definitive estimates adjusted for publication bias and related small-study effects but rather to assess the robustness of results to the assumption of bias.
Similar approaches have been used by other authors. Network meta-regression models fitted within a Bayesian framework were previously developed to assess the impact of novelty bias and risk of bias within trials [36, 37]. Network meta-regression to assess the impact of small-study effect was specifically used by Dias et al. in a re-analysis of a network of published head-to-head randomized trials of selective serotonin reuptake inhibitors . Along the line of the regression-based approach of Moreno et al. in conventional MA, the authors introduced a measure of study size as a regression variable within the NMA model and identified a mean bias in pair-wise effect sizes. More recently, Moreno et al. used a similar approach to adjust for small-study effects in several conventional MAs of similar interventions and outcomes and illustrated their method using the dataset of Turner et al. . Our approach differed in that we extended this meta-regression approach to NMAs. We used the standard error of treatment effect estimate as the regressor. As well, we specified an additive between-trial variance rather than a multiplicative overdispersion parameter. With the latter, the estimated multiplicative parameter may be < 1, which implies less heterogeneity than would be expected by chance alone. Selection model approaches have been considered recently. Chootrakool et al. introduced an approximated normal model based on empirical log-odds ratio for NMAs within a frequentist framework and applied Copas selection models for some groups of trials in the network selected according to funnel plot asymmetry . Mavridis et al. presented a Bayesian implementation of the Copas selection model extended to NMA and applied their method on the network of Turner et al. . In the Copas selection model, the selection probability depends on both the estimates of the treatment effects and their standard errors. In the extension to NMA, an extra correlation parameter ρ, assumed equal for all comparisons, needs to be estimated. When applied to published data of the network of Turner et al., the selection model we proposed and the treatment-specific selection model of Mavridis et al. yielded close results.
The 2 adjustment models rely on the assumption of exchangeability of selection processes across the network; that is, biases, if present, operate in a similar way in trials across the network. In this case study, all studies were, by construction, industry-sponsored, placebo-controlled trials registered with the FDA, and for all drugs, results of entire studies remained unreported depending on the results . Thus, the assumption of exchangeability of selection processes is plausible. More generally, if we have no information to distinguish different reporting bias mechanisms across the network, an exchangeable prior distribution is plausible, "ignorance implies exchangeability" [42, 43]. However, the assumption may not be tenable in other contexts in which reporting biases may affect the network in an unbalanced way. It may operate differently in placebo-controlled and head-to-head trials , in older and more recent trials (because of trial registries), and for drug and non-drug interventions . In more complex networks involving head-to-head trials, the 2 adjustment models could be generalized to allow the expected publication bias or small-study bias for active-active trials to differ from that of the expected bias in trials comparing active and inactive treatments . In head-to-head trials, the direction of bias is uncertain but assumptions in defining could be that the sponsored treatment is favored (sponsorship bias) [45, 46] or that the newest treatment is favored (optimism bias) [37, 47, 48]. If treatment j is the drug provided by the pharmaceutical that sponsored the trial and treatment k is not, would be equal to 1. Or would be equal to 1 if treatment j is newer than treatment k. However, disentangling the sources of bias operating on direct and indirect evidence would be difficult, especially if reporting bias and inconsistency are twisted together or if the assumed bias directions are in conflict on a loop.
The models we described have limitations. First, they would result in poor estimation of bias and effect sizes when the conventional MAs within the network include small numbers of trials . Second, for the selection model, we specified the weight function. If the underlying assumptions (ie, a logistic link form and the chance of a trial being selected related to standard error) are wrong, the estimated selection model will be wrong. However, alternative weight functions (e.g., probit link) or conditioning (e.g., on the magnitude of effect size) could be considered. Finally, it was implemented with a weakly informative prior, which mainly suggested that the propensity for results to be published may decrease with increasing standard error. There is a risk that prior information overwhelms observed data, especially if the number of trials is low. Although they were somewhat arbitrarily set, our priors for the selection model parameters were in line with the values in previous studies using the Copas selection model [12, 49]. Different patterns of selection bias could be tested, for instance, by considering various prior modes for p min and p max , the probabilities of publication when the standard error takes its minimum and maximum values across the network .
In conclusion, addressing publication bias and related small-study effects in NMAs was feasible in this case study. Validity may be conditioned by sufficient numbers of trials in the network and assuming that conventional MAs constituting the network share a common mean bias. Simulation analyses are required to determine under which condition such adjustment models are valid. Application of such adjustment models should be replicated on more complex networks, ideally representing the totality of the data as in Turner's, but our results confirm that authors and readers should interpret NMAs with caution when reporting bias has not been addressed.
Grant support was from the French Ministry of Health Programme Hospitalier de Recherche Clinique National (PHRC 2011 MIN01-63) and European Union Seventh Framework Programme (FP7 – HEALTH.2011.4.1-2) under grant agreement n° 285453 (http://www.open-project.eu). The funders had no role in study design, data collection and analysis, decision to publish, or preparation of the manuscript.
The authors thank Laura Smales (BioMedEditing, Toronto, Canada) for editing the manuscript.
- Lumley T: Network meta-analysis for indirect treatment comparisons. Stat Med. 2002, 21 (16): 2313-2324. 10.1002/sim.1201.View ArticlePubMedGoogle Scholar
- Lu G, Ades AE: Combination of direct and indirect evidence in mixed treatment comparisons. Stat Med. 2004, 23 (20): 3105-3124. 10.1002/sim.1875.View ArticlePubMedGoogle Scholar
- Salanti G, Higgins JPT, Ades AE, Ioannidis JPA: Evaluation of networks of randomized trials. Stat Methods Med Res. 2008, 17 (3): 279-301.View ArticlePubMedGoogle Scholar
- Dwan K, Altman DG, Arnaiz JA, Bloom J, Chan AW, Cronin E, Decullier E, Easterbrook PJ, Von Elm E, Gamble C, et al: Systematic review of the empirical evidence of study publication bias and outcome reporting bias. PLoS One. 2008, 3 (8): e3081-10.1371/journal.pone.0003081.View ArticlePubMedPubMed CentralGoogle Scholar
- Song F, Parekh S, Hooper L, Loke YK, Ryder J, Sutton AJ, Hing C, Kwok CS, Pang C, Harvey I: Dissemination and publication of research findings: an updated review of related biases. Health Technol Assess. 2010, 14 (8): 193-View ArticleGoogle Scholar
- Salanti G, Kavvoura FK, Ioannidis JP: Exploring the geometry of treatment networks. Ann Intern Med. 2008, 148 (7): 544-553.View ArticlePubMedGoogle Scholar
- Li T, Puhan MA, Vedula SS, Singh S, Dickersin K: Network meta-analysis-highly attractive but more methodological research is needed. BMC Med. 2011, 9: 79-10.1186/1741-7015-9-79.View ArticlePubMedPubMed CentralGoogle Scholar
- Trinquart L, Abbé A, Ravaud P: Impact of reporting bias in network meta-analysis of antidepressant placebo-controlled trials. PLoS One. 2012, 7 (4): e35219-10.1371/journal.pone.0035219.View ArticlePubMedPubMed CentralGoogle Scholar
- Hedges LV: Modeling publication selection effects in meta-analysis. Stat Sci. 1992, 7 (2): 246-255. 10.1214/ss/1177011364.View ArticleGoogle Scholar
- Silliman NP: Hierarchical selection models with applications in meta-analysis. JASA. 1997, 92 (439): 926-936.View ArticleGoogle Scholar
- Larose DT, Dey DK: Modeling publication bias using weighted distributions in a Bayesian framework. Comput Stat Data Anal. 1998, 26: 279-302. 10.1016/S0167-9473(97)00039-X.View ArticleGoogle Scholar
- Copas J, Shi JQ: Meta-analysis, funnel plots and sensitivity analysis. Biostatistics. 2000, 1 (3): 247-262. 10.1093/biostatistics/1.3.247.View ArticlePubMedGoogle Scholar
- Duval S, Tweedie R: Trim and fill: a simple funnel-plot-based method of testing and adjusting for publication bias in meta-analysis. Biometrics. 2000, 56 (2): 455-463. 10.1111/j.0006-341X.2000.00455.x.View ArticlePubMedGoogle Scholar
- Sutton AJ, Song F, Gilbody SM, Abrams KR: Modelling publication bias in meta-analysis: a review. Stat Methods Med Res. 2000, 9 (5): 421-445. 10.1191/096228000701555244.View ArticlePubMedGoogle Scholar
- Copas JB, Shi JQ: A sensitivity analysis for publication bias in systematic reviews. Stat Methods Med Res. 2001, 10 (4): 251-265. 10.1191/096228001678227776.View ArticlePubMedGoogle Scholar
- Preston C, Ashby D, Smyth R: Adjusting for publication bias: modelling the selection process. J Eval Clin Pract. 2004, 10 (2): 313-322. 10.1111/j.1365-2753.2003.00457.x.View ArticlePubMedGoogle Scholar
- Bowden J, Jackson D, Thompson SG: Modelling multiple sources of dissemination bias in meta-analysis. Stat Med. 2010, 29 (7–8): 945-955.View ArticlePubMedGoogle Scholar
- Carpenter J, Rucker G, Schwarzer G: Assessing the sensitivity of meta-analysis to selection bias: a multiple imputation approach. Biometrics. 2011, 67 (3): 1066-1072. 10.1111/j.1541-0420.2010.01498.x.View ArticlePubMedGoogle Scholar
- Rucker G, Carpenter JR, Schwarzer G: Detecting and adjusting for small-study effects in meta-analysis. Biom J. 2011, 53 (2): 351-368. 10.1002/bimj.201000151.View ArticlePubMedGoogle Scholar
- Rufibach K: Selection models with monotone weight functions in meta analysis. Biom J. 2011, 53 (4): 689-704. 10.1002/bimj.201000240.View ArticlePubMedGoogle Scholar
- Moreno SG, Sutton AJ, Ades AE, Stanley TD, Abrams KR, Peters JL, Cooper NJ: Assessment of regression-based methods to adjust for publication bias through a comprehensive simulation study. BMC Med Res Methodol. 2009, 9: 2-10.1186/1471-2288-9-2.View ArticlePubMedPubMed CentralGoogle Scholar
- Moreno SG, Sutton AJ, Thompson JR, Ades AE: Abrams KR. 2012, Cooper NJ: A generalized weighting regression-derived meta-analysis estimator robust to small-study effects and heterogeneity. Stat MedGoogle Scholar
- Turner EH, Matthews AM, Linardatos E, Tell RA, Rosenthal R: Selective publication of antidepressant trials and its influence on apparent efficacy. N Engl J Med. 2008, 358 (3): 252-260. 10.1056/NEJMsa065779.View ArticlePubMedGoogle Scholar
- Higgins JP, Whitehead A: Borrowing strength from external trials in a meta-analysis. Stat Med. 1996, 15 (24): 2733-2749. 10.1002/(SICI)1097-0258(19961230)15:24<2733::AID-SIM562>3.0.CO;2-0.View ArticlePubMedGoogle Scholar
- Lu G, Ades AE: Assessing evidence inconsistency in mixed treatment comparisons. JASA. 2006, 101 (474): 447-459.View ArticleGoogle Scholar
- Lu G, Ades A: Modeling between-trial variance structure in mixed treatment comparisons. Biostatistics. 2009, 10 (4): 792-805. 10.1093/biostatistics/kxp032.View ArticlePubMedGoogle Scholar
- Stanley TD: Meta-regression methods for detecting and estimating empirical effects in the presence of publication selection*. Oxf Bull Econ Stat. 2008, 70 (1): 103-127.Google Scholar
- Moreno SG, Sutton AJ, Turner EH, Abrams KR, Cooper NJ, Palmer TM, Ades AE: Novel methods to deal with publication biases: secondary analysis of antidepressant trials in the FDA trial registry database and related journal publications. BMJ. 2009, 339: b2981-10.1136/bmj.b2981.View ArticlePubMedPubMed CentralGoogle Scholar
- Rucker G, Schwarzer G, Carpenter JR, Binder H, Schumacher M: Treatment-effect estimates adjusted for small-study effects via a limit meta-analysis. Biostatistics. 2011, 12 (1): 122-142. 10.1093/biostatistics/kxq046.View ArticlePubMedGoogle Scholar
- Song F, Eastwood AJ, Gilbody S, Duley L, Sutton AJ: Publication and related biases. Health Technol Assess. 2000, 4 (10): 1-115.Google Scholar
- Hedges LV: [Selection models and the file drawer problem]: comment. Stat Sci. 1988, 3 (1): 118-120. 10.1214/ss/1177013013.View ArticleGoogle Scholar
- Bedrick EJ, Christensen R, Johnson W: A new perspective on priors for generalized linear models. JASA. 1996, 91 (436): 1450-1460.View ArticleGoogle Scholar
- Wu Y, Shih WJ, Moore DF: Elicitation of a beta prior for Bayesian inference in clinical trials. Biom J. 2008, 50 (2): 212-223. 10.1002/bimj.200710390.View ArticlePubMedGoogle Scholar
- Salanti G, Ades AE, Ioannidis JP: Graphical methods and numerical summaries for presenting results from multiple-treatment meta-analysis: an overview and tutorial. J Clin Epidemiol. 2010, 64 (2): 163-171.View ArticlePubMedGoogle Scholar
- Spiegelhalter DJ, Best NG, Carlin BP, Van Der Linde A: Bayesian measures of model complexity and fit. J R Stat Soc Ser B Stat Methodol. 2002, 64 (4): 583-639. 10.1111/1467-9868.00353.View ArticleGoogle Scholar
- Dias S, Welton NJ, Marinho VCC, Salanti G, Higgins JPT, Ades AE: Estimation and adjustment of bias in randomized evidence by using mixed treatment comparison meta-analysis. J R Stat Soc Ser A Stat Soc. 2010, 173 (3): 613-629. 10.1111/j.1467-985X.2010.00639.x.View ArticleGoogle Scholar
- Salanti G, Dias S, Welton NJ, Ades AE, Golfinopoulos V, Kyrgiou M, Mauri D, Ioannidis JP: Evaluating novel agent effects in multiple-treatments meta-regression. Stat Med. 2010, 29 (23): 2369-2383.PubMedGoogle Scholar
- Dias S, Welton NJ, Ades AE: Study designs to detect sponsorship and other biases in systematic reviews. J Clin Epidemiol. 2010, 63 (6): 587-588. 10.1016/j.jclinepi.2010.01.005.View ArticlePubMedGoogle Scholar
- Moreno SG, Sutton AJ, Ades AE, Cooper NJ, Abrams KR: Adjusting for publication biases across similar interventions performed well when compared with gold standard data. J Clin Epidemiol. 2011, 64 (11): 1230-1241. 10.1016/j.jclinepi.2011.01.009.View ArticlePubMedGoogle Scholar
- Chootrakool H, Shi JQ, Yue R: Meta-analysis and sensitivity analysis for multi-arm trials with selection bias. Stat Med. 2011, 30 (11): 1183-1198.PubMedGoogle Scholar
- Mavridis D, Sutton A, Cipriani A, Salanti G: A fully Bayesian application of the Copas selection model for publication bias extended to network meta-analysis. Stat Med. 2012, 10.1002/sim.5494. [Epub ahead of print].Google Scholar
- McCandless LC, Gustafson P, Levy AR, Richardson S: Hierarchical priors for bias parameters in Bayesian sensitivity analysis for unmeasured confounding. Stat Med. 2012, 31 (4): 383-396. 10.1002/sim.4453.View ArticlePubMedGoogle Scholar
- Gelman A, Carlin JB, Stern HS, Rubin DB: Bayesian Data Analysis. 2004, New York: Chapman Hall/CRC, 2Google Scholar
- Rising K, Bacchetti P, Bero L: Reporting bias in drug trials submitted to the food and drug administration: review of publication and presentation. PLoS Med. 2008, 5 (11): e217-10.1371/journal.pmed.0050217. discussion e217.View ArticlePubMedPubMed CentralGoogle Scholar
- Lathyris DN, Patsopoulos NA, Salanti G, Ioannidis JP: Industry sponsorship and selection of comparators in randomized clinical trials. Eur J Clin Invest. 2010, 40 (2): 172-182. 10.1111/j.1365-2362.2009.02240.x.View ArticlePubMedGoogle Scholar
- Lexchin J, Bero LA, Djulbegovic B, Clark O: Pharmaceutical industry sponsorship and research outcome and quality: systematic review. BMJ. 2003, 326 (7400): 1167-1170. 10.1136/bmj.326.7400.1167.View ArticlePubMedPubMed CentralGoogle Scholar
- Bero L, Oostvogel F, Bacchetti P, Lee K: Factors associated with findings of published trials of drug-drug comparisons: why some statins appear more efficacious than others. PLoS Med. 2007, 4 (6): e184-10.1371/journal.pmed.0040184.View ArticlePubMedPubMed CentralGoogle Scholar
- Chalmers I, Matthews R: What are the implications of optimism bias in clinical research?. Lancet. 2006, 367 (9509): 449-450. 10.1016/S0140-6736(06)68153-1.View ArticlePubMedGoogle Scholar
- Rucker G, Schwarzer G, Carpenter J: Arcsine test for publication bias in meta-analyses with binary outcomes. Stat Med. 2008, 27 (5): 746-763. 10.1002/sim.2971.View ArticlePubMedGoogle Scholar
- The pre-publication history for this paper can be accessed here:http://www.biomedcentral.com/1471-2288/12/150/prepub
This article is published under license to BioMed Central Ltd. This is an Open Access article distributed under the terms of the Creative Commons Attribution License (http://creativecommons.org/licenses/by/2.0), which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.