- Research article
- Open Access
- Open Peer Review

# Network meta-analysis on the log-hazard scale, combining count and hazard ratio statistics accounting for multi-arm trials: A tutorial

- Beth S Woods
^{1}Email author, - Neil Hawkins
^{1, 2}and - David A Scott
^{1}

**10**:54

https://doi.org/10.1186/1471-2288-10-54

© Woods et al; licensee BioMed Central Ltd. 2010

**Received:**9 May 2009**Accepted:**10 June 2010**Published:**10 June 2010

## Abstract

### Background

Data on survival endpoints are usually summarised using either hazard ratio, cumulative number of events, or median survival statistics. Network meta-analysis, an extension of traditional pairwise meta-analysis, is typically based on a single statistic. In this case, studies which do not report the chosen statistic are excluded from the analysis which may introduce bias.

### Methods

In this paper we present a tutorial illustrating how network meta-analyses of survival endpoints can combine count and hazard ratio statistics in a single analysis on the hazard ratio scale. We also describe methods for accounting for the correlations in relative treatment effects (such as hazard ratios) that arise in trials with more than two arms. Combination of count and hazard ratio data in a single analysis is achieved by estimating the cumulative hazard for each trial arm reporting count data. Correlation in relative treatment effects in multi-arm trials is preserved by converting the relative treatment effect estimates (the hazard ratios) to arm-specific outcomes (hazards).

### Results

A worked example of an analysis of mortality data in chronic obstructive pulmonary disease (COPD) is used to illustrate the methods. The data set and WinBUGS code for fixed and random effects models are provided.

### Conclusions

By incorporating all data presentations in a single analysis, we avoid the potential selection bias associated with conducting an analysis for a single statistic and the potential difficulties of interpretation, misleading results and loss of available treatment comparisons associated with conducting separate analyses for different summary statistics.

## Keywords

- Relative Treatment Effect
- Survival Endpoint
- Cumulative Count
- Random Effect Variance
- Treatment Effect Model

## Background

Network meta-analyses enable us to combine trials that compare different sets of treatments, and form a network of evidence, within a single analysis [1] and to use all available direct and indirect evidence to inform a given comparison between treatments. Network meta-analysis is based on the assumption that, on a suitable scale, we can add and subtract within-trial estimates of relative treatment effects i.e. the difference in effect between treatments A & B (*d*
_{
AB
}) is equal to the difference in effects between treatments A & C and B & C (*d*
_{
AB
}= *d*
_{
AC
}- *d*
_{
BC
}) [1–3].

In this paper we show how network meta-analyses of survival endpoints can be conducted on the hazard ratio scale when some or all trials report cumulative count data. We also describe how trials with more than two arms reporting relative treatment effects (such as hazard ratios) should be included.

A survival endpoint is one where, over time, an increasing number of patients experience an event. Although death is the ultimate survival endpoint, many other endpoints may also be considered as survival endpoints. For example, in the study of epilepsy, seizure freedom may be regarded as a survival endpoint as over time the number of patients experiencing one or more seizures can only increase and the probability of being seizure free decreases. In contrast, the number of patients with a greater than 50% reduction in seizure frequency relative to baseline is not a survival endpoint as the number of patients achieving this response can increase or decrease over time.

Data on survival endpoints may be expressed either in the form of hazard ratio statistics derived from parametric or non-parametric methods of survival analysis [4, 5] or as cumulative count statistics (the total number of subjects who have experienced an event at a specific time point). Unlike cumulative count statistics, hazard ratio statistics from survival analysis account for censoring, incorporate time to event information, and may be adjusted for co-variables. Where both hazard ratio and cumulative count statistics are available for a given study, use of the hazard ratio data may therefore be considered preferable.

The methods described in this paper allow hazard ratio and cumulative count survival statistics to be combined within a single network meta-analysis on the log-hazard scale; this might be termed a multi-statistic evidence synthesis. Treatment effects can then be estimated based on an inclusive set of data, and separate analyses for hazard ratio and count statistics are avoided [6–9].

Network meta-analyses should account for the correlations in relative treatment effect estimates that arise from trials with more than two treatment arms (multi-arm trials) [10]. These correlations are accounted for 'by default' when count statistics for individual trial arms are included in a network meta-analysis [10]. When hazard ratio statistics are used, we show how these correlations can be accounted for by deriving estimates of the mean log hazards (and their variances) for individual trial arms.

## Methods

We use an example data set describing mortality in randomised controlled trials of treatments for chronic obstructive pulmonary disease (COPD). The data set consists of a subset of trials from Baker et al [11]. We chose five trials providing mortality data for the following comparators: salmeterol, fluticasone, salmeterol fluticasone combination (SFC) or placebo. Three of the trials report count data [12–14], one reports hazard ratio data from a two arm trial [15] and one reports hazard ratio data from a multi (four)-arm trial [16]. Hazard ratio data were obtained from the trial publications and an existing meta-analysis of hazard ratio data [17].

The results of this example analysis are purely illustrative of the methodology and do not provide any indication of the comparative effectiveness of treatments for decision-making purposes, as the data set omits relevant direct and indirect data.

- (a)
Perform a meta-analysis of count statistics on the log hazard ratio scale;

- (b)
Reflect correlation in relative treatment effect estimates from multi-arm trials;

- (c)
Combine count and hazard ratio statistics in a single analysis on the log-hazard ratio scale; and

- (d)
Include a random effect in to the analysis, whilst preserving the correlation in relative treatment effects for multi-arm trials.

We also discuss how other possible presentations of survival data, not available from our motivating example, could be incorporated into an analysis on the log-hazard ratio scale.

All analyses described were conducted using WinBUGS [18]. The WinBUGS code for the fixed and random effects analyses are presented as an Appendix.

*(a) Meta-analysing count statistics on the log hazard ratio scale*

where *r*
_{
s,k
}is the cumulative count of subjects who have experienced an event in arm *k* of study *s*; *n*
_{
s,k
}is the total number of subjects in arm *k* of study *s*; and *F*
_{
s,k
}is the cumulative probability of a subject having experienced an event (or 'failure').

*α*

_{ s }and a treatment effect coefficient

*β*

_{ k }:

where *β*
_{1} = 0 for the reference treatment (placebo in our example) and *β*
_{
b
}represents the treatment effect for the baseline treatment in study *s*. The fixed study level 'baseline' term is a nuisance parameter, included to ensure that the treatment effect estimates are informed by within trial differences between treatment arms and not by differences in baseline event rates across trials.

*β*

_{ k }coefficient is equal to both the log cumulative hazard ratio and the log hazard ratio:

where *h*
_{
s,b
}represents the hazard for the baseline treatment in study *s*. This identity allows us to combine the count statistics analysed on the log cumulative hazard scale with the hazard ratio data analysed on the log hazard scale. It also demonstrates that analysis of count statistics on the log hazard scale does not require a stronger assumption than proportional hazards.

*(b) Reflecting correlations in relative treatment effects from multi-arm trials*

Estimates of relative treatment effects from trials with more than two treatment arms will be correlated [10]. In our example the TORCH trial [16] is the only multi-arm trial reporting hazard ratio data. Estimated treatment effects from TORCH will be correlated, for example, the hazard ratio comparing SFC to placebo and the hazard ratio comparing salmeterol to placebo will be correlated due to their joint dependence on the time to event data in the placebo arm.

If a network meta-analysis is based on estimates of treatment effect in individual trial arms rather than estimates of relative treatment effect between arms ("contrast" statistics), this correlation will automatically be captured in the analysis [10]. This is the case when a network meta-analysis is based on count statistics. However, if the network meta-analysis is conducted based on estimates of relative treatment effect, this correlation between arms will not automatically be captured [10].

For multi-arm trials reporting hazard ratio statistics, this problem can be addressed by converting the log hazard ratios (contrast statistics) to log hazards (arm-specific statistics). Log hazards for individual trial arms are derived by nominally setting the log hazard for the baseline treatment *b*for the trial to zero. The mean log hazards for the other treatments are then equal to the log hazard ratios compared to baseline treatment.

Where *se*
^{2}
_{
i,j
}is the variance of the log hazard ratio comparing arm *i* to arm *j* and *se*
_{
i
}is the standard error of the log hazard for arm *i*.

Hazard ratio and log hazard ratio statistics

Author/Trial (Date) | Treatment | Base | HR | HR | HR | ||
---|---|---|---|---|---|---|---|

Burge/ISOLDE (2000) [15] | Fluticasone | Placebo | 0.76 | 0.51 | 1.13 | -0.276 | 0.203 |

Calverly/TORCH (2007) [16] | SFC | Placebo | 0.811 | 0.670 | 0.982 | -0.209 | 0.098 |

Salmeterol | Placebo | 0.857 | 0.710 | 1.035 | -0.154 | 0.096 | |

Fluticasone | Placebo | 1.056 | 0.883 | 1.264 | 0.055 | 0.092 | |

SFC | Salmeterol | 0.946 | 0.777 | 1.151 | -0.056 | 0.100 | |

SFC | Fluticasone | 0.768 | 0.636 | 0.927 | -0.264 | 0.096 |

Log hazard statistics for a multi-arm trial (TORCH [16])

Comparator | Log-hazard | se(log-hazard) |
---|---|---|

SFC | -0.209 | 0.072 |

Salmeterol | -0.154 | 0.070 |

Fluticasone | 0.055 | 0.063 |

Placebo | 0.000 | 0.066 |

In order to estimate standard errors of the log hazards for each treatment, we required estimates of the uncertainty associated with four treatment contrasts. In some cases this data may not be available and thus the methods presented in equations 7 and 8 may not be feasible. For example, hazard ratios and associated measures of uncertainty may only be available for each active treatment relative to a single common comparator (e.g. placebo) as is commonly reported in the published literature.

*(c) Combining count and hazard ratio statistics in a network meta-analysis*

*k*to

*b*are incorporated in the network meta-analysis model using a normal likelihood:

where
is the log hazard ratio estimate for study s comparing treatments *k* to *b* and
is the corresponding variance.

*s*comparing treatments

*k*and

*b*equal to the difference between the two treatment coefficients:

where *β*
_{1} = 0 for the reference treatment (in our example placebo) and *β*
_{
b
}represents the treatment effect for the baseline treatment in study *s*. As in equation 5, the *β*
_{
k
}coefficient is equal to the log hazard ratio for treatment *k* compared to the reference treatment.

*b*:

where
is the log hazard for treatment arm *k* from study *s* and
is the associated variance.

*α*

_{ s }and a treatment effect coefficient

*β*

_{ k }:

where *β*
_{1} = 0 for the reference treatment (placebo in our example) and *β*
_{
b
}represents the treatment effect for the baseline treatment in study *s*. The fixed study level 'baseline' term is a nuisance parameter, included to ensure that the treatment effect estimates are informed by within trial differences between treatment arms and not by differences in baseline event rates across trials. As the *β*
_{
k
}coefficient is equal to the log hazard ratio for the cumulative count data, the log hazard ratio data and the log hazard data, they can be combined within a single analysis. Where an individual study reports both cumulative count and hazard ratio data, only one set of data should be included in the analysis to avoid double counting.

*(d) Incorporating a random effect*

In a random effects analysis of a network containing multi-arm trial contrast data, the correlation in the random effects must also be taken in to account. Again this is due to the joint dependence of the multiple contrast estimates on common trial arms.

*re*

_{ s,k }is the random effect deviation for arm

*k*of study

*s*and is assumed to be normally distributed with zero mean and variance

*σ*

^{2}/2 where

*σ*

^{2}is the random effect variance for a treatment contrast:

This approach assumes that *σ*
^{2} is the same for all treatments and consequently that the random effect variance will be the same for all treatment contrasts. The assumption of a common random effect variance across treatment contrasts implies that the covariance for any pair of treatment contrasts from the same study will equal half the treatment contrast random effect variance [10].

A vague prior for the study specific baseline *α*
_{s}~*N*(0,10^{6}) is used to ensure estimates of treatment effect are informed by within trial differences between treatment arms, and not by differences in absolute response between trials. A vague prior is also used for the treatment effect coefficients with *β*
_{
k
}~ *N*(0,10^{6}) and *β*
_{1} = 0 (representing placebo).

Each model was run for 40,000 burn-in simulations and 200,000 runs which were then thinned every 20^{th} simulation to reduce autocorrelation.

Two sets of initial values were used and convergence was assessed by examining caterpillar plots and Brooks Gelman-Rubin (BGR) statistics. The deviance information criteria (DIC) was used to compare the fit of the fixed and random effects models [19].

## Results

Network meta-analysis results

Comparator | Hazard ratio (95% CI) vs. placebo - fixed effects | Hazard ratio (95% CI) vs. placebo - random effects |
---|---|---|

Fluticasone | 0.99 (0.84, 1.16) | 0.89 (0.39, 1.42) |

Salmeterol | 0.82 (0.68, 0.98) | 0.73 (0.29, 1.23) |

SFC | 0.78 (0.64, 0.93) | 0.69 (0.26, 1.21) |

random effect SD | - | 0.36 (0.31) |

DIC | 25.25 | 25.73 |

Again it must be noted that this example data set is not an appropriate basis for answering the clinical question of which is the most efficacious treatment with respect to the mortality endpoint.

## Discussion

In this tutorial we have described how network meta-analysis can be conducted on the log-hazard ratio scale when the data on the survival endpoint available from the network of studies takes varying forms: count statistics and hazard ratio statistics from two and multi-arm trials.

In highlighting the importance of network meta-analysis as an extension to conventional pairwise meta-analysis Ades et al point out that "...to ignore indirect evidence either makes the unwarranted claim that it is irrelevant, or breaks the established precept of systematic review that synthesis should embrace all available evidence" [6]. A similar criticism can be levelled at meta-analyses that omit data points on the basis of the summary measure reported. This may result in intentional or unintentional selection bias. Furthermore, even if separate analyses are conducted for each summary measure available, this may have undesirable consequences by: producing conflicting results; producing coherent results that would not be supported by an 'all embracing' analysis; and unlinking some comparators from the network entirely.

Other authors have presented methods for combining arm-based mean change from baseline data with contrast mean difference data [21] and for combining mean time, median time and count data within a parametric survival model [22]. Further research is required in this area, for example in to methods for combining incidence rate data with count data, as these summary measures are often reported interchangeably across trials.

Other approaches to combining alternative reporting of time to event data have been discussed [8, 23–26], recommended [27] and implemented [28, 29]. These approaches involve the approximation of the log hazard ratio and its variance using available count statistics. The methods used in this tutorial avoid the use of approximations, instead utilising cumulative count statistics directly in the analysis.

*r*

_{ s,k }should be calculated outside the model to ensure it is a whole number. A log hazard for each trial arm ln (

*h*

_{ s,k }) is then derived from the cumulative failure probability

*F*

_{ s,k }and median survival time

*T*

_{ s,k }:

The arm-specific log hazard estimates are then included in a treatment effect model that takes the same form as that for hazard ratio reporting multi-arm trials (see equation 14). It should be noted that analysis of median survival times requires the strong assumption of a constant hazard in each trial arm. The code for incorporating median survival time data is also provided in the Appendix.

Where multiple statistics are reported for a trial, hazard ratio statistics should be used in preference to median survival time and count data as hazard ratio statistics incorporate information about time to event and censoring, and analysis of hazard ratio statistics does not require assumptions stronger than proportional hazards. If both median survival time and count data are reported, a judgement is required to weigh up the relative merits of each presentation, as median survival time incorporates information about censoring but its analysis requires the stronger assumption of constant hazards.

Finally, this tutorial discusses methods for running analyses on the log hazard ratio scale. However, the question of which is the most appropriate scale for a given analysis (the scale on which transitivity and exchangeability are most likely to hold) - is an empirical question. Further work is required to develop methods for selecting the most appropriate scale for a given data set.

## Conclusions

Meta-analysis of summary statistics continues to play an important role in medical research [30, 31]. Where only summary statistics are available, the analyst performing pairwise or network meta-analysis may be faced with multiple summary statistics for a given endpoint. We present methods for meta-analysing different statistics summarising survival data on the hazard ratio scale. By incorporating all data presentations in a single analysis, we avoid the potential selection bias associated with conducting an analysis for a single statistic only, and the potential difficulties of interpretation, misleading results and loss of available treatment comparisons associated with conducting separate analyses for different summary measures. The methods described also allow use of the most informative statistic available from each study.

## Appendix - WinBUGS code

*A) Fixed effects analysis*

model{

#Define Prior Distributions

#On tx effect mean

beta[1] < -0

for (tt in 2:nTx){

beta[tt]~dnorm(0,1.0E-6)

}

#On individual study baseline effect

for(ss in 1:nStudies){

alpha[ss] ~ dnorm(0,1.0E-6)

}

#Fit data

#For hazard ratio reporting studies

for(ii in 1:LnObs ){

Lmu[ii] < - alpha[Lstudy[ii]]*multi[ii] + beta[Ltx[ii]] - beta[Lbase[ii]]

Lprec[ii] < - 1/pow(Lse[ii],2)

Lmean[ii] ~ dnorm(Lmu[ii],Lprec[ii])

}

#For binary data reporting studies

for(ss in 1:BnObs){

logCumHaz[ss] < - alpha[Bstudy[ss]] + beta[Btx[ss]] - beta[Bbase[ss]]

cumFail[ss] < - 1-exp(-1*exp(logCumHaz[ss]))

Br[ss] ~ dbin(cumFail[ss], Bn[ss])

}

# Calculate HRs

for (hh in 1:nTx) {

hr[hh] < -exp(beta[hh])

}

# Ranking plot

for (ll in 1:nTx) {

for (mm in 1:nTx) {

rk[ll,mm] < - equals(ranked(beta[],mm),beta[ll])

}

}

}

# Data

# Data set descriptors

list(LnObs = 5, BnObs = 8, nTx = 4, nStudies = 5)

# Log hazard ratio and log hazard data

Lstudy[] Ltx[] Lbase[] Lmean[] Lse[] multi[]

1 1 1 0 0.066 1

1 2 1 0.055 0.063 1

1 3 1 -0.154 0.070 1

1 4 1 -0.209 0.072 1

2 2 1 -0.276 0.203 0

END

# Binary data

Bstudy[] Btx[] Bbase[] Br[] Bn[]

3 3 1 1 229

3 1 1 1 227

4 2 1 4 374

4 3 1 3 372

4 4 1 2 358

4 1 1 7 361

5 3 1 1 554

5 1 1 2 270

END

# Initial values

list(alpha = c(-0.50,-0.50,-0.50,-0.50,-0.50), beta = c(NA,-0.5,-0.5,-0.5))

list(alpha = c(0.50,0.50,0.50,0.50,0.50), beta = c(NA,0.5,0.5,0.5))

*B)Random effects analysis* (changes required to incorporate random effect in bold)

model{

#Define Prior Distributions

#on random tx effect variance

**sd~dunif(0,5)**

**reTau < - 2/pow(sd,2)**

#On tx effect mean

beta[1] < -0

for (tt in 2:nTx){

beta[tt]~dnorm(0,1.0E-6)

}

#On individual study baseline effect

for(ss in 1:nStudies){

alpha[ss] ~ dnorm(0,1.0E-6)

}

**#Define random effect**

**for (ss in 1:nStudies){**

**for(tt in 1:nTx){**

**re[ss,tt]~dnorm(0,reTau)**

**}**

**}**

#Fit data

#For hazard ratio reporting studies

for(ii in 1:LnObs ){

Lmu[ii] < - alpha[Lstudy[ii]]*multi[ii] **+ re[Lstudy[ii],Ltx[ii]] -**

**re[Lstudy[ii],Lbase[ii]]** + beta[Ltx[ii]] - beta[Lbase[ii]]

Lprec[ii] < - 1/pow(Lse[ii],2)

Lmean[ii] ~ dnorm(Lmu[ii],Lprec[ii])

}

#For binary data reporting studies

for(ss in 1:BnObs){

logCumHaz[ss] < - alpha[Bstudy[ss]] **+ re[Bstudy[ss],Btx[ss]] -**

**re[Bstudy[ss],Bbase[ss]]** + beta[Btx[ss]] - beta[Bbase[ss]]

cumFail[ss] < - 1-exp(-1*exp(logCumHaz[ss]))

Br[ss] ~ dbin(cumFail[ss], Bn[ss])

}

# Calculate HRs

for (hh in 2:nTx) {

hr[hh] < -exp(beta[hh])

}

# Ranking plot

for (ll in 1:nTx) {

for (mm in 1:nTx) {

rk[ll,mm] < - equals(ranked(beta[],mm),beta[ll])

}

}

}

# Data as for fixed effects analysis

############################

# Initial values

list(alpha = c(-0.50,-0.50,-0.50,-0.50,-0.50), beta = c(NA,-0.5,-0.5,-0.5),**sd = 0.1**)

list(alpha = c(0.50,0.50,0.50,0.50,0.50), beta = c(NA,0.5,0.5,0.5),**sd = 1**)

*C) Additional code required for data reported as median survival times*

for (ii in 1:medianNObs ){

medianMu[ii] < - alpha[medianStudy[ii]] + beta[medianTx[ii]] -

beta[medianBase[ii]]

prob[ii] < - exp(-median[ii]*exp(medianMu[ii]))

medianR[ii] ~ dbin(prob[ii],medianN[ii])

}

## Declarations

### Acknowledgements

James Roger and Oliver Keene are acknowledged for their valuable comments on the methods presented in this paper.

## Authors’ Affiliations

## References

- Caldwell DM, Ades AE, Higgins JPT: Simultaneous comparison of multiple treatments: combining direct and indirect evidence. BMJ. 2005, 331: 897-900. 10.1136/bmj.331.7521.897.View ArticlePubMedPubMed CentralGoogle Scholar
- Ades AE: A chain of evidence with mixed comparisons: model for multi-parameter synthesis and consistency of evidence. Stat Med. 2003, 22 (19): 2995-3016. 10.1002/sim.1566.View ArticlePubMedGoogle Scholar
- Hasselblad V: Meta-analysis of multitreatment studies. Med Decis Making. 1998, 18: 37-43. 10.1177/0272989X9801800110.View ArticlePubMedGoogle Scholar
- Collett D: Modelling survival data in medical research. 2003, Chapman & Hall/CRC, SecondGoogle Scholar
- Kleinbaum DG, Klein M: Survival analysis: a self-learning text. 2005, Springer, U.S, SecondGoogle Scholar
- Ades A, Sculpher M, Sutton AJ, Cooper NJ, Abrams KR, Welton N, Lu G: Bayesian Methods for Evidence Synthesis in Cost-Effectiveness Analysis. Pharmacoeconomics. 2006, 24: 1-19. 10.2165/00019053-200624010-00001.View ArticlePubMedGoogle Scholar
- Chan AW, Hróbjartsson A, Haahr MT, Gøtzsche PC, Altman DG: Empirical evidence for selective reporting of outcomes in randomized trials: comparison of protocols to published articles. JAMA. 2004, 291 (20): 2457-65. 10.1001/jama.291.20.2457.View ArticlePubMedGoogle Scholar
- Parmar MKB, Torri V, Stewart L: Extracting summary statistics to perform meta-analyses of the published literature for survival endpoints. Stat Med. 1998, 17: 2815-34. 10.1002/(SICI)1097-0258(19981230)17:24<2815::AID-SIM110>3.0.CO;2-8.View ArticlePubMedGoogle Scholar
- Whitehead A: Meta-Analysis of Controlled Clinical Trials. 2002, John Wiley and Sons: ChichesterView ArticleGoogle Scholar
- Salanti G, Higgins JP, Ades AE, Ioannidis JP: Evaluation of networks of randomised trials. Stat Methods Med Res. 2008, 17 (3): 279-301. 10.1177/0962280207080643.View ArticlePubMedGoogle Scholar
- Baker WL, Baker EL, Coleman C: Pharmacologic treatments for chronic obstructive pulmonary disease: a mixed-treatment comparison meta-analysis. Pharmacotherapy. 2009, 29 (8): 891-905. 10.1592/phco.29.8.891.View ArticlePubMedGoogle Scholar
- Boyd G, Morice AH, Pounsford JC, Siebert M, Peslis N, Crawford C: An evaluation of salmeterol in the treatment of chronic obstructive pulmonary disease (COPD). Eur Respir J. 1997, 10: 815-21.PubMedGoogle Scholar
- Calverley P, Pauwels R, Vestbo J, Jones P, Pride N, Gulsvik A, Anderson J, Maden C, Trial of Inhaled Steroids and Long-Acting β2-Agonists (TRISTAN) Study Group: Combined salmeterol and fluticasone in the treatment of chronic obstructive pulmonary disease: a randomized controlled trial. Lancet. 2003, 361: 449-56. 10.1016/S0140-6736(03)12459-2.View ArticlePubMedGoogle Scholar
- Celli B, Halpin D, Hepburn R, Byrne N, Keating ET, Goldman M: Symptoms are an important outcome in chronic obstructive pulmonary disease clinical trials: results of a 3-month comparative study using the breathlessness, cough, and sputum scale (BCSS). Resp Med. 2003, 97 (suppl A): S35-43.View ArticleGoogle Scholar
- Burge PS, Calverley PMA, Spencer S, Anderson JA, Maslen TK, on behalf of the ISOLDE Study Investigators: Randomised, double-blind, placebo-controlled study of fluticasone propionate in patients with moderate to severe chronic obstructive pulmonary disease: the ISOLDE trial. BMJ. 2000, 320: 1297-303. 10.1136/bmj.320.7245.1297.View ArticlePubMedPubMed CentralGoogle Scholar
- Calverley PMA, Anderson JA, Celli B, Ferguson GT, Jenkins C, Jones PW, Yates JC, Vestbo J, for the TORCH Investigators: Salmeterol and fluticasone propionate and survival in chronic obstructive pulmonary disease. N Engl J Med. 2007, 356: 775-89. 10.1056/NEJMoa063070.View ArticlePubMedGoogle Scholar
- Sin DD, Wu L, Anderson JA: Inhaled corticosteroids and mortality in chronic obstructive pulmonary disease. Thorax. 2005, 60: 992-7. 10.1136/thx.2005.045385.View ArticlePubMedPubMed CentralGoogle Scholar
- Lunn DJ, Thomas A, Best N, Spiegelhalter D: WinBUGS - a Bayesian modelling framework: concepts, structure, and extensibility. Stat Comput. 2000, 10: 325-337. 10.1023/A:1008929526011.View ArticleGoogle Scholar
- Spiegelhalter DJ, Best NJ, Carlin BP, Van der Linde A: Bayesian measures of model complexity and fit (with discussion). J R Stat Soc. 2002, 64: 1-34. 10.1111/1467-9868.00353.View ArticleGoogle Scholar
- The BUGS project: DIC: Deviance Information Criterion. Accessed 10/02/2010, [http://www.mrc-bsu.cam.ac.uk/bugs/winbugs/dicpage.shtml#q9]
- Welton NJ: Mixed treatment comparisons: outcome measures. Lecture notes from Indirect and Multiple Comparisons course. 2008, Leicester UniversityGoogle Scholar
- Welton NJ, Cooper NJ, Ades AE, Lu G, Sutton AJ: Mixed treatment comparison with multiple outcomes reported inconsistently across trials: Evaluation of antivirals for treatment of influenza A and B. Stat Med. 2008, 27: 5620-39. 10.1002/sim.3377.View ArticlePubMedGoogle Scholar
- Ma J, Liu W, Hunter A, Zhang W: Performing meta-analysis with incomplete statistical information in clinical trials. BMC Med Res Methodol. 2008, 8 (1): 56-10.1186/1471-2288-8-56.View ArticlePubMedPubMed CentralGoogle Scholar
- Michiels S, Piedbois P, Burdett S, Syz N, Stewart L, Pignon JP: Meta-analysis when only the median survival times are known: a comparison with individual patient data results. Int J Technol Assess. 2005, 21 (1): 119-25.View ArticleGoogle Scholar
- Tierney JF, Stewart LA, Ghersi D, Burdett S, Sydes MR: Practical methods for incorporating summary time-to-event data into meta-analysis. Trials. 2007, 8: 16-10.1186/1745-6215-8-16.View ArticlePubMedPubMed CentralGoogle Scholar
- Williamson PR, Tudur Smith C, Hutton JL, Marson AG: Aggregate data meta-analysis with time-to-event outcomes. Stat Med. 2002, 21: 3337-51. 10.1002/sim.1303.View ArticlePubMedGoogle Scholar
- Higgins JPT, Deeks JJ, (editors): Chapter 7: Selecting studies and collecting data. Cochrane Handbook for Systematic Reviews of Interventions Version 5.0.0 (updated February 2008). The Cochrane Collaboration. Edited by: Higgins JPT, Green S. 2008, [http://www.cochrane-handbook.org]
- Green JA, Kirwan JM, Tierney JF, Symonds P, Fresco L, Collingwood M, Williams CJ: Survival and recurrence after concomitant chemotherapy and radiotherapy for cancer of the uterine cervix: a systematic review and meta-analysis. Lancet. 2001, 358 (8): 781-86. 10.1016/S0140-6736(01)05965-7.View ArticlePubMedGoogle Scholar
- Langendijk JA, Leemans CR, Buter J, Berkhof J, Slotman BJ: The Additional Value of Chemotherapy to Radiotherapy in Locally Advanced Nasopharyngeal Carcinoma: A Meta-Analysis of the Published Literature. J Clin Oncol. 2004, 22 (22): 4604-12. 10.1200/JCO.2004.10.074.View ArticlePubMedGoogle Scholar
- Lyman GH, Kuderer NM: The strengths and limitations of meta-analyses based on aggregate data. BMC Med Res Methodol. 2005, 5 (1): 14-10.1186/1471-2288-5-14.View ArticlePubMedPubMed CentralGoogle Scholar
- Riley RD, Simmonds MC, Look MP: Evidence synthesis combining individual patient data and aggregate data: a systematic review identified current practice and possible methods. J Clin Epidemiol. 2007, 60 (5): 431-9. 10.1016/j.jclinepi.2006.09.009.View ArticlePubMedGoogle Scholar
- The pre-publication history for this paper can be accessed here:http://www.biomedcentral.com/1471-2288/10/54/prepub

### Pre-publication history

## Copyright

This article is published under license to BioMed Central Ltd. This is an Open Access article distributed under the terms of the Creative Commons Attribution License (http://creativecommons.org/licenses/by/2.0), which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.