Skip to main content

A systematic survey shows that reporting and handling of missing outcome data in networks of interventions is poor

Abstract

Background

To provide empirical evidence about prevalence, reporting and handling of missing outcome data in systematic reviews with network meta-analysis and acknowledgement of their impact on the conclusions.

Methods

We conducted a systematic survey including all published systematic reviews of randomized controlled trials comparing at least three interventions from January 1, 2009 until March 31, 2017.

Results

We retrieved 387 systematic reviews with network meta-analysis. Description of missing outcome data was available in 63 reviews. Intention-to-treat analysis was the most prevalent method (71%), followed by missing outcome data investigated as secondary outcome (e.g., acceptability) (40%). Bias due to missing outcome data was evaluated in half the reviews with explicit judgments in 18 (10%) reviews. Only 88 reviews interpreted their results acknowledging the implications of missing outcome data and mostly using the network meta-analysis results on missing outcome data as secondary outcome. We were unable to judge the actual strategy applied to deal with missing outcome data in 65% of the reviews due to insufficient information. Six percent of network meta-analyses were re-analyzed in sensitivity analysis considering missing outcome data, while 4% explicitly justified the strategy for dealing with missing outcome data.

Conclusions

The description and handling of missing outcome data as well as the acknowledgment of their implications for the conclusions from network meta-analysis are deemed underreported.

Peer Review reports

Background

Missing outcome data (MOD) are a distinct type of missing information attributed to multifaceted causes that prematurely terminate the participation in a research study, including a clinical trial. These causes may relate to the design and conduct of the clinical trial or be completely irrelevant to the clinical trial structure. Besides sample size losses, discontinuation may distort the balance of the baseline characteristics between the compared groups in a trial leading to confounding and selection bias [1,2,3,4,5,6]. Furthermore, participants remaining in the trial might not be representative of the population originally recruited and reduce directness of the findings [1,2,3, 7].

Missing outcome data are an integral part of a systematic review as they emerge inevitably through the inclusion of clinical trials with reported participant losses. Since the risk of bias due to MOD at the trial level is likely to translate into a similar risk at the meta-analysis level, integration of trials with MOD may lead to meta-analysis results that do not reflect the impact of the interventions in practice [8,9,10,11]. Consequently, strategic management of MOD in a systematic review is particularly necessary [4, 5].

There is guidance for systematic reviewers, meta-analysts, journal editors and reviewers in order to improve reporting and handling of MOD at meta-analysis level [9, 12,13,14]. In addition, statistical methods, software and tutorials for dealing with MOD of varying complexity have been developed over time [8, 11, 14,15,16,17,18,19,20,21,22,23]. Nevertheless, MOD in systematic reviews with network meta-analysis have received very little attention. Network meta-analysis (NMA) is an extension of pairwise meta-analysis that allows direct (i.e., trials investigating the same comparison) and indirect (i.e., different sets of trials that compare each intervention with a common comparator) evidence to be synthesized simultaneously in a single analysis in order to yield relative intervention effects for all possible comparisons and by extent, to rank the multiple competing interventions [24]. While methods for handling MOD in pairwise meta-analysis may be applicable to NMA, addressing MOD in the context of NMA holds an additional degree of complexity. This complexity inherently stems from the additional assumptions required to integrate different pieces of evidence from multiple trials in a single analysis and obtain internally coherent relative treatment effects for all pairwise comparisons [25]. Therefore, the presence of MOD in a network of interventions might materially impact on additional parameters beyond the standard meta-analytic ones (i.e., mean effect and between-study variance) including incoherence of the estimates between direct and indirect evidence for a particular comparison, probabilistic statements on the comparability of all competing interventions and intervention ranking.

Currently, publications on MOD in the NMA context are restricted to a post hoc evaluation that uses the informative missingness odds ratio (IMOR) parameter [26] to assess the impact of missing binary outcome data on NMA results; an extension of the IMOR parameter for missing continuous outcome data [18] and a meta-epidemiological study that investigates the association between specific trial characteristics and the likelihood of premature discontinuation in antipsychotic trials for schizophrenia [27].

While the recommendations on the reporting and handling of MOD in conventional systematic reviews are of great relevance and importance also for systematic reviews with NMA, they can only partially assist the interested audience since they do not reflect upon the additional NMA components, namely, consistency assumption and intervention ranking. Properly established guidelines cannot exist without empirical evidence and to our knowledge there is no empirical evidence on the reporting and handling of MOD in networks with multiple interventions (e.g., similar to [12]). The aim of this study was to bridge this particular knowledge gap by providing evidence about the impact of MOD on the credibility of a systematic review with NMA, including the reporting and handling of MOD as well as the acknowledgment of their implications.

Methods

Eligible systematic reviews

We used several databases to identify NMAs of at least three interventions. First, we based our sample on the published database of Zarin et al. [28]. This database includes 456 systematic reviews with NMA of randomized controlled trials (RCTs) in all languages with at least four different interventions from inception until April 14, 2015. The authors excluded systematic reviews (i) of diagnostic test accuracy studies or genetic studies or observational studies or mixture of RCTs and observational studies; (ii) those that included a smaller number of trials than interventions; and (iii) had implemented naïve indirect comparison methods.

Second, we considered also the databases by Tan et al. and Bafeta et al. as they included at least three interventions and we added these additional potential NMAs in our database [29, 30]. These databases covered a period from 1997 [29] to July 2012 [30]. Third, we searched in addition the database of Nikolakopoulou et al. [31] (they searched from inception to December 2012) to locate possible systematic reviews that might have been missed by Zarin et al. [28].

Finally, to make the sample of systematic reviews current, we conducted our own search using the eligibility criteria (i) – (iii) considered in Zarin et al. [28] for eligible systematic reviews published from August 2012 to March 2017 that assessed at least 3 interventions. For that search, we applied the strategies developed by Petropoulou et al. [32] (and used in Zarin et al. [28] as well) for MEDLINE, EMBASE and the Cochrane Database of Systematic Reviews.

We included reviews published from 2009 and onwards, because the new Cochrane risk of bias tool was published during 2009 and hence, we expected the reviewers to have routinely assessed the included studies also in terms of bias due to MOD [12]. We considered whether RCTs with ‘non-standard’ design, such as quasi, crossover, factorial, cluster, split-mouth, contralateral and split-face/body RCTs, were among the eligible trials. Since the methodology to handle MOD has been developed primarily in the context of standard RCTs, we decided to exclude systematic reviews that included RCTs with ‘non-standard’ design. Furthermore, we excluded commentaries, letters, editorials, case-series, protocols, methodological articles, overviews of systematic reviews (MOD have been addressed already within the systematic review), cost-effectiveness reviews of multiple interventions that did not perform de novo systematic review with NMA (but used results from published NMAs, published meta-analyses and selected trials) and systematic reviews that investigated MOD as a single primary outcome usually termed as ‘acceptability’ or ‘withdrawal due to adverse events or any reason’.

Eligible network meta-analyses

From each eligible systematic review, we selected only one primary outcome. When the authors described more than one primary outcomes, we gave priority to binary outcomes, since established methods to handle MOD in systematic reviews address mainly binary outcomes. In the presence of many binary primary outcomes, we selected a patient-important outcome following the hierarchy of outcomes as defined by Akl et al. [33].

Screening

Two reviewers (LMS, JJYN) double-screened in a standardized approach the titles and abstracts of the systematic reviews already included in Bafeta et al. [30], Tan et al. [29] and Nikolakopoulou et al. [31] as well as those retrieved from the updated literature search. Reasons for exclusion were documented. Then, potentially relevant systematic reviews were screened in full-text. In case of conflicts, a third reviewer (HJS) was consulted.

Data collection process and data items

We developed a data extraction form and piloted on randomly-selected eligible systematic reviews in order to determine the finalized content of the extraction form. One reviewer (LMS) extracted all necessary information from the eligible systematic reviews. A second reviewer (JJYN) randomly selected and checked 10% of the extracted systematic reviews for potential errors. Disagreements were resolved by discussion between the reviewers; in case of no concordance, the authors consulted a third reviewer (HJS).

Systematic review level

We extracted items that referred to specific domains defined separately for the systematic review as a whole and the selected NMA. Specifically, at the systematic review level, we extracted information on general characteristics of the review, such as year of publication, disease condition, protocol availability and incorporation of the GRADE (Grading of Recommendations, Assessment, Development and Evaluation) approach. Items referring to reporting and handling of MOD in the protocol and published report of the systematic review included the definition of MOD, MOD strategy and adherence to the pre-specified strategy (if a protocol exists), possible explanation provided by the authors to support their strategy, use of the last-observation-carried-forward (LOCF) approach (i.e., for each participant who withdrew during the course of a trial with longitudinal follow-up, the last observed continuous outcome is forwarded to the remaining pre-specified outcome assessments in order to fill-in the subsequent missing responses [1, 2]) and assessment of bias due to MOD. To locate these items, we also searched the PROSPERO registry (if a registration number was provided) or available Web addresses and published additional file.

Lastly, we extracted information on whether and how authors dealt with the implications of MOD on NMA components (i.e. treatment effects, consistency parameter, intervention ranking, heterogeneity and possible model fit parameters); for instance, we noted authors’ judgments when they used the risk of bias assessment for single studies or the GRADE approach for a body of evidence, remarks on the impact of dropout prevalence and reasons, and discussion of the results of any moderator analyses in the context of MOD, such as subgroup analysis where studies were split into subgroups of low and high risk of bias due to MOD, meta-regression analysis using dropout prevalence as covariate, and sensitivity analysis by excluding trials without intention-to-treat analysis (namely, analysis of all participants in the group they were originally randomized irrespectively protocol compliance or withdrawal [1,2,3]). This information was sought in the abstract, results, discussion and conclusion sections of the publication.

Selected network meta-analysis level

For the selected NMAs, we extracted information on network topology (number of interventions and included studies and network shape), type of intervention-comparison and outcome definition (as defined by Turner et al. [34]) and effect measure. For the selected primary outcome, we recorded the reported choice of analysis set (i.e., (modified) intention-to-treat and per-protocol) and then, we judged the actual method of analysis rather than the reported (e.g., available case analysis or imputation with or without LOCF). Furthermore, information was extracted on the accountability of the uncertainty induced by missingness and specific scenarios considered when imputing MOD.

We evaluated whether the authors applied a sensitivity analysis and what strategies they considered to inspect further the impact of MOD for the selected primary outcome. We documented whether the authors reported any changes in the inferences after sensitivity analysis. Lastly, we determined which systematic reviews provided extractable data information for the selected primary outcome and MOD.

Results

Systematic reviews and network meta-analysis selection

A total of 410 studies located in Zarin et al. [28] were relevant for screening. After supplementing with studies identified in Nikolakopoulou et al. [31], Bafeta et al. [30] and Tan et al. [29], a total of 447 reviews were considered for full-text screening. Of those, 125 were excluded for several reasons (Fig. 1).

Fig. 1
figure1

Flow diagram of systematic reviews with network meta-analysis. ITC, indirect treatment comparison; MA, meta-analysis; NMA, network meta-analysis; RCT, randomized controlled trial

The bibliographic search from August 2012 to March 2017 yielded a total of 446 possibly relevant articles for title and abstract screening. Of those, 46 were duplicate records and 137 had no relevance to systematic review and NMA (e.g., studies and methodology of genetic, laboratory and animal content, cohort studies, cross-sectional studies, case-control studies and case reports). Then, a total of 263 articles were considered for full-text screening. Of those, 198 were excluded for various reasons (Fig. 1). A list with the excluded articles and their reasons can be found in Additional file 1. In total, 387 systematic reviews were considered eligible for data extraction (see, Additional file 2).

Characteristics of selected systematic reviews

The majority were published during 2014 (n = 92, 24%), specialized on cardiology (n = 71, 18%) and oncology (n = 46, 12%) (see Fig. 2 for further details), were non-Cochrane reviews (n = 380, 98%), did not mention in their publication whether a protocol was available (n = 279, 72%) and did not incorporate the GRADE approach in their results (n = 358, 92%, Fig. 3). Sixty-six reviews (17%) mentioned that a protocol was developed but they did not make the protocol available (Table 1). Only 38 (10%) reviews registered or published a protocol with the majority being found between 2013 and 2015.

Fig. 2
figure2

Bubble chart of 387 systematic reviews with network meta-analysis by year of publication and health specialty

Fig. 3
figure3

Mosaic plot of 387 systematic reviews with network meta-analysis to visualize the frequency of incorporation of GRADE approach by year of publication. GRADE, Grading of Recommendations, Assessment, Development and Evaluation

Table 1 Characteristics of 387 systematics reviews in the context of missing outcome data

Addressing and handling of missing outcome data in protocols

Among the systematic reviews that made their protocol available (n = 37, 10%), the majority (n = 23, 62%) did not provide any strategy to address MOD in their analysis (Table 1). Five (13%) reviews planned to handle MOD as a secondary outcome, 9 (24%) planned to follow intention-to-treat analysis, whereas 2 (5.4%) planned to exclude participants with MOD or trials without complete outcome data. Of those 14 (38%) reviews that provided at least one pre-planned strategy to address MOD, only 2 (14%) did not address MOD eventually.

Prevalence in defining missing outcome data

Overall, only 63 (16%) reviews explicitly defined either in their protocol or publication what the reviewers considered as MOD and most were published between 2011 and 2015 (Table 1). For example, Cipriani et al. [35] reported that as secondary outcome “we defined treatment discontinuation (acceptability) as the number of patients who terminated the study early for any reason during the first 8 weeks of treatment (dropouts)”. Other quotes regarding the definition of MOD can be found in Additional file 3.

Further in the definition of MOD, only 23 (6%) reviews explicitly mentioned that the LOCF approach has been already applied in the included trials (most found during 2013 and 2014). However, only 5 (1.3%) reviews clearly distinguished between LOCF and completely MOD (i.e., missing outcomes relating to participants who did not provide any measurements apart from their baseline characteristics before leaving the trial [12]). For example, Filippini et al. [36] imputed completely MOD under the ‘all missing events’ assumption while considering also the LOCF data: “we assumed that the treated- and control-group participants who dropped out and were not included in the study analysis both had the outcome (relapse or disability progression)”.

Prevalence and handling of missing outcome data in systematic reviews

Seventy percent (n = 273) of the systematic reviews explicitly reported that there are MOD in the included trials with most being published between 2011 and 2015, whereas only one systematic review clearly indicated through the risk of bias assessment that no incomplete outcome data exist in any of the included trials (Table 1). Of these 273 reviews, more than half (n = 153, 56%) applied at least one strategy to address MOD in their analysis, especially those reviews published during 2014, whereas the remaining reviews did not describe any such strategy.

Among the 153 reviews that described at least one strategy to handle missingness, the majority (n = 109, 71%) used intention-to-treat analysis but only 25 (23%) reviews explicitly reported what they meant by intention-to-treat analysis (Table 1). Forty percent (n = 61) of the reviews considered MOD as one of the studied outcomes, whereas 4 (2.6%) reviews excluded participants with MOD from the primary analyses. The remaining reviews employed at least one sensitivity analysis (n = 16, 10%), followed by subgroup analysis (n = 6, 4%) and meta-regression analysis (n = 3, 2%) using a specific dropout characteristic (e.g., studies with and without complete outcome data or proportion of dropouts).

Out of 109 systematic reviews that applied intention-to-treat analysis (mainly published during 2013 and 2014), 46 (42%) extracted outcome data as reported in the trials, 31 (28%) applied intention-to-treat analysis de novo and only 25 (80%) reported specific scenarios about MOD, 13 (12%) applied a combination of ‘as reported’ and ‘de novo intention-to-treat analysis’, whereas 19 (17%) reviews provided no explicit information to comprehend how intention-to-treat analysis was implemented (see, Additional file 4).

Evaluation of bias due to missing outcome data in systematic reviews

Almost half the systematic reviews (n = 198, 51%) assessed bias due to MOD using a specific tool to evaluate risk of bias of the trials and most were published during 2013 and 2014; only 18 (9.1%) reviews evaluated explicitly the domain by providing justification for the expressions of low, high, or unclear risk of bias (Table 1). The Cochrane risk of bias tool was the preferred tool in 147 (74%) reviews followed by the Jadad scale (n = 68, 34%) (see, Additional file 5). Nineteen percent (n = 72) of the reviews stated that bias due to MOD has been evaluated but results on that domain are not displayed at all (n = 15, 21%) or an overall score is provided for each trial (n = 37, 51%) or results on the evaluation of risk of bias are omitted altogether (n = 19, 26%). Other domains of risk of bias were evaluated in 25 (6.5%) systematic reviews, whereas no evaluation of risk of bias occurred in the remaining (n = 92, 24%).

Acknowledgment of missing outcome data implications

Out of 273 (70%) systematic reviews that explicitly reported presence of MOD in the included trials, only 88 (32%) interpreted their results while accounting for MOD (mostly found between 2011 and 2014) (Table 1), primarily referring to the implications of MOD on NMA treatment effects (n = 84, 95%) using the NMA results on dropout as secondary outcome (n = 58, 66%) or the sensitivity analysis results (n = 16, 18%). The discussion section (n = 74, 84%) followed by the abstract (n = 46, 52%) were the sections in which the authors primarily discussed the implications of MOD.

Characteristics of selected network meta-analyses

We selected one primary outcome for each systematic review that explicitly or implicitly denoted the presence of MOD in the included trials (n = 273). A median of 18 (IQR: 11–36) RCTs were included in the 273 selected NMAs that assessed a median of 6 (IQR: 4–9) interventions. The majority of the NMAs referred to a full network (n = 200, 73%), compared pharmacological interventions against placebo (n = 150, 55%) and investigated a subjective primary outcome (n = 127, 46%). Binary outcomes were more prevalent (n = 175, 64%) with odds ratio being the most frequently used effect measure (n = 120, 68%). Figure 4 illustrates the distribution of the network shape by intervention-comparator type (left) and the distribution of effect measure by outcome type (right) in 273 selected NMAs.

Fig. 4
figure4

Stacked bar chart of intervention-comparator type and network shape (left) and a stacked bar chart of outcome type and effect measure in 273 selected network meta-analyses (right)

Addressing missing outcome data in selected primary outcomes

A median of 5 RCTs (IQR: 3–12) handled intermediate missing outcomes using the LOCF approach (11 NMAs). Only 113 (41%) NMAs explicitly mentioned the analysis plan adopted (Table 2), and most were published during 2012 and 2014, with intention-to-treat analysis being the most prevalent analysis plan (n = 105, 93%). While judging the actual method of primary analysis rather than the reported method, we concurred only with 14 (12%) out of 113 NMAs (see, Additional file 6). We were able to judge the actual method applied in the primary analysis in 95 (35%) out of 273 NMAs that were published mostly between 2011 and 2014; we judged the great majority of NMAs to have analyzed the data as reported in the included trials (n = 57, 60%). Among the 16 NMAs that were judged to have imputed MOD in the primary analysis, 8 (50%) assumed all MOD to be non-events, 1 (6.2%) assumed all MOD to be events and 7 (44%) provided no information on the missingness scenarios. For the scenarios selected to impute MOD (9 NMAs), the reviewers considered no trial information on the reasons for dropout.

Table 2 Characteristics of 273 selected network meta-analyses

Sensitivity analysis on missing outcome data for selected primary outcomes

Most NMAs were not re-analyzed in a sensitivity analysis (n = 257, 94%) to investigate the implications of MOD on the primary analysis results further (Table 2). Out of 16 NMAs with at least one sensitivity analysis (most published during 2014 and 2015), only 2 (12%) reported that changes were detected when compared with the primary analysis results. Specifically, Delahoy et al. [37] reported that “Both sensitivity analyses [one after excluding missing outcome data and one after imputing all missing cases as failures] supported the findings of the base-case analysis, although statistical significance was not demonstrated”, whereas Cui et al. [38] stated that “[After using per-protocol instead of intention-to-treat analysis] two outcomes for apixaban and dabigatran in preventing total VTE were significantly affected”. Overall, for 12 (4.4%) out of 273 NMAs there was an explicit explanation to support the strategy applied to handle MOD in primary and sensitivity analysis (see, Additional file 7) and the majority was published during 2014.

Availability of primary and missing outcome data for extraction

Only 55 (20%) out of 273 NMAs provided extractable MOD at arm- or trial-level for every included trial (Table 2) with most being published between 2011 and 2015. In 27 (10%) NMAs, MOD were reported at intervention-level, whereas the remaining 191 (70%) NMAs provided no numerical information on MOD. Overall, 39 (14%) NMAs provided necessary numerical information for each arm in every trial (published mostly between 2011 and 2014) in order to enable extraction of both primary and missing outcome data.

Discussion

In this systematic survey, we reviewed a total of 387 systematic reviews with NMA on at least three interventions. Only a fraction described in their protocol a strategy to address MOD or provided a definition of MOD. Overall, there was insufficient information in most systematic reviews with NMA to judge the actual strategy applied to handle MOD. Only a small fraction of NMAs were re-analyzed in at least one sensitivity analysis for MOD and an even smaller number explicitly justified the strategy applied in primary and/or sensitivity analysis to address MOD.

This is the first systematic survey that provides comprehensive evidence about the reporting and handling of MOD in networks of interventions. It suggests that planning, reporting and conduct of analyzing the impact of MOD in NMA results is poor and will benefit from related guidance.

We believe that this work is representative because it vastly expands our previous work in mental health Cochrane reviews [12]. We evaluated twice the systematic reviews covering a wide range of several health specialties [12]. Another strength of our study includes searching a multitude of sources that include 4 previously published collections of NMAs and an update of the original searches using an extensive previously published search strategy [32].

The weaknesses of our study are the lack of guidance for handling MOD in NMA itself. Without this guidance, the authors of the systematic reviews could not follow proper reporting and analysis standards. The overall poor quality of the reporting of the eligible systematic reviews raised a number of challenges for our work. First, only a fraction of reviews provided numerical information on MOD (82 out of 387), whereas 70% of the reviews with potential MOD in the included trials provided only implicit information on MOD and primarily through the assessment of bias due to MOD. As a result, we were not able to appraise the precise extent of MOD in the reviews and by extent, in each clinical field. Second, we were not able to provide an actual estimation of the use of the LOCF approach in the included studies since only 11 networks made available such information (again, mostly through the assessment of bias due to MOD). Therefore, we might have underestimated the extent of the LOCF approach in systematic reviews with NMA. Third, we strived to judge as objectively as possible the actual method applied in the primary analysis by developing a classification of judgments which was then applied in all eligible reviews. However, considering also the particular nature of this task this judgment remained subjective. Finally, we did not attempt to contact the authors of the reviews that mentioned a protocol but did not make the protocol publicly available (1 out of 5 reviews). It is unlikely, though, that these protocols would have addressed MOD more explicitly.

The results of our study are comparable with our previous work [12] focusing on Cochrane reviews with two interventions in mental health and with the work of Akl et al. [13] who investigated Cochrane and non-Cochrane reviews with focus on binary outcome data. We observed that only a small fraction of reviews provided a clear definition for MOD or distinguished between LOCF and completely MOD. Furthermore, most of the reviews addressed MOD either as a secondary outcome or by applying an imputation strategy for the purpose of complying with the intention-to-treat principle. Sensitivity analysis was among the least prevalent methods to address MOD, whereas the interpreted results were hardly discussed in the context of MOD. Akl et al. [13] encountered also a low prevalence in the reporting and handling of MOD. Exclusion of missing participants from the analysis and application of sensitivity analysis were the most and least prevalent methods, respectively. Assessment of bias due to MOD was primarily a characteristic of Cochrane but not of other reviews.

Nevertheless, contrary to our previous work [12], we were not able to judge the actual method applied in the majority of the selected NMAs. This difficulty primarily stems from insufficient information about important components that are necessary to derive the actual method of primary analysis explicitly or implicitly. Additional file 6 shows examples of systematic reviews with explicit, implicit and unclear judgments alongside the justification of our judgments.

Our systematic survey revealed that systematic reviews with NMA rarely develop protocols prior to their conduct, while those that make protocols available in public, tend not to provide a strategy to address MOD. A carefully planned protocol should include explicit information on collecting MOD for each arm in every trial, on the methodology considered (e.g., imputation or likelihood-based methods) and assumptions about missing cases (source and distribution across trials and interventions). Ideally, the reviewers should also attempt to distinguish between completely MOD and LOCF data and record the trials with information on LOCF. In that way, the actual extent of MOD and their implications on the conclusions can be appraised more accurately.

We found that many systematic reviews favored intention-to-treat analysis but only a handful of those provided further information on the technique and assumptions made about the missing cases. Indeed, a great majority of these reviews merely mentioned that, ‘where possible, data were extracted under the intention-to-treat principle as reported in the included trials’. However, the reviewers made no efforts to evaluate the validity of the analysis claimed in these trials, as they seem to have taken at face value the reported results. The latter increases the risk of misconducting an intention-to-treat analysis in the review, since intention-to-treat analysis is notorious for being described and implemented inadequately in trials [39]. Without possessing at least the number analyzed out of the number randomized in the included trials, the reported analysis in the selected outcome may not be following the genuine intention-to-treat principle.

Without information available on post-withdrawal outcomes, MOD are eventually analyzed using untestable assumptions [40]. Therefore, a set of clinically plausible sensitivity analyses needs to be selected to explore the deviations from the assumption made in the primary analysis and by extent, the robustness of the inferences [40, 41]. Nevertheless, our survey demonstrated that sensitivity analyses in the context of MOD are very rarely planned in systematic reviews with NMAs (n = 16, 4.1%). We previously found that only 16% of the eligible meta-analyses applied at least one sensitivity analysis, whereas only half communicated their findings while considering the results of the sensitivity analyses. It is possible that more careful protocol development and peer review of Cochrane reviews provided some protection against this shortcoming. Gamble and Hollis [17] first shed light on the methods used in Cochrane reviews with meta-analyses to address MOD, and they found that half of the systematic reviews planned and performed sensitivity analysis, whereas interpretation of the sensitivity analyses was either unclear or omitted, overall.

Conclusions

Our comprehensive empirical systematic survey indicates that the quality of reporting and handling MOD in systematic reviews with NMAs is inadequate. Despite the published literature on MOD, reviewers remain either unaware of the presence and importance of MOD or unable to plan to deal with MOD in systematic reviews of multiple interventions. In addition, the poor handling of MOD attests to limited knowledge of the reviewers regarding the existing relevant methodology. Therefore, further methods development, reporting guidelines and education amongst reviewers is deemed necessary to allow addressing the implications on the quality of the inferences delivered to the healthcare system. To ensure a transparent, detail-oriented reporting and methodologically appropriate management of MOD in systematic reviews, peer reviewers and journal editors may benefit from training as well.

Abbreviations

GRADE:

Grading of Recommendations, Assessment, Development and Evaluation

IMOR:

Informative missingness odds ratio

LOCF:

Last-observation-carried-forward

MOD:

Missing outcome data

NMA:

Network meta-analysis

RCT:

Randomized controlled trials

References

  1. 1.

    Little RJA, Rubin DB. Statistical analysis with missing data. 2nd ed. New Jersey: John Wiley & Sons, Inc., Publication; 2002.

    Google Scholar 

  2. 2.

    Molenberghs G, Kenward M. Missing data in clinical studies. 1st ed. West Sussex: John Wiley & Sons Ltd; 2007.

    Google Scholar 

  3. 3.

    Fleiss JL, Levin B, Paik MC. Missing data. In: Fleiss JL, Levin B, Paik MC. Statistical Methods for Rates and Proportions. 3rd ed. New Jersey: John Wiley & Sons, Inc., Publication; 2004.

  4. 4.

    Akl EA, Kahale LA, Ebrahim S, Alonso-Coello P, Schünemann HJ, Guyatt GH. Three challenges described for identifying participants with missing data in trials reports, and potential solutions suggested to systematic reviewers. J Clin Epidemiol. 2016;76:147–54.

    Article  Google Scholar 

  5. 5.

    Guyatt GH, Ebrahim S, Alonso-Coello P, Johnston BC, Mathioudakis AG, Briel M, et al. GRADE guidelines 17: assessing the risk of bias associated with missing participant outcome data in a body of evidence. J Clin Epidemiol. 2017;87:14–22.

    Article  Google Scholar 

  6. 6.

    Sterne JA, Hernán MA, Reeves BC, Savović J, Berkman ND, Viswanathan M, et al. ROBINS-I: a tool for assessing risk of bias in non-randomised studies of interventions. BMJ. 2016;355:i4919.

    Article  Google Scholar 

  7. 7.

    Guyatt GH, Oxman AD, Kunz R, Woodcock J, Brozek J, Helfand M, et al. GRADE guidelines: 8. Rating the quality of evidence--indirectness. J Clin Epidemiol. 2011;64(12):1303–10.

    Article  Google Scholar 

  8. 8.

    Higgins JP, White IR, Wood AM. Imputation methods for missing outcome data in meta-analysis of clinical trials. Clin Trials. 2008;5:225–39.

    Article  Google Scholar 

  9. 9.

    Akl EA, Kahale LA, Agoritsas T, Brignardello-Petersen R, Busse JW, Carrasco-Labra A, et al. Handling trial participants with missing outcome data when conducting a meta-analysis: a systematic survey of proposed approaches. Syst Rev. 2015;4:98.

    Article  Google Scholar 

  10. 10.

    Hewitt CE, Kumaravel B, Dumville JC, Torgerson DJ. Assessing the impact of attrition in randomized controlled trials. J Clin Epidemiol. 2010;63:1264–70.

    Article  Google Scholar 

  11. 11.

    White IR, Higgins JP, Wood A. Allowing for uncertainty due to missing data in meta-analysis—part 1: two-stage methods. Stat Med. 2008;27:711–27.

    Article  Google Scholar 

  12. 12.

    Spineli LM, Pandis N, Salanti G. Reporting and handling missing outcome data in mental health: a systematic review of cochrane systematic reviews and meta-analyses. Res Synth Methods. 2015;6:175–87.

    Article  Google Scholar 

  13. 13.

    Akl EA, Carrasco-Labra A, Brignardello-Petersen R, Neumann I, Johnston BC, Sun X, et al. Reporting, handling and assessing the risk of bias associated with missing participant data in systematic reviews: a methodological survey. BMJ Open. 2015;5:e009368.

    Article  Google Scholar 

  14. 14.

    Akl EA, Johnston BC, Alonso-Coello P, Neumann I, Ebrahim S, Briel M, et al. Addressing dichotomous data for participants excluded from trial analysis: a guide for systematic reviewers. PLoS One. 2013;8:e57132.

    CAS  Article  Google Scholar 

  15. 15.

    White IR, Welton N, Wood A, Ades AE, Higgins JP. Allowing for uncertainty due to missing data in meta-analysis—part 2: hierarchical models. Stat Med. 2008;27:728–45.

    Article  Google Scholar 

  16. 16.

    Ebrahim S, Akl EA, Mustafa RA, Sun X, Walter SD, Heels-Ansdell D, et al. Addressing continuous data for participants excluded from trial analysis: a guide for systematic reviewers. J Clin Epidemiol. 2013;66:1014–21.

    Article  Google Scholar 

  17. 17.

    Gamble C, Hollis S. Uncertainty method improved on best-worst case analysis in a binary meta-analysis. J Clin Epidemiol. 2005;58:579–88.

    Article  Google Scholar 

  18. 18.

    Mavridis D, White IR, Higgins JP, Cipriani A, Salanti G. Allowing for uncertainty due to missing continuous outcome data in pairwise and network meta-analysis. Stat Med. 2015;34:721–41.

    Article  Google Scholar 

  19. 19.

    Dimitrakopoulou V, Efthimiou O, Leucht S, Salanti G. Accounting for uncertainty due to ‘last observation carried forward’ outcome imputation in a meta-analysis model. Stat Med. 2014;34:742–52.

    Article  Google Scholar 

  20. 20.

    Yuan Y, Little RJ. Meta-analysis of studies with missing data. Biometrics. 2009;65:487–96.

    Article  Google Scholar 

  21. 21.

    Turner NL, Dias S, Ades AE, Welton NJ. A Bayesian framework to account for uncertainty due to missing binary outcome data in pairwise meta-analysis. Stat Med. 2015;34:2062–80.

    CAS  Article  Google Scholar 

  22. 22.

    White IR, Higgins JP. Meta-analysis with missing data. Stata J. 2009;9:57–69.

    Google Scholar 

  23. 23.

    Mavridis D, Chaimani A, Efthimiou O, Leucht S, Salanti G. Addressing missing outcome data in meta-analysis. Evid Based Ment Heal. 2014;17:85–9.

    Article  Google Scholar 

  24. 24.

    Salanti G. Indirect and mixed-treatment comparison, network, or multiple-treatments meta-analysis: many names, many benefits, many concerns for the next generation evidence synthesis tool. Res Synth Methods. 2012;3:80–97.

    Article  Google Scholar 

  25. 25.

    Li T, Puhan MA, Vedula SS, Singh S, Dickersin K. Network meta-analysis-highly attractive but more methodological research is needed. BMC Med. 2011;9:79.

    Article  Google Scholar 

  26. 26.

    Spineli LM, Higgins JP, Cipriani A, Leucht S, Salanti G. Evaluating the impact of imputations for missing participant outcome data in a network meta-analysis. Clin Trials. 2013;10:378–88.

    Article  Google Scholar 

  27. 27.

    Spineli LM, Leucht S, Cipriani A, Higgins JP, Salanti G. The impact of trial characteristics on premature discontinuation of antipsychotics in schizophrenia. Eur Neuropsychopharmacol. 2013;23:1010–6.

    CAS  Article  Google Scholar 

  28. 28.

    Zarin W, Veroniki AA, Nincic V, Vafaei A, Reynen E, Motiwala SS, et al. Characteristics and knowledge synthesis approach for 456 network meta-analyses: a scoping review. BMC Med. 2017;15:3.

    Article  Google Scholar 

  29. 29.

    Tan SH, Bujkiewicz S, Sutton A, Dequen P, Cooper N. Presentational approaches used in the UK for reporting evidence synthesis using indirect and mixed treatment comparisons. J Health Serv Res Policy. 2013;18:224–32.

    Article  Google Scholar 

  30. 30.

    Bafeta A, Trinquart L, Seror R, Ravaud P. Analysis of the systematic reviews process in reports of network meta-analyses: methodological systematic review. BMJ. 2013;347:f3675.

    Article  Google Scholar 

  31. 31.

    Nikolakopoulou A, Chaimani A, Veroniki AA, Vasiliadis HS, Schmid CH, Salanti G. Characteristics of networks of interventions: a description of a database of 186 published networks. PLoS One. 2014;9:e86754.

    Article  Google Scholar 

  32. 32.

    Petropoulou M, Nikolakopoulou A, Veroniki AA, Rios P, Vafaei A, Zarin W, et al. Bibliographic study showed improving statistical methodology of network meta-analyses published between 1999 and 2015. J Clin Epidemiol. 2017;82:20–8.

    Article  Google Scholar 

  33. 33.

    Akl EA, Kahale LA, Agarwal A, Al-Matari N, Ebrahim S, Alexander PE, et al. Impact of missing participant data for dichotomous outcomes on pooled effect estimates in systematic reviews: a protocol for a methodological study. Syst Rev. 2014;3:137.

    Article  Google Scholar 

  34. 34.

    Turner RM, Davey J, Clarke MJ, Thompson SG, Higgins JP. Predicting the extent of heterogeneity in meta-analysis, using empirical data from the Cochrane database of systematic reviews. Int J Epidemiol. 2012;41:818–27.

    Article  Google Scholar 

  35. 35.

    Cipriani A, Furukawa TA, Salanti G, Geddes JR, Higgins JP, Churchill R, et al. Comparative efficacy and acceptability of 12 new-generation antidepressants: a multiple-treatments meta-analysis. Lancet. 2009;373:746–58.

    CAS  Article  Google Scholar 

  36. 36.

    Filippini G, Del Giovane C, Vacchi L, D’Amico R, Di Pietrantonj C, Beecher D, et al. Immunomodulators and immunosuppressants for multiple sclerosis: a network meta-analysis. Cochrane Database Syst Rev. 2013;6:CD008933.

    Google Scholar 

  37. 37.

    Delahoy P, Thompson S, Marschner IC. Pregabalin versus gabapentin in partial epilepsy: a meta-analysis of dose-response relationships. BMC Neurol. 2010;10:104.

    Article  Google Scholar 

  38. 38.

    Cui J, Wu B, Liu C, Li Z. A systematic review and adjusted indirect comparison of oral anticoagulants. Orthopedics. 2014;37:763–71.

    Article  Google Scholar 

  39. 39.

    Hollis S, Campbell F. What is meant by intention to treat analysis? Survey of published randomised controlled trials. BMJ. 1999;319:670–4.

    CAS  Article  Google Scholar 

  40. 40.

    White IR, Horton NJ, Carpenter J, Pocock SJ. Strategy for intention to treat analysis in randomised trials with missing outcome data. BMJ. 2011;342:d40.

    Article  Google Scholar 

  41. 41.

    White IR, Carpenter J, Horton NJ. Including all individuals is not enough: lessons for intention-to-treat analysis. Clin Trials. 2012;9:396–407.

    Article  Google Scholar 

Download references

Acknowledgements

The authors would like to thank Konstantinos Pateras and Chrysostomos Kalyvas for their assistance with the article searches.

Funding

This work was supported by the German Research Foundation (Deutsche Forschungsgemeinschaft under grant number SP 1664/1–1) to LMS. The funder had no involvement in study design; in the collection, analysis and interpretation of data; in the writing of the report; and in the decision to submit the article for publication.

Availability of data and materials

The authors declare that all data supporting the findings of this study are available within the article and its additional files.

Author information

Affiliations

Authors

Contributions

LMS conceived the study. LMS and HJS designed the study. LMS and JJNP acquired the data and LMS analyzed the data. LMS drafted the article. JJNP and HJS revised the article critically for important intellectual content. All authors approved the final version of the article.

Corresponding author

Correspondence to Loukia M. Spineli.

Ethics declarations

Ethics approval and consent to participate

Not applicable.

Consent for publication

Not applicable.

Competing interests

The authors declare that they have no competing interests.

Publisher’s Note

Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Additional files

Additional file 1:

Appendix A. List of excluded articles by reason for exclusion. (DOC 223 kb)

Additional file 2:

Appendix B. List of included articles by year of publication. (DOC 254 kb)

Additional file 3:

Appendix C. List of verbatim definitions of missing outcome data. (DOCX 23 kb)

Additional file 4:

Appendix D. Description of intention-to-treat analyses according to reports in the reviews. (DOCX 35 kb)

Additional file 5:

Appendix E. Tools used for the risk of bias assessment in 387 systematic reviews. (DOCX 49 kb)

Additional file 6:

Appendix F. Examples of systematic reviews with explicit, implicit and unclear judgments. (DOCX 57 kb)

Additional file 7:

Appendix G. Explanations to support the strategy applied in primary and sensitivity analysis. (DOCX 22 kb)

Rights and permissions

Open Access This article is distributed under the terms of the Creative Commons Attribution 4.0 International License (http://creativecommons.org/licenses/by/4.0/), which permits unrestricted use, distribution, and reproduction in any medium, provided you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons license, and indicate if changes were made. The Creative Commons Public Domain Dedication waiver (http://creativecommons.org/publicdomain/zero/1.0/) applies to the data made available in this article, unless otherwise stated.

Reprints and Permissions

About this article

Verify currency and authenticity via CrossMark

Cite this article

Spineli, L.M., Yepes-Nuñez, J.J. & Schünemann, H.J. A systematic survey shows that reporting and handling of missing outcome data in networks of interventions is poor. BMC Med Res Methodol 18, 115 (2018). https://doi.org/10.1186/s12874-018-0576-9

Download citation

Keywords

  • Missing outcome data
  • Systematic review
  • Network meta-analysis
  • Intention-to-treat analysis
  • Empirical research