Skip to main content

Assessing transferability in systematic reviews of health economic evaluations – a review of methodological guidance



For assessing cost-effectiveness, Health Technology Assessment (HTA) organisations may use primary economic evaluations (P-HEs) or Systematic Reviews of Health Economic evaluations (SR-HEs). A prerequisite for meaningful results of SR-HEs is that the results from existing P-HEs are transferable to the decision context (e.g, HTA jurisdiction). A particularly pertinent issue is the high variability of costs and resource needs across jurisdictions. Our objective was to review the methods documents of HTA organisations and compare their recommendations on considering transferability in SR-HE.


We systematically hand searched the webpages of 158 HTA organisations for relevant methods documents from 8th January to 31st March 2019. Two independent reviewers performed searches and selected documents according to pre-defined criteria. One reviewer extracted data in standardised and piloted tables and a second reviewer checked them for accuracy. We synthesised data using tabulations and in a narrative way.


We identified 155 potentially relevant documents from 63 HTA organisations. Of these, 7 were included in the synthesis. The included organisations have different aims when preparing a SR-HE (e.g. to determine the need for conducting their own P-HE). The recommendations vary regarding the underlying terminology (e.g. transferability/generalisability), the assessment approaches (e.g. structure), the assessment criteria and the integration in the review process.


Only few HTA organisations address the assessment of transferability in their methodological recommendations for SR-HEs. Transferability considerations are related to different purposes. The assessment concepts and criteria are heterogeneous. Developing standards to consider transferability in SR-HEs is desirable.

Peer Review reports


Economic evaluations play an important role when making pricing and reimbursement decisions on health technologies. To support these decisions Health Technology Assessments (HTAs) often include cost-effectiveness data [1]. These can be based on primary economic evaluations (P-HEs) or Systematic Reviews of Health Economic evaluations (SR-HEs) [2]. P-HEs are time and resource intensive. As not all decision-makers have the ability to commission de-novo analyses, decisions may need to rely on existing similar studies carried out somewhere else. SR-HEs have the potential to meet the increasing demand on incorporating cost-effectiveness considerations [3]. Another possible advantage is the ability to assess consistency between P-HEs, and that a higher confidence may be placed in the results from SR-HEs compared to a single P-HE. That said, a prerequisite to meaningful SR-HEs is that the results from included P-HEs are transferable to the context of interest (e.g. between jurisdictions or over time). But - especially due to concerns regarding the transferability of P-HEs - the usefulness of SR-HEs has been questioned [2, 3]. A P-HE would be considered transferable, if a similar level of cost-effectiveness of the invention could be achieved in the context of interest [4]. However, in the worst case no P-HE can be identified that would be transferable to the own context. In addition, a P-HE specifically developed for the own context will always be better applicable than previously developed P-HEs (suited for different jurisdictions), assuming that data and methodological quality are consistent.

There are many reasons why health economic evaluations might not be transferable between different decision contexts, e.g. regarding methodological aspects [5], geographical settings or time periods [2, 6, 7]. Therefore, considering transferability of the results from included P-HEs is important when conducting a SR-HE [8]. Several tools for assessing transferability of P-HEs have been developed and suggested [6, 9,10,11], but there is no widely agreed approach [12, 13].

Our objective was to review the methods documents of HTA organisations in regard to determine how transferability is recommended to be considered when performing SR-HEs in the context of HTAs. We specifically focussed on methods for assessing transferability in SR-HEs, as methods and positions on the transferability of decision model input data are already addressed elsewhere [6, 14]. Moreover, we focused our analysis on HTA organisations because they prepare recommendations for specific jurisdictions, which makes transferability issues particularly relevant.

This is the second part of a larger research project of our team on this topic. In the first part we reviewed the methodological recommendations of evidence synthesis producing organisations on assessing context suitability (e.g. transferability) of evidence on effectiveness [15].


This review is reported according to the Preferred Reporting Items for Systematic Reviews and Meta-Analysis (PRISMA 2020 Statement) [16], in so far as it is applicable to methodological research.

There was no published protocol for this review. Unless otherwise indicated, we specified all described methods in advance.

Search strategy

We performed structured searches on the webpages of HTA organisations. In a first step, we identified HTA organisations through publicly available member lists of the following HTA umbrella organisations: Health Technology Assessment international (HTAi), International Agency of Health Technology Assessment (INAHTA), European Network of HTA Agencies (EUnetHTA) and Red de Evaluación de Tecnologías en Salud de las Américas (RedETSA). In a second step, two independent reviewers performed structured searches on the webpages of identified HTA organisations. The searches were performed from 8th January to 31st March 2019. To allow for a thorough search on webpages with different and sometimes complex site-structures, we checked each section of the webpages carefully. We used machine based and browser translation tools to identify English or German language documents on foreign language websites. Both reviewers downloaded and stored identified documents, independently. The list of identified documents was compared and synchronised manually. In cases where several versions of one document existed, only the latest version was considered. We removed duplicates manually.

Eligibility criteria and screening

We screened all identified documents against the following pre-defined eligibility criteria:

  1. (i)

    Publication type: Methods documents for the preparation of HTAs (e.g. guidelines, handbooks, manuals, standard operation procedures)

  2. (ii)

    Documents include recommendations for appraising transferability when conducting SR-HEs

  3. (iii)

    The appraisal process is specified, e.g. in form of concrete methods, questionnaires or tools

  4. (iv)

    Recommendations on external validity, generalisability, extrapolation, transferability or applicability are considered (according to the definitions of Burford et al. [4])

  5. (v)

    Languages: English, German

There is no consensus on the terminology of “transferability” and other related terms such as “generalisability” are sometimes used interchangeably [4, 10]. According to Burford et al. we define transferability as to “whether when implementing an intervention in a particular setting or population, the level of [cost]-effectiveness of the intervention (i.e., the effect size) will be similar to that observed in the systematic review.” [4] However, due to heterogeneity we considered all related terms, which are defined in Table 1. Moreover, we decided to reassign the different terms and definitions to the corresponding terminology of Burford et al. [4]. This was an important step when synthesising evidence, to archive a uniform terminology and to distinguish heterogeneity related to terminology from heterogeneity related to other aspects.

Table 1 Terms and definitionsa

Two independent reviewers performed full-text screening. Discrepancies were resolved by discussion and arbitration as necessary.

Data extraction

We performed data extraction using standardised and piloted data extraction sheets. We developed the data extraction sheets inductively, by reviewing included methods documents. We piloted the extraction sheets on a sample of included documents. Final data extraction was performed by one reviewer and checked by a second reviewer. We resolved discrepancies by discussion, and arbitration if necessary. Data were extracted verbatim to avoid interpretation bias.

We collected information on the following issues: the types of considered health technologies, the terminology, the purpose of the SR-HE, the recommended assessment approach and the assessment criteria. All data extraction items are defined in Table 1.

Evidence was synthesised using tabulations and in a narrative way.


Literature search

We identified 158 HTA organisations. Webpage searches resulted in 156 potentially relevant publications from 63 of these organisations. After deduplication, a total of 151 publications remained and were assessed for eligibility. Seven documents from eight organisations were included in our synthesis. Two organisations (Gesundheit Österreich GmbH (GÖG)/Ludwig Boltzmann Institut (LBI)) share the same methods document and are therefore considered together in the following. The selection process is illustrated in Fig. 1. We provide a list of excluded documents as supplement (see Additional file 1).

Fig. 1
figure 1

PRISMA Flow-Chart of document selection

We included methods documents from the following organisations: Agency for Care Effectiveness (ACE) [17], European Network for Health Technology Assessment (EUnetHTA) [18], GÖG/ LBI [19], Health Information and Quality Authority (HIQA) [20], Health Quality Ontario (HQA) [21], National Institute for Health and Care Excellence (NICE) [22], Swedish Agency for Health Technology Assessment and Assessment of Social Services (SBU) [23].

Guidance characteristics

Table 2 provides an overview of the characteristics of included methods documents, including the types of considered health technologies (e.g. non-pharmaceuticals) and the transferability-related terminology. The organisations mainly use the terms transferability, applicability and generalisability with varying and sometimes missing definitions.

Table 2 Guidance characteristics

Review purpose

SR-HEs can have several purposes. Four organisations [17, 21,22,23] use them to determine the need to conduct a P-HE. They consider a SR-HE sufficient if one or more P-HEs without major limitations and sufficient transferability are identified. If deemed necessary, the P-HEs are updated and/or adapted to the decision context. In contrast, GÖG/LBI questions whether a SR-HE can be used to answer questions on cost-effectiveness, because transferability is often limited. That said, GÖG/LBI [19] acknowledges that SR-HEs are useful to explore how and why interventions may be more or less effective, resource or cost intensive and to collect important information for performing a P-HE. EUnetHTA [18] suggests using SR-HEs if a P-HE cannot be conducted (eg, due to limited resources). In this case SR-HEs can be used for explanatory purposes or to identify the most relevant P-HE for informing a particular decision.

Assessment approach concepts

We developed definitions for different aspects of the assessment approach concepts to describe how the transferability assessment is recommended to be operationalised and provide these in Table 1. An overview of the recommended assessment approach concepts can be found in Table 3. Further details are available as supplement (see Additional file 2).

Table 3 Assessment approach concepts

Aspects of transferability can be considered at different steps during the preparation of a SR-HE: At the selection-level, when defining and applying eligibility criteria and at the assessment-level, when assessing the studies quality (and transferability). All included organisations [17,18,19,20,21,22,23] recommend a transferability assessment at the assessment level. Five organisations [17, 19,20,21,22] additionally recommend to consider aspects of transferability at the selection level. Moreover, NICE [22] proposes to limit the study setting to the UK or countries with similar healthcare systems, if necessary.

The assessment approaches vary regarding the target data for assessing transferability, the assessment structure, the provision of guidance for completion, the combination of different assessments (transferability and methodological quality) and the consideration of results derived from sensitivity analyses.

Target data

Regarding the target data for assessing transferability, three organisations [19, 20, 22] consider the transferability of cost and effectiveness data. One organisation [23] considers the transferability of cost data. For three organisations [17, 18, 21] it is unclear, whether their recommendations consider only cost data, or both.

Structuring of the assessment approach

Regarding the assessment structure, six [17, 19,20,21,22,23] concepts can be defined as structured approaches, comprising for example checklists or questionnaires. Two concepts are non-structured approaches, which may include examples for potentially relevant assessment criteria [18, 21].

Guidance on completion

Regarding the provision of guidance in the form of instructions or item descriptions for assessors, four organisations [17, 18, 21, 23] do not provide any instructions or item descriptions, while three organisations [19, 20, 22] provide some. These mainly include explanations and examples of assessment criteria, explanations and examples why something might have an impact on transferability and/or guidance on what should be considered in the assessment.

Combination of different assessments

Regarding the combination of different assessments, six organisations [17,18,19,20,21,22] recommend assessing transferability independently from other aspects (standalone assessment), e.g. the methodological quality while one organisation [23] recommends a combined assessment. Specifically, SBU [23] recommends combining the assessment of transferability and methodological quality, including a joint overall judgement. Although NICE [22], HQA [21] and HIQA [20] recommend a standalone assessment, they link the assessments of transferability and methodological quality: NICE [22] and HQA [21] recommend to assess the methodological quality only if the assessed study is sufficiently transferable, while HIQA [20] recommends to assess transferability only for studies with acceptable quality.

Consideration of results from sensitivity analyses

Four [17, 19, 20, 22] organisations recommend to consider information derived from sensitivity analyses in the context of transferability assessments. For example, according to ACE, “any key drivers of the economic model and areas of uncertainty identified by the sensitivity analysis should be included in the evaluation report.” [17].

Assessment criteria

The assessment criteria describe which factors might affect transferability and are recommended by HTA organisations for consideration in transferability assessments. An overview of all recommended assessment criteria is provided in Table 4. More details can be found as supplement (see Additional file 3).

Table 4 Assessment criteria

PICO (Population, Intervention, Comparison, Outcome)

With exception of SBU [23], all organisations consider population characteristics in their recommended assessments [17,18,19,20,21,22]. Four organisations [18,19,20,21] take demographic characteristics, epidemiology (e.g., incidence/prevalence) and values of health sate preferences into account. EUnetHTA [18] and GÖG/LBI [19] additionally consider life expectancy, compliance and comorbidities. Other items are risk factors [18], severity of disease [18], ethnicity [20] and case mix [19].

Four organisations consider intervention characteristics. Three of them [17, 21, 22] pose the question, of whether the intervention is appropriate for the research question of the SR-HE. SBU [23] recommends to compare the care and type of intervention with the current intervention in the target context. ACE [17] additionally includes a question regarding the similarity between the study’s comparator and the comparator proposed in the research question.

Three organisations consider outcome characteristics [20,21,22]. All of them take health state preferences in terms of QALYs into consideration. Furthermore, HQA [21] and NICE [22] consider the appropriateness of measures. HQA [21] also checks whether estimates of treatment effects are from the best available source.

Health system

Health system characteristics are considered by six [17,18,19,20,21,22] organisations. Three of them pose the question of whether the health system, in which the study was conducted, can be compared to that of the context of interest [17, 21, 22]. Others recommend to examine health system characteristics like available treatment options and unit prices [18], availability of the health technology of interest [19] or incentives to healthcare professionals and institutions [20, 21].

Clinical practice

Six organisations [17,18,19,20,21, 23] consider variation in clinical practice. Some of them specify aspects including provider characteristics [18], treatment practice [20], clinical guidelines [20], care pathways [20], range of treatments [20], organisation of prescribing [19] and provision of health services by different professional groups [19, 23].


Six [17,18,19,20,21, 23] out of seven organisations recommend the consideration of cost data. The following items are suggested: completeness of cost data [17], unit prices/costs [18, 23], absolute and relative prices/costs [19], the medical costing approach [19], relative prices/costs [20, 21] or the value of various costing elements (e.g. charges or fees) [20].

Methodological aspects

All organisations consider methodological aspects of the studies [17,18,19,20,21,22,23], in particular a study’s perspective which determines the costs and consequences that were considered in the analysis [17,18,19,20,21,22,23]. Five organisations [18,19,20,21,22] also consider the appropriateness of discounting. Moreover, HIQA [20] considers the appropriateness of the model used to extrapolate data to the context of interest, while SBU [23] questions whether costs and effects were studied or whether effects were assumed to be equal.


Our review summarises the methodological recommendations of seven HTA organisations for considering transferability in the context of SR-HEs. However, the few hits of our structured search show that this topic still receives little attention in methodological recommendations. In accordance, a review of Mathes et al. [13] identified 13 HTA methods documents addressing SR-HEs, of which only four included recommendations for assessing transferability. According to Luhnen et al. [24], who analysed the methods applied for SR-HE in published HTAs, only 10% of SR-HEs included in HTA reports include an assessment of transferability.

Review purpose

The included organisations have different objectives when performing a SR-HE. For example, GÖG/LBI [19] use them in an exploratory way, while other organisations recourse to identified PH-Es for cost-effectiveness evaluations as far as they do not have any major limitations and sufficient transferability. This may be the case because the different organisations take different stances on the transferability of P-HE: some accept them and consider their limitations, other deem their use generally inappropriate. Not surprising, these general judgements seem to impact the overall approach and related recommendations for considering transferability. That said, the assessment approach concepts and assessment criteria also differ between organisations with similar views on the usefulness of SR-HEs. Some differences might be explained by the varying degree of details/item descriptions provided.

Assessment approach concepts

We identified various concepts for assessing transferability in the context of SR-HEs. It is conspicuous that none of the included HTA organisations recommends a previously published tool without adoptions (only HIQA [20] recommends the use of “any” published tool), suggesting that there is no widely accepted tool. There are several potential reasons for this. For example, some published tools vary in scope and content (e.g. combing methodological study quality, reporting quality, transferability) compared to recommended approaches of HTA organisations [3]. Further, complexity and expenditure of time might limit feasibility of tools for practical application in preparing HTAs. Kim et al. [25] also question the suitability of these tools for local authorities due to their technical and complex nature. Therefore, they developed a decision framework and practical guidance, which might better suit the specific requirements for preparing evidence for local decision making.

The recommendations differ regarding various aspects of their assessment approach concepts (e.g. assessment structuring or target data). All organisations recommend the consideration of transferability in the study assessment. More than half of the organisations also recommend considering aspects of transferability as an eligibility criterion [17, 19,20,21,22]. However, in most cases there is no clear guidance when (e.g. whether studies are directly excluded in the selection process, or post-hoc after performing the study assessment) and which (i.e. what means insufficient) studies should be excluded due to transferability reasons.

Most recommended assessment approach concepts are structured (e.g. recommending a checklist, questionnaire or successive steps for the assessment). Comparing these assessment approach concepts with published tools for assessing transferability (identified through systematic reviews from Goeree et al. [3] and Munthe-Kaas et al. [26]) reveals some differences. For example, Welte et al. [11] and Drummond et al. [6] developed decision charts for assessing the transferability of economic evaluation results, but none of organisations recommends a decision chart. Further, Boulenger et al. [10] developed a checklist, including a score which represents the percentage of checklist items that were adequately or partially addressed in the study, although such a score was not recommended by any HTA organisation.

Regarding the combination of assessment approaches (transferability and methodological study quality), all except one organisations recommend to assess transferability separately from methodological study quality. Nevertheless, three organisations [20,21,22] link the assessments of transferability and methodological quality by making them interdependent, i.e. by only assessing methodological quality for studies with acceptable transferability or vice versa. Comparable recommendations can be found for published tools: According to Späth et al. [27] assessors should judge whether methodological minimum requirements are met before assessing transferability. Welte et al. [11] and Drummond et al. [6] consider insufficient methodological quality as a knock-out criterion. Moreover, Antonanzas et al. [9] recommend to evaluate the methodological study quality initially as a part of the “general transferability index”. Here, a poor rating for several methodological quality aspects can lead to a rating as generally non-transferable. Thus, all these tools put the assessment of methodological study quality first. Interestingly, only one HTA organisation recommends to assess transferability solely for studies with acceptable quality [20], while two organisations recommend the opposite [21, 22], that is to assess methodological study quality solely for studies with sufficient transferability. Furthermore, these approaches are contrary to recommendations for preparing effectiveness reviews. In case of effectiveness reviews there is no acknowledged guidance that recommends the exclusion of studies due to limited study quality [28,29,30].

Assessment criteria

The recommendations include various assessment criteria. Overall, the assessment criteria are heterogeneous and vary in scope and content. We found no hint that differences in terminology would explain this heterogeneity.

We assigned the different assessment criteria (in form of items) to the following domains: population, intervention, comparator, health system, clinical practice, costs and methodological aspects. Comparing these domains and the assigned items between the different organisations shows that the items vary between the organisations, while the domains have a significant overlap. This might be due to the fact that several organisations provide broad assessment criteria, while others include more details. The application of general and unspecific assessment criteria might be explained by the broad scope of the transferability assessment, concrete that the guidance of most HTA organisations refer to all types of health technologies. In this case it is difficult to find a good balance between level of detail and applicability across different types of health technologies. However, unspecific assessment criteria in conjunction with missing instructions leave plenty of scope for user interpretation, which might in turn lead to unsystematic or inconsistent assessments. Only three out of seven organisations provide guidance on completion to support assessors [19, 20, 22].

We further compared the assessment criteria included in the recommended assessment approaches with those of tools from Drummond et al. [6], Welte et al. [11] and Antonanzas et al. [9] and with empirically assessed criteria, identified by the review of Sculpher et al. [7]. The comparison shows heterogeneity both between published tools and regarding the assessment criteria recommended by different HTA organisations. In particular, it is conspicuous that all published tools include the intervention and comparator treatment as assessment criteria, while these are less frequently recommended by the HTA organisations. A possible explanation would be that these aspects are already considered by the HTA organisations when defining and applying eligibility criteria and therefore are not considered in the study assessment. Nevertheless, we identified some assessment criteria that were recommended by the majority of included HTA organisations and in addition by the previously published tools and the empirical evidence. These include demographics, epidemiology, health state preferences, healthcare system (in general), clinical practice (in general), perspective and discount rate.


This review is not without limitations. First, we only considered English and German language methods documents. Second, we did not contact HTA organisations for unpublished documents Thus, we were also not able to clarify unspecific and insufficient descriptions. Third, our literature search was conducted in 2019. However, methodological developments take time and HTAs methods documents are usually updated infrequently. Therefore, it can be assumed that there have been no major changes in this context since then.


Different approaches exist on how to consider transferability of P-HEs, when performing a SR-HE. Some tools have already been published and suggested [3]. However, the included HTA organisations mainly recommend using their own or adapted tools for assessing transferability. There is no commonly used approach/tool for transferability assessments.

The methodological recommendations differ regarding assessment approach concepts and assessment criteria. The structure (e.g, checklist, questionnaire), the step in the SR-HE preparation process, at which transferability should be considered and the link between transferability and methodological study quality also vary between HTA organisations, as well as compared to previously published tools. Differences can also be observed regarding the assessment criteria.

Transferability considerations may depend on the review purpose and should fit the relevant medical area and specific decision contexts. Obviously, generic guidance for all types of health technologies must always allow flexibility to be applicable to the different types of health technologies. A solution might be the development of an assessment tool which comprises a set of core items (assessment criteria which are relevant to most research questions and health technologies) and additional ‘add on’ items (assessment criteria which are only relevant to specific research questions or health technologies, e.g. diagnostics or public health interventions). Regardless of whether the items are core or add-on items, a context-specific formulation, instead of standardised questions would be necessary. Moreover, because of this complexity and inherent heterogeneity, the provision of instructions, explanations and examples appears to be of particular importance. The generated list of assessment criteria provides a comprehensive overview of potentially relevant criteria for assessing transferability. The list might be used as a starting point for determining the relevant items of a transferability assessment tool (e.g, core and add-on items) or when determining the relevant criteria that might affect transferability in a particular decision problem. There are some assessment criteria, which were considered by the majority of included HTA organisations and by the previously published tools and empirical evidence [6, 7, 9, 11]. This suggests that these might have the potential be core items in future assessment tools.

Availability of data and materials

All data generated or analysed during this study are included in this published article and its supplementary information files.



Agency for Care Effectiveness


European Network of HTA Agencies


Gesundheit Österreich GmbH


Health Information and Quality Authority


Health Quality Ontario


Health Technology Assessment


Health Technology Assessment international


International Agency of Health Technology Assessment


Ludwig Boltzmann Institut


National Institute for Health and Care Excellence


Primary Health Economic Evaluation


Population, Intervention, Comparison, Outcome


Preferrred Reporting Items for Systematic Reviews and Meta-Analysis


Quality Adjusted Life Years


Red de Evaluación de Tecnologías en Salud de las Américas


Swedish Agency for Health Technology Assessment and Assessment of Social Services


Systematic Review of Health Economic Evaluations


  1. Garattini L, Casadei G. Health technology assessment: for whom the bell tolls? Eur J Health Econ. 2008;9(4):311–2.

    Article  Google Scholar 

  2. Anderson R. Systematic reviews of economic evaluations: utility or futility? Health Econ. 2010;19(3):350–64.

    Article  Google Scholar 

  3. Goeree R, He J, O'Reilly D, Tarride JE, Xie F, Lim M, et al. Transferability of health technology assessments and economic evaluations: a systematic review of approaches for assessment and application. Clinicoecon Outcomes Res. 2011;3:89–104.

    Article  Google Scholar 

  4. Burford B, Lewin S, Welch V, Rehfuess E, Waters E. Assessing the applicability of findings in systematic reviews of complex interventions can enhance the utility of reviews for decision making. J Clin Epidemiol. 2013;66(11):1251–61.

    Article  Google Scholar 

  5. Jacobsen E, Boyers D, Avenell A. Challenges of systematic reviews of economic evaluations: a review of recent reviews and an obesity case study. Pharmacoeconomics. 2020;38(3):259–67.

    Article  Google Scholar 

  6. Drummond M, Barbieri M, Cook J, Glick HA, Lis J, Malik F, et al. Transferability of economic evaluations across jurisdictions: ISPOR Good Research Practices Task Force report. Value Health. 2009;12(4):409–18.

    Article  Google Scholar 

  7. Sculpher MJ, Pang FS, Manca A, Drummond MF, Golder S, Urdahl H, et al. Generalisability in economic evaluation studies in healthcare: a review and case studies. Health Technol Assess. 2004;8(49):3–4 1-192.

    Article  Google Scholar 

  8. Mandrik O, Severens JL, Bardach A, Ghabri S, Hamel C, Mathes T, et al. Critical appraisal of systematic reviews with costs and cost-effectiveness outcomes: An ISPOR good practices task force report. Value Health. 2021;24(4):463–72.

    Article  Google Scholar 

  9. Antonanzas F, Rodríguez-Ibeas R, Juárez C, Hutter F, Lorente R, Pinillos M. Transferability indices for health economic evaluations: methods and applications. Health Econ. 2009;18(6):629–43.

    Article  Google Scholar 

  10. Boulenger S, Nixon J, Drummond M, Ulmann P, Rice S, de Pouvourville G. Can economic evaluations be made more transferable? Eur J Health Econ. 2005;6(4):334–46.

    Article  Google Scholar 

  11. Welte R, Feenstra T, Jager H, Leidl R. A decision chart for assessing and improving the transferability of economic evaluation results between countries. Pharmacoeconomics. 2004;22(13):857–76.

    Article  Google Scholar 

  12. Wijnen B, Van Mastrigt G, Redekop WK, Majoie H, De Kinderen R, Evers S. How to prepare a systematic review of economic evaluations for informing evidence-based healthcare decisions: data extraction, risk of bias, and transferability (part 3/3). Expert Rev Pharmacoecon Outcomes Res. 2016;16(6):723–32.

    Article  Google Scholar 

  13. Mathes T, Walgenbach M, Antoine SL, Pieper D, Eikermann M. Methods for systematic reviews of health economic evaluations: a systematic review, comparison, and synthesis of method literature. Med Decis Mak. 2014;34(7):826–40.

    Article  Google Scholar 

  14. Barbieri M, Drummond M, Rutten F, Cook J, Glick HA, Lis J, et al. What do international pharmacoeconomic guidelines say about economic data transferability? Value Health. 2010;13(8):1028–37.

    Article  Google Scholar 

  15. Weise A, Büchter R, Pieper D, Mathes T. Assessing context suitability (generalizability, external validity, applicability or transferability) of findings in evidence syntheses in healthcare-An integrative review of methodological guidance. Res Synth Methods. 2020;11(6):760–79.

    Article  Google Scholar 

  16. Page MJ, McKenzie JE, Bossuyt PM, Boutron I, Hoffmann TC, Mulrow CD, et al. The PRISMA 2020 statement: an updated guideline for reporting systematic reviews. BMJ. 2021;372:n71.

    Article  Google Scholar 

  17. Ministry of Health Singapore. Medical technologies evaluation methods and process guide. Singapore; 2018. Available from:

  18. EUnetHTA Joint Action 2. Workpackage 8. HTA Core Model® version 3.0 (Pdf). 2016. Available from:

    Google Scholar 

  19. Gesundheit Österreich GmbH. Methodenhandbuch für Health Technology Assessment Version 1. Vienna; 2012. Available from:

  20. Health Information and Quality Authority. Guidelines for the retrieval and interpretation of economic evaluations of health technologies in Ireland. Dublin; 2014. Available from:

  21. Health Quality Ontario. Health Technology Assessments: Methods and Process Guide. Version 2.0. Toronto (ON): Queen’s Printer for Ontaro; 2018. Available from:

  22. National Institute for Health and Care Excellence. Developing NICE guidelines: the manual (PMG20). 2014. Available from:

  23. Swedish Agency for Health Technology Assessment and Assessment of Social Services. Assessment of methods in healthcare. A handbook. Preliminary Version. 2018. Available from:

  24. Luhnen M, Prediger B, Neugebauer EAM, Mathes T. Systematic reviews of economic evaluations in health technology assessment: a review of characteristics and applied methods. Int J Technol Assess Health Care. 2018;34(6):537–46.

    Article  Google Scholar 

  25. Kim D, Bacon R, Neumann P. Assessing the transferability of economic evaluations: a decision framework. In: Isaranuwatchai W, Archer RA, Teerawattananon Y, Culyer AJ, editors. Non-Communicable Disease Prevention Best Buys, Wasted Buys and Contestable Buys; 2019. p. 91–118.

    Chapter  Google Scholar 

  26. Munthe-Kaas H, Nøkleby H, Nguyen L. Systematic mapping of checklists for assessing transferability. Syst Rev. 2019;8(1):22.

    Article  Google Scholar 

  27. Späth HM, Carrère MO, Fervers B, Philip T. Analysis of the eligibility of published economic evaluations for transfer to a given health care system. Methodological approach and application to the French health care system. Health Policy. 1999;49(3):161–77.

    Article  Google Scholar 

  28. Higgins J, Thomas J, Chandler JE, Cumpston M, Li T, Page M, et al. Cochrane Handbook for Systematic Reviews of Interventions Version 6.0. Cochrane; 2019. Available from:

  29. Centre for Reviews and Dissemination. Systematic Reviews: CRD Guidance for Undertaking Reviews in Health Care: York publishing Services Ltd; 2008. Available from:

  30. Aromataris E, Munn Z. JBI Manual for Evidence Synthesis 2021. Available from:

    Google Scholar 

Download references


Not applicable.


Open Access funding enabled and organized by Projekt DEAL.

Author information

Authors and Affiliations



Study design: AW, DP, TM. Literature Search: AW, RBB. Data extraction: AW, RBB. Data analysis and interpretation: AW, RBB, DP, TM. Writing the first draft of the manuscript: AW. Revisions of the manuscript for important intellectual content: AW, RBB, DP, TM. Final approval of the manuscript: AW, RBB, DP, TM.

Corresponding author

Correspondence to Alina Weise.

Ethics declarations

Ethics approval and consent to participate

Not applicable.

Consent for publication

Not applicable.

Competing interests

The authors declare that they have no competing interests.

Additional information

Publisher’s Note

Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Supplementary Information

Additional file 1.

List of Excluded Documents.

Additional file 2.

Assessment Approach Concepts.

Additional file 3.

Assessment Criteria.

Rights and permissions

Open Access This article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons licence, and indicate if changes were made. The images or other third party material in this article are included in the article's Creative Commons licence, unless indicated otherwise in a credit line to the material. If material is not included in the article's Creative Commons licence and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this licence, visit The Creative Commons Public Domain Dedication waiver ( applies to the data made available in this article, unless otherwise stated in a credit line to the data.

Reprints and permissions

About this article

Check for updates. Verify currency and authenticity via CrossMark

Cite this article

Weise, A., Büchter, R.B., Pieper, D. et al. Assessing transferability in systematic reviews of health economic evaluations – a review of methodological guidance. BMC Med Res Methodol 22, 52 (2022).

Download citation

  • Received:

  • Accepted:

  • Published:

  • DOI: