Skip to main content

Systematically reviewing and synthesizing evidence from conversation analytic and related discursive research to inform healthcare communication practice and policy: an illustrated guide



Healthcare delivery is largely accomplished in and through conversations between people, and healthcare quality and effectiveness depend enormously upon the communication practices employed within these conversations. An important body of evidence about these practices has been generated by conversation analysis and related discourse analytic approaches, but there has been very little systematic reviewing of this evidence.


We developed an approach to reviewing evidence from conversation analytic and related discursive research through the following procedures:

• reviewing existing systematic review methods and our own prior experience of applying these

• clarifying distinctive features of conversation analytic and related discursive work which must be taken into account when reviewing

• holding discussions within a review advisory team that included members with expertise in healthcare research, conversation analytic research, and systematic reviewing

• attempting and then refining procedures through conducting an actual review which examined evidence about how people talk about difficult future issues including illness progression and dying


We produced a step-by-step guide which we describe here in terms of eight stages, and which we illustrate from our ‘Review of Future Talk’. The guide incorporates both established procedures for systematic reviewing, and new techniques designed for working with conversation analytic evidence.


The guide is designed to inform systematic reviews of conversation analytic and related discursive evidence on specific domains and topics. Whilst we designed it for reviews that aim at informing healthcare practice and policy, it is flexible and could be used for reviews with other aims, for instance those aiming to underpin research programmes and projects. We advocate systematically reviewing conversation analytic and related discursive findings using this approach in order to translate them into a form that is credible and useful to healthcare practitioners, educators and policy-makers.

Peer Review reports


Our objective in this paper is to describe a step-by-step guide aimed at those wishing to systematically review conversation analytic and related discourse analytic evidence on relatively specific topics or domains (we are not writing here about the methodology of doing either conversation or discourse analysis). The guide is particularly tailored to reviews where the aim is to inform healthcare practice and policy. Throughout, we draw on a review we conducted to examine evidence about how people talk about sensitive future matters, including illness progression, death and dying [1].

We first provide some background, briefly outlining the rationale and core procedures of systematic reviewing, then providing an overview of the value and methods of conversation analytic and related discursive research. We then present our step-by-step guide. In describing the stages, we examine some distinctive features of conversation analytic and related discursive research which must be addressed when conducting reviews and which mean that established review procedures need combination, alteration, and adaptation for systematic reviews of conversation analytic and related discursive evidence.

Systematic review and synthesis

The rationale and development of systematic review methods have been extensively discussed [26]. In brief, the overall purpose is to sum up best available research evidence in relation to a specific question. The process entails employing recognised and replicable procedures to find, evaluate, and draw together the findings of relevant research. Whilst any reviewer of the literature might well aim to be systematic in their reviewing, the term systematic review is used to mean a specific approach. Compared to traditional (or informal) literature reviews and summaries, systematic reviews aim to be more comprehensive, formalised and transparent, and less dependent upon individual reviewers’ interests – interests which can open traditional summaries to bias [7]. In the field of healthcare research and practice, the findings of this kind of review are seen as more credible than other forms of literature review [8].

Procedures followed in systematically reviewing and synthesizing quantitative evidence are well-established [4]. For qualitative research, methods are rather more diverse and contested [2, 9, 10]. However quantitative and qualitative review approaches share some core procedures. After formulating the review question(s) and scope, extensive searches for evidence are conducted, often with a particular emphasis on electronic databases, using sets of keywords to interrogate these. There follows a progressive sifting of identified publications by applying explicit inclusion and exclusion criteria, then an appraisal of quality using a ranking tool and/or checklist. For publications included in the final set, characteristics of the studies they report - such as the design and participants, and details of the findings - are ‘extracted', that is, summarised and recorded using standardised forms. The final stages involve synthesis of the evidence - comparing and integrating findings, and consulting extensively with interested parties so as to draw conclusions and formulate explicit recommendations [11, 12]. For quantitative research, synthesis, i.e. combining the findings of multiple studies, usually involves applying statistical procedures. For qualitative research, an increasingly diverse range of approaches exists for combining findings of multiple studies [2, 6, 10, 13]; these can be understood as falling into two broad sets of approaches [2]. One set, which has been termed ‘aggregative synthesis’ [9], entails a focus on describing and summarising findings ‘(often in a highly structured and detailed way) and translating the studies into one another.’ ([2], p 8/11). The other set of approaches can be termed interpretive syntheses [9]; these ‘seek to push beyond the original data to a fresh interpretation of the phenomena under review’ ([2], p 8/11), and their ‘primary concern is with the development of concepts and theories that integrate those concepts.’ ([9], p 2/13).

Conversation analysis

The vast majority of healthcare delivery - from diagnosis to decision-making, and from implementing procedures to measuring their effects - is accomplished in and through conversations between people. The quality and effectiveness of healthcare depend enormously upon how people manage these conversations, and thus on the communication practices employed within them [14, 15]. Over the past four decades, and particularly in the last fifteen years, great strides have been taken in scientific understandings of human communication practices and behaviours – particularly those derived from naturalistic observations of large numbers of communication episodes [1618] rather than from experiments or qualitative interview studies. A substantial contribution to this progress has been made through conversation analytic investigations. Despite the name, conversation analysis is applicable in any setting where people interact, including: family conversations [19, 20]; consultations with doctors [18], nurses [21], psychotherapists [22], and physiotherapists [23]; surgical procedures [24, 25]; and interactions in legal [26], mediation [27, 28], and social support settings [29, 30].

Although many conversation analytic studies have collected and analysed data from health and social care settings, the approach has been developed, applied and published largely by those working in the academic domains of linguistics, sociology and social psychology. It is currently less familiar to those working in medical and health services research, amongst whom the term discourse analysis is somewhat more familiar than conversation analysis. Discourse analysis is an ‘umbrella term’ that encompasses a wide range of approaches to analysing texts and talk [31]. In contrast, conversation analysis is a single, specific, defined, and bounded research approach with an established set of perspectives and methods [32]. Some discourse analytic approaches share areas of common ground with conversation analysis [33] and the review methods we describe allow for this kind of discourse analytic work to be incorporated into a review. However, in order to avoid cumbersome wording, hereafter we use only the term conversation analysis.

Conversation analytic studies rely on audio and, increasingly, audio-visual recordings of interactions between people. Recording is planned and conducted so as to minimize the intrusiveness and effects of recording on behaviour [3436], aiming to capture what would go on whether or not the research were in progress [31]. Whilst it is impossible to prove definitively that data captured reflect what would have occurred had recording equipment not been present [37], there are good reasons to assume data is valid in important respects [36]. Recordings are subjected to repeated listening and viewing, and collections of the phenomenon/a of interest are made. For instance, when investigating healthcare, collections might entail episodes where bad news is delivered [38]; where the topic of alcohol or smoking is raised [39]; where patients resist a treatment proposal [40]; or where consultations get brought to a close [41]. Collected episodes are closely scrutinized to generate descriptions of typical and atypical features of communication sequences. These features include: who does what and in what order; what phrases and words are used, and what body movement patterns can be observed. Episodes are transcribed using established conventions [42] which include information about pacing, intonation and overlapping speech, as well as the words used. Analysis draws heavily on previously established findings about communication practices and their functioning [43]. Once practices and patterns of communication have been identified and described in close detail by reference to specific (and often numerous) data sequences, empirical findings are used to generate understandings about the functioning and outcomes of particular practices.

Whilst there have been some literature reviews examining conversation analytic evidence in relation to specific phenomena and domains [44, 45], to the best of our knowledge only one systematic review has been published [7]. This pilot review by Nowak examined German language research on doctor patient talk. Whilst drawing upon a number of approaches to synthesizing qualitative research, Nowak’s review was ‘largely designed in accordance with the research process of the “meta-narrative review”’(p. 430) - a pre-existing off the shelf review approach. Whilst we too draw considerably on existing systematic review procedures, we propose that no pre-existing off the shelf approach is adequate for handling conversation analytic evidence. Thus in the review we conducted, whilst we drew extensively on components of existing review approaches, we also developed new components fitted to the distinctive features of conversation analytic work for which existing quantitative and qualitative review approaches could not provide a solution. Also, Nowak’s systematic review [7] involved generating ‘new theoretical concepts’ (p430, see also p436) within the synthesis phase by using a grounded theory approach. Our approach does not involve use of interpretive processes to develop new theoretical concepts, but entails aggregating findings so as to draw out clinical, policy and/or educational implications.

The significant knowledge conversation analytic studies have generated about verbal and embodied communication practices and their consequences has been little accessed and recognized in healthcare policy, education and practice. This reflects the fact that many studies have been framed in terms of sociological and linguistic concerns, theories and debates, and reported in sociological and linguistic publications. The evidence thus remains largely confined within its parent academic fields. Our paper is motivated by a conviction that this knowledge should no longer remain unavailable to clinical practice and education. Systematically reviewing this kind of evidence is particularly timely because conversation analytic findings are increasingly being used to underpin quantitative evaluation [46], communication training [47], and interventions which have proven effective in enhancing health and social care practice [4850].

Background to the review of future talk

We conducted a review of evidence about how people talk about sensitive and uncertain future matters including illness progression, dying and death. The review protocol can be found at the PROSPERO website [51], an initial summary of findings is reported elsewhere [1], and a more extensive report is in preparation. The work was initiated in a context of growing debate and policies proposing that members of the public [52] and healthcare professionals [53, 54] should talk more than they do about individuals’ death and dying, and that this should lead on to explicit planning for end of life care. At the same time, it is clear that both public [55] and professionals [56] find broaching this topic difficult, and patients and families report very unsatisfactory experiences [57]. Some of the review team knew of conversation analytic studies investigating how people talk about these sensitive topics and documenting the consequences of different ways of talking about them in settings including HIV counselling [58] and oncology clinics [59]. We also knew these had largely been reported in sociology and linguistics publications. We concluded that drawing together evidence in this area would enable us to generate useful, practice-relevant information. We recognised that applying a systematic review approach would enhance the likelihood that findings would be seen as credible by our intended audiences.


In order to develop our approach, we reviewed methodological reports, reviews, and discussions of existing approaches – particularly those about systematic reviews of social scientific research and evidence [2, 3, 9, 10, 12]. We then drafted an outline plan for the proposed steps in our review by drawing upon both this existing literature and the review team’s and advisory group members’ expertise in systematically reviewing quantitative, mixed methods and qualitative research [6064] and in conversation analysis [6567]. We discussed and reached consensus on these proposals with our review advisory group. An iterative process followed in which trying out, reflecting upon, and refining methods for each stage of the review culminated in the guide we present here.

In the following sections, we describe our review approach in terms of eight stages. The approach is tailor-made for working with conversation analytic and related discursive evidence, and we illustrate from our ‘Review of Future Talk’. The stages vary in the degree to which they are based upon and borrow from established and previously reported review practices. For those that are similar, we cite original sources; for those that are dissimilar, we provide detailed explanation, description and some additional files containing various templates. In discussion, we reflect on the challenges and value of systematically reviewing this kind of evidence, and note some possible adaptations and developments of our approach.

Findings: step-by-step guide

Table 1 summarises the eight stages of reviewing. Our proposals should be treated not as rules but as guidelines to be applied flexibly to individual cases. Despite the linear layout of our table and description, in reality the process involves considerable overlap and looping between stages. A note on managing the process: reviews require handling large amounts of data and performing various operations on that data, and may also involve geographically spread teams. Technologies that allow teams to organise the data and communicate efficiently include online reference management software and online file storage. Thus in our review, we maintained a review record in electronic document form. Each reviewer completed and revised sections, and consecutively numbered versions as they added to the record. We shared these and other files via an online file storage programme [68]. Electronic database searches were downloaded to online reference management software [69] which allowed checking for and removal of duplicates, and maintenance of different folders for original searches, and for included and excluded papers. Email discussions, phone conversations, and face to face meetings were also important elements of the process.

Stage 1: Articulate purpose and audiences, then articulate review question and scope

In explicitly articulating the purpose of the review, including its intended audience(s), reviewers build the essential foundations for subsequent deliberations about the review question(s) and scope, and for making decisions about the relevance of individual papers and specific bodies of work. In terms of process, defining purpose and audiences requires reading and deliberation within the review team, and consultations with a range of people with relevant expertise and insights, including practitioners and academics. These consultations comprise face-to-face discussions and circulation and revision of drafts. Only once purpose and audiences are clear should reviewers begin to formulate the review question(s) and scope.

In the Review of Future Talk, deliberation and consultation led to the following definition of the review purpose: “To inform healthcare practice, policy and training with regards providing opportunities for communication about sensitive future matters, including death, dying and planning for end of life”. The phrase ‘with regards providing opportunities for communication about….’ articulates an agnostic stance towards the rights or wrongs of providing such opportunities, and was incorporated as a result of both clinical and conversation analytic perspectives expressed during consultations. The review purpose remained unchanged throughout and provided an anchor point of certainty amidst the sometimes perplexing task of deliberating about whether particular bodies of work and individual publications should be included.

The next step involves articulating the review question(s) and the scope. Defining scope means deciding as precisely as possible which communication practices and tasks, and which conversational participants and settings, will be treated as relevant. This is not easy because communication practices, tasks and activities are not neatly demarcated, and they do not fall into mutually exclusive categories. People generally do more than one thing at the same time through their communication; and any particular communicative task can be attempted and accomplished via multiple practices: think, for instance, of the multiple ways in which one can attempt to ascertain information, including asking direct questions; issuing ‘fishing’ comments; conveying confusion; and raising concerns. (An academic discussion bearing on

Table 1 Stages of systematically reviewing and synthesizing evidence from conversation analytic and related discursive studies

this point can be found in those sections of Levinson’s “Pragmatics” text which examine the ‘Literal Force Hypothesis’ [70]). Furthermore, by their very nature, communication practices and tasks do not carry explicit or self-evident ‘labels’. For these reasons, finalising the questions and defining the scope for a conversation analytic review is a lengthy process. In practice it involves initial searches for and reading of potentially relevant publications, and discussions between reviewers and advisors. This is similar to processes used in established approaches for reviews on complex topics [9, 12].

In the Review of Future Talk, the review questions and scope were redefined and specified with increasing precision over the first six months of the two year project. The resulting primary question was: ‘What evidence exists about how people initiate and pursue talk about sensitive future matters including death, dying and planning for end of life?.’ Defining the scope (see below) required reaching clarity about what would count as ‘sensitive future matters’ for the purposes of the review. Some aspects were clear: studies about talk on future matters that were not directly personal (e.g. talk about global climate change) and studies examining talk about future positive achievements were ruled out by the review’s purpose of informing a particular area of healthcare practice. However, others were less clear: for instance what we meant by ‘sensitive', and whether to include studies that examined people’s talk about the future in relation to currently existing troubles.

In the Review of Future Talk, the final definition of the scope in terms of 'talk about sensitive future matters' was as follows

For the purpose of this review we define talk about sensitive future matters as talk where there is reference to states, events and/or actions:

  • In the domain of individual persons (rather than, e.g. the Earth’s climate)

  • Spanning those that are uncertain to certain, contingent or not

  • That may or will happen in relation to individual persons, and are oriented to - or orientable to - in the specific context as negative or as having potential negative implication(s)

  • That may or will happen some time after the current interactional episode

We include:

  • Studies where talk about future sensitive matters is inherent to the activity examined in the research, and also those where it is adjunctive and occasional

  • Studies of talk about future sensitive matters whether or not talk includes or aims at making plans or decisions about future actions in relation to individuals’ care and lives

We do not include:

  • Studies where analysis examines talk that is exclusively focused on possible future actions in relation to currently existing troubles (as is found in many studies of advice giving)

Stage 2: Specify eligibility criteria

Eligibility criteria specify a priori which kinds of evidence will be included in a systematic review. In quantitative reviews, criteria are generally narrow [4], with only certain study designs eligible for inclusion e.g. randomised controlled trials. Similarly, reviews of conversation analytic evidence should be restricted to studies that rely on detailed inductive analysis of audio- or audio-visually recorded naturalistic interactions. Studies where recorded naturalistic data are analysed solely or primarily using coding frameworks are excluded. Furthermore, to be relevant, studies must include not only interactional data but also analysis that explicitly attends to the topic or phenomenon of interest. Because of the richness and complexity of communication, it is common - and rather frustrating - to find publications where data extracts show participants directly engaging with the matters that are of interest to the reviewers, but where the analytic focus of the publication itself is on other matters. Commonly in systematic reviews, limits are set in terms of how long ago evidence was published. In our view, given the cumulative nature of conversation analytic research, the fact that the term conversation analysis was not used before the 1970s, and the relative stability of communication behaviours, it is logical to include publications from any date in reviews of conversation analytic evidence. Other eligibility criteria should be defined for individual reviews; considerations should include: whether or not to exclude studies outside healthcare; whether to restrict to studies analysing data from only one language; and whether to include unpublished studies such as graduate theses.

In the Review of Future Talk, we included studies of talk about future sensitive matters whether the setting was formal and institutional (e.g. health or social care episodes) or informal (e.g. friend and family conversations). This decision was consistent with the conversation analytic view that practices used in institutional interactions are grounded in, rather than distinct from, everyday communication practices [71]. We did, however, exclude studies of large-group interactions, such as classrooms, as these are so different to healthcare consultations which usually involve just two or three people. We excluded studies where data involved languages other than English because of the possibility that different languages might entail significantly different practices for talking about the future, and/or different consequences of practices. We did, however, keep copies and notes on non-English studies that we identified. This allowed us to make preliminary observations about whether practices identified in the main review had been identified in other languages. We included only studies published in peer-reviewed journals or published books, and excluded conference presentations and graduate theses. In so doing, we treated the peer review process as a form of quality control upon the publications included in our review (although we acknowledge that, like any quality control, peer review is not without flaws).

Stage 3: Search for studies

3a) Identify potential sources of publications, search sources

As noted above, many conversation analytic investigations relevant to healthcare are published outside clinical journals and in disparate fields including linguistics and sociology. For this reason, diverse sources need to be searched. Doing so is established practice in systematic reviews of complex interventions and those where social science literature is examined [3, 72]. Therefore, besides interrogating electronic databases using standardised sets of search terms, other sources are used. These include the review team’s existing knowledge, and knowledge amongst the conversation analytic and academic healthcare community accessed via personal contacts, forums such as electronic discussion lists, and online bibliographies. ‘Snowball sampling’ – i.e. citation tracking and reference searching of publications identified through these various means should also be used. With regards formulating search terms for use with electronic databases, the services of a librarian/information specialist should be sought if possible ([4], Section 6.3.1).

In the Review of Future Talk, we tested and refined sets of ‘word groups’ in order to maximize sensitivity and specificity of the electronic database searching. Terms that we found most useful in identifying studies that applied conversation analytic and related discursive methodologies to our substantive topic were: (Group 1) communicat* OR interact* AND (Group 2) audio* OR video* OR discourse-analysis OR conversation-analysis OR sequential-analysis OR linguistic*. Details of all the word groups we used can be found in Additional file 1. We also searched for publications from sources including our own Endnote databases, the bibliography section of the ‘Ethno/CA News’ website [73], and an enquiry to the ‘Languse’ internet discussion list [74]. Once we had identified papers from these sources, we searched for potentially relevant papers amongst citations of these using the ‘Google Scholar’ database [75]. At the time of the review, we did not have the resources to call upon the services of an information specialist. Whilst this may have resulted in a less than optimal search strategy, we believe it did not have a huge impact on our review because, as we explain below, five out of the 18 papers we eventually included were not listed in any of the multiple electronic databases we searched.

3b) Scan identified publications and make final selection for inclusion

Each round of searching usually identifies a relatively large number of publications compared to the number finally included. Identified publications need sifting to decide which fall within the review’s scope and eligibility criteria. This can usually be judged merely by examining title and abstract, and whilst established guidance states that it is desirable for two reviewers to do so ([4] Section 7.2.4), for reasons of practicality it is not uncommon for just one reviewer to perform this initial sifting [76]. Where decisions cannot be made from title and abstract alone, the full paper must be obtained and the data extracts and analysis sections read closely. At this point, for the sake of reliability, it is ideal practice for two reviewers to undertake reading and judgements separately. Even after closely reading extracts and analysis, there are often boundary cases for which decisions about inclusion are not straightforward. After these have been read by at least two members of the team, they should be discussed in order to reach reasoned consensus decisions about inclusion. Where a publication has been read and excluded, notes should be kept on the decision made and the reasoning behind the exclusion as this helps later report writing, and expedites any process of revisiting or even revising decisions.

In the Review of Future Talk, we identified over 2000 publications through our broad search strategy. Eighteen publications were included in the final review. We opted to search nine different electronic databases (ISI Web of Science, Amed, Embase, CINAHL, Medline, PsycINFO, ASSIA, Sociological Abstracts CSA, Google Scholar) because we were interested in whether any would stand out as particularly useful or not for conversation analytic publications. The least useful databases for us in terms of the proportion of publications identified to those actually included were: (a) PsycINFO where searching identified 844 publications, only three of which were finally included and two of those were also found in other databases; (b) Sociological Abstracts where searching identified 284 papers, none of which were included in the final review. We found the ISI, Embase and Medline databases produced fewer ‘false positives’ - each yielding fewer than 160 ‘hits’; three publications which were found in these databases and not found from any other source were included in the final review. Notably, ten of the finally included papers were not identified in any of our electronic searches. After completing our review, we checked back and found that five of these ten were listed in the electronic databases, but had not been identified in our searches, and that the other five were not listed in any of the databases.

Of the final 18 publications we included, four were found exclusively from electronic database searching, 10 were found through reviewing our existing knowledge, one was a serendipitous find, and the other three were each identified twice – both in the databases and via our existing knowledge. Eight of the 18 were listed in the 2011 version of the specialist ‘EMCA news’ specialist bibliography [73]. In our discussion, we consider the pros and cons of searching various sources, particularly electronic databases, for this kind of review.

In the searching and sifting stage, we found 15 publications for which it was not possible to make definitive inclusion or exclusion decisions without detailed reading. Each of the two main reviewers read and then discussed them in order to reach consensus decisions. Five of these 15 were included in the final 18. Our discussions about these ‘boundary cases’ and our notes on reasons for exclusions and inclusions were important in reaching a final version of the review’s scope.

Stage 4: Describe characteristics of - rather than appraise - included studies

Existing techniques, guidance and discussion about appraising the quality of quantitative ([4], Chapter 8) and qualitative [77, 78] research have very limited application to conversation analytic research for the important reason that conversation analytic perspectives, methods and findings are incompatible with the binary categories - qualitative and quantitative - familiar in healthcare research [43]. The primary data and findings of conversation analyses are not numerical and statistical (although studies increasingly include tabulations and descriptive statistics as part of their findings [71, 79, 80]) so conversation analytic work does not fall within the scope of quantitative healthcare-related research. The conversation analytic approach is also incompatible with conventional understandings of qualitative enquiry as entailing investigating meanings, views and understandings via interpretive analysis, most commonly using transcribed interview data [8183]. In conversation analysis, the main data always comprise directly recorded interactions rather than qualitative interviews, and conversation analysts explicitly and strictly avoid using data to impute psychological states, perceptions and motivations [84, 85]. Conversation analysis produces systematic and empirically grounded descriptions of concrete practices and their interactional consequences and functioning, it does not involve the kind of interpretation and theory generation that characterise in qualitative healthcare research [84]. These distinctive features mean that no existing tools for quality appraisal of research are suitable.

So, what can be done in terms of characterising the contribution made by each conversation analytic study included in a systematic review? Rather than reaching a single assessment of each study’s quality, or ranking studies, two broad dimensions must be considered in relation to each study’s value and contribution: (1) the type and amount of data, and (2) the detail and depth of analysis. These two cannot be collapsed into a meaningful, single, quality assessment. The type and amount of data: conversation analytic studies vary with regards whether audio or audio-visual recordings are used, what amount of data is analysed, how many settings and participants are involved, and how diverse is the range of settings and participants. It is inappropriate to assume that more data is better: studies that document practices in substantial detail regularly involve quite small datasets. Also, one type of conversation analytic research involves bringing the cumulative findings of past work about the use and functioning of interactional practices to bear upon single episodes of interaction [86]. On the other hand, some studies examining larger datasets examine practices in less detail, but may significantly contribute to reviews by providing evidence about how widespread a practice is, its frequency of use within settings, and by showing recurrent patterns in its consequences such as the kinds of response it prompts from patients. The detail and depth of analysis: studies vary greatly in the detail and depth to which they analyse particular interactional practices or phenomena. Variations include the degree to which they examine when in their interactions (and communication sequences) people use some particular practice(s); and whether or not they examine only the words used or other important language features such as grammar, pauses, and intonation. They also vary in the degree to which they investigate the consequences and/or social functioning of the practice. If analysis examines few of these features, this is not necessarily equivalent to lower quality – studies often explicitly set out to examine restricted aspects of a phenomenon, practice or domain in great depth. Studies also vary in terms of the extent to which analysis is grounded in previous empirical findings. Again, it is not logical to treat this as a simple matter of analytic quality, because it is impossible for earlier studies to refer to later findings.

Thus, reviewers should not claim that studies with more analytic detail, depth and grounding provide stronger evidence, nor that studies documenting the practices or phenomena among more numerous or diverse people and settings provide stronger evidence. Rather, studies documenting a practice ‘more widely’ contribute one type of evidence, while those documenting a practice ‘more deeply’ contribute another type of evidence. Rather than applying conventional quality appraisal tools, conversation analytic reviewers must collect and present information on several dimensions of the studies. Any proposals concerning the strength of evidence about particular practices or phenomena need to be described and justified in terms of these various dimensions. Reviewers need to record these details using a customised template designed to capture characteristics of each study, its dataset, and its analysis.

The characteristics we recorded for studies in the Review of Future Talk are listed in the subsection below. Additional file 2 provides a formatted version of the template we used.

Characteristics recorded for studies in the Review of Future Talk

Data characteristics:

  • Size of overall dataset in minutes / hours, and number of interactions

  • Number of episodes from the overall dataset upon which analysis relies

  • Number of episodes from the collection that appear in the publication

  • Number and description of sites

  • Number and description of institutional contexts (e.g. hospital ward, outpatient clinic, family conversation)

  • Whether practice(s) is/are observed in more than one individual/dyad

  • Whether practice(s) is/are observed in more than one group (e.g. do both doctors and patients use it)

Analysis characteristics: Does analysis:

  • Predominantly examine more than only one party’s turns; i.e. attend to sequence?

  • Examine data in fine-grained detail?

  • Examine more than just the topical/semantic content; i.e. does it attend to aspects of grammatical, pragmatic, and/or prosodic content?

  • Include examination of aspects of the sequential environment in which practice(s) occur(s)?

  • Include examination of aspects of turn and/or sequence design?

  • Include examination of interactional effects and consequences?

  • Include examination of atypical cases?

  • Support central analytic claims by direct references to data/extracts?

  • Use established analytic findings as ‘tools’ in the analysis?

Stage 5: Data extraction

Data extraction is the term conventionally used in systematic reviewing for the work of recording findings, claims, and data from each included publication [3, 4]. Besides recording findings, ‘facesheet’ data are gathered – these document basic details such as study title, date of publication, authors, and journal or book title. For recording findings, reviews of conversation analytic evidence require customised templates to collect information on the kinds of phenomena and analytic dimensions that conversation analytic studies report, and they also need to handle the fact that studies often describe more than one practice or phenomenon. Development of templates should include blind testing on a diversity sample of the included papers, with two or more reviewers completing templates for several papers independently, and then comparing results. This allows ambiguities and missing elements to be identified and then resolved in subsequent drafts. The subsection below lists the analytic dimensions recorded for each paper in the Review of Future Talk. A version of our data extraction template can be found in Additional file 3. Another set of data is also collected: original data extracts from each publication. Where only part of a publication’s findings are relevant to the review, only the associated data extracts are collected. It is worth noting that extracts comprise edited sections of transcripts rather than the original data, so cannot be used for the purpose of going beyond existing evidence to build new findings; rather they are used to support the collation and synthesis of the findings, and for illustrating reports.

Data extraction categories used in the Review of Future Talk

  • Phenomenon (in brief)

  • Phenomenon in author’s own words

  • Research question for this finding (if applicable)

  • Number of episodes pertaining to this finding

  • Archetypal sequence

  • Features of the talk in which the phenomenon is produced – i.e. aspects of the sequential/interactional context in which it arises

  • What are the implications of these environmental features?

  • Sequence and/or turn design features of the phenomenon

  • What are the interactional effects of these design features?

  • In sum, what is the overarching function of the phenomenon?

  • Author-proposed implications

  • Any other implications

  • Reviewer’s notes

Stage 6: Collate and synthesise data

At this analytic stage the studies are organized into logical categories [3]. There is no definitive or ‘correct’ organisation, rather the process must be driven by the review’s purpose and questions. An obvious way to organise conversation analytic evidence is in terms of the practices documented; other options include organizing according to study setting or the kinds of participants recorded. Next, findings are analysed within each category so as to combine understandings about particular practices. Tables summarising the characteristics and findings of the included studies are compiled.

Synthesis begins with an overall description of the amount of information uncovered through the review [3]. Findings are then collated and summarised using an aggregative approach – drawing together findings in ways that involve describing, summarising and what has been termed ‘translating into’ each other [2]. At this stage, reviewers of conversation analytic evidence may draw upon, and indeed systematically search for, other literature in order to expand the insights provided by the review. Doing so is established practice in review approaches for evaluating complex interventions [12]. Synthesis also involves identifying gaps in the evidence. This requires understanding what is actually done and required in practice, and comparing this with those communication practices and actions that have been investigated and documented by research. Synthesis culminates in generation of evidence-based, reasoned proposals about the implications for the review’s audience(s).

A specific and distinctive feature of conversation analytic findings must be dealt with when synthesizing the evidence and formulating explicit recommendations: conversation analyses find consistently that there are always multiple ways to perform any communicative activity, with each way having an array of advantages and disadvantages for any individual context and endeavour. It is therefore not possible to produce conclusions such as: ‘Practice X should be used, and practice Y should not’; or ‘X works, and Y does not’. Such blanket recommendations are incompatible with a scientific approach that takes seriously the complexity of human communication and the way that communication practices are always fitted to individual contexts and interlocutors. Reasoned proposals or implications generated in reviews of conversation analytic evidence thus need to take a form similar to that recommended within an existing approach called explanatory review (sometimes known as realist synthesis) [12], along the lines of: “In circumstances such as A, try practice B, or when using C, watch out for D” ([12], p S1:24). The proposals should be tested, extended and refined by reporting them to and discussing them with relevant audiences and experts before they are finalised.

In the Review of Future Talk, we organised the findings in terms of practices. These included: ‘agenda setting questions’; ‘use of hypotheticals’; ‘allusive, vague, or euphemistic talk’; and ‘features other than words that display sensitivity’. We then considered two alternative ways of ordering these categories for reporting. One was to order them according to how much evidence there was about each – in terms of both analytic detail and depth, and ‘quantity/diversity’ of data, participants and settings in which the practice had been observed. The alternative was to mirror the interactional sequences we were interested in, that is, to start with practices used in attempting to initiate talk about sensitive future matters, then report on those used in pursuing such talk, and finally those that closed talk about these matters. Given that the primary purpose of our review was to provide practice, education and policy relevant information (rather than, for instance, to set a research agenda), we decided this latter ordering would be the most helpful for our intended audiences. As we moved to synthesis, we drew on seminal and recent studies and reviews in order to strengthen findings and extend the usefulness of the review. Specifically, we used these to add information about how and why particular practices had particular effects on encouraging or discouraging talk about future sensitive matters. For instance, work on how questions function within medical interactions [44] offered additional insights into the mechanisms through which talk about future troubles is encouraged by the question-asking practices identified in our review. However, as noted, we did not aim to generate new findings or theories on the basis of our review or through re-analysis of data extracts; in this sense, the synthesis approach we used was aggregative rather than interpretive.

Stage 7: Sensitivity and subgroup analyses

Conventionally, sensitivity analysis involves assessing post hoc the effects of including or excluding particular findings [9]. Subgroup analyses examine whether the findings vary in relation to particular characteristics of included studies (or their participants) [87]. Subgroup analyses can be used to examine the effects of including studies yielded from particular sources, and so inform design of search strategies for subsequent reviews.

As noted, only three publications included in the Review of Future Talk were identified exclusively through electronic database searching. We decided to perform a subgroup analysis examining the usefulness of particular sources for finding publications in this kind of review. We examined the three publications that had been identified exclusively through electronic database searching – those that had not been found via any source. We found that one of these documented a practice about which more extensive evidence was provided by other publications, and that the other two involved analysis that examined wording but not other important features of communication, and which focused largely on clinicians’ talk rather than including examination of patients’ responses. On the other hand, all three of these studies documented the practices across relatively large numbers of cases. The value to the review of these three studies was in adding to the extent of evidence, rather than adding details about practices’ structure and functioning. We concluded that in our particular review we would have drawn the same conclusions in terms of implications had we not included these three publications (i.e. had we not searched electronic databases), but that on the other hand, these three strengthened the credibility of the review because they contributed evidence that the identified practices are widespread in their use.

Stage 8: Reporting the review

Reviews should be reported in a form that is accessible, useful and credible to the audiences for whom it is designed. Consulting with potential users and asking them for comments on draft versions is thus an important element of reporting. The kinds of applied and clinical journal to which reports are likely to be submitted often impose strict word length limitations. It is thus useful to make additional information such as reproductions of extracts from included papers available e.g. through an extended version published on the journal’s website. Most if not all clinical journals will expect quantitative systematic reviews to be reported in a format laid out in the ‘PRISMA statement’ – well established guidance for reporting systematic reviews that evaluate healthcare interventions [87]. Although this guidance is specific to one type of quantitative systematic review [88], it is advisable to use some of its features in reporting reviews of conversation analytic evidence - including using flow diagrams to set out information about numbers of publications initially found, numbers sifted out, and numbers finally included. Reports should also include tables summarising as briefly as is feasible the characteristics of each study, and studies’ findings and claims. Specific guidance on reporting reviews other than the type covered by the PRISMA statement is emerging (e.g. [88]) and should be consulted in writing journal reports.

We presented and discussed emerging findings from our Review of Future Talk with several clinician and educator audiences including people with different professional backgrounds, more and less experience, and working in different settings. We also held discussions with individual colleagues before attempting to write a report for publication. This proved useful in alerting us to which elements seemed of most interest to our audiences, and to the kinds of concerns they might express. For instance, some clinicians reacted to our proposals with concerns and questions about the extent to which changing their communication behaviours would constitute manipulating patients, and we thus chose to address this concern within written manuscripts. On the other hand, our verbal presentations about findings reassured us that although most clinicians saw themselves as already having tacit knowledge of the practices we described, they found it useful rather than patronizing to have this knowledge made explicit. Presentations and discussions also gave us the opportunity to seek advice from clinical and education colleagues about our proposed implications. We were careful to make it clear that these implications were extrapolations from the data – and thus needed to be treated more tentatively than the empirical findings of the original studies.


We have offered an eight-stage guide tailor-made for conducting systematic reviews of conversation analytic evidence, shaped particularly for reviews that aim to provide useful information to professionals, policy makers, and educators. The process supports systematic location, collation and examination of evidence derived from conversation analytic and related discursive work. It could be adapted for reviews with other aims, such as literature reviews where the aim is inform research agendas, or to underpin doctoral theses. A possible adaptation of the approach would be for reviews to include re-analysis or secondary analysis of original data, along the lines of meta-analysis in quantitative reviews. Whilst theoretically this could be a very fruitful means of extending conversation analytic knowledge, there are practical problems particularly in relation to evidence about healthcare because of the restrictions that usually (and appropriately) exist in terms of sharing data.

Systematic review work is time consuming and laborious. Systematically searching for evidence – particularly via electronic databases – produces a low yield of eligible publications relative to the time and effort involved. Despite the limited contribution of publications identified this way in our own review, we nevertheless argue that it is worth spending the time required for two reasons. Firstly, this kind of process both counteracts the natural tendency for reviewers to focus only upon publications of which they are already aware and helps ensure searching of the breadth of academic fields in which conversation analytic work is published. Secondly, the systematic review process has an established reputation and credibility in applied academic fields such as healthcare [4], education [6] and social care [89]. This means that using a systematic review approach is likely to maximize the chances of conversation analytic evidence making its way into applied fields, and of being seen as credible amongst those who wield influence in the fields of healthcare practice, policy and education. It is also worth noting that although in our particular review electronic searching yielded relatively little evidence, this would not necessarily be the case in reviews of conversation analytic evidence on other topics.

Conversation analytic and related discursive studies have generated a significant, substantial and cumulative body of knowledge about healthcare communication. This knowledge is little accessed by practitioners, educators and policy makers. Systematically reviewing evidence from this form of study offers the prospect of making useful knowledge available to practitioners, educators and policy makers in a credible form. However, there are distinctive challenges in reviewing this kind of evidence. These can be managed by applying the approach to reviewing which we have presented here.


  1. Parry R, Land V, Seymour J: Communicating face-to-face about sensitive future matters including end of life: a systematic review of evidence from fine-grained observational research [abstract]. Palliat Med. 2012, 26: 539-

    Google Scholar 

  2. Barnett-Page E, Thomas J: Methods for the synthesis of qualitative research: a critical review. BMC Med Res Methodol. 2009, 9: 59-10.1186/1471-2288-9-59.

    Article  PubMed  PubMed Central  Google Scholar 

  3. Petticrew M, Roberts H: Systematic Reviews in the Social Sciences: A Practical Guide. 2006, New York: John Wiley

    Book  Google Scholar 

  4. Higgins J, Green S: Cochrane Handbook for Systematic Reviews of Interventions Version 5.1.0: The Cochrane Collaboration. 2011,,

    Google Scholar 

  5. Centre for Reviews and Dissemination University of York: Systematic Reviews: CRD’s guidance for undertaking reviews in health care. 2009, York: CRD, []

    Google Scholar 

  6. An Introduction to Systematic Reviews. Edited by: Gough D, Oliver S, Thomas J. 2012, London: Sage

  7. Nowak P: Synthesis of qualitative linguistic research—A pilot review integrating and generalizing findings on doctor–patient interaction. Patient Educ Couns. 2011, 82: 429-441. 10.1016/j.pec.2011.01.022.

    Article  PubMed  Google Scholar 

  8. Briner R, Denyer D: Systematic review and evidence synthesis as a practice and scholarship tool. The Oxford Handbook of Evidence-based Management. Edited by: Rousseau D. 2012, New York: Oxford University Press

    Google Scholar 

  9. Dixon-Woods M, Cavers D, Agarwal S, Annandale E, Arthur A, Harvey J, Hsu R, Katbamna S, Olsen R, Smith L: Conducting a critical interpretive synthesis of the literature on access to healthcare by vulnerable groups. BMC Med Res Methodol. 2006, 6: 35-10.1186/1471-2288-6-35.

    Article  PubMed  PubMed Central  Google Scholar 

  10. Ring N, Ritchie K, Mandava L, Jepson R: A guide to synthesising qualitative research for researchers undertaking health technology assessments and systematic reviews. 2010, NHS Quality Improvement Scotland, []

  11. Tranfield D, Denyer D, Smart P: Towards a methodology for developing evidence-informed management knowledge by means of systematic review. Br J Manag. 2003, 14: 407-222.

    Article  Google Scholar 

  12. Pawson R, Greenhalgh T, Harvey G, Walshe K: Realist review - a new method of systematic review designed for complex policy interventions. J Health Serv Res Policy. 2005, 10 (Suppl 1): 21-34. 10.1258/1355819054308530.

    Article  PubMed  Google Scholar 

  13. Kastner M, Tricco AC, Soobiah C, Lillie E, Perrier L, Horsley T, Welch V, Cogo E, Antony J, Straus SE: What is the most appropriate knowledge synthesis method to conduct a review? Protocol for a scoping review. BMC Med Res Methodol. 2012, 12 (1): 114-10.1186/1471-2288-12-114.

    Article  PubMed  PubMed Central  Google Scholar 

  14. Haskard Zolnierek K, DiMatteo M: Physician communication and patient adherence to treatment: a meta-analysis. Medical Care. 2009, 47 (8): 826-834. 10.1097/MLR.0b013e31819a5acc.

    Article  PubMed Central  Google Scholar 

  15. Stewart M: Effective physician-patient communication and health outcomes: a review. Can Med Assoc J. 1995, 152 (9): 1423-1433.

    CAS  Google Scholar 

  16. Stivers T, Mondada L, Steensig J: The Morality of Knowledge in Conversation. 2011, New York: Cambridge University Press

    Book  Google Scholar 

  17. Tomasello M: The New Psychology of Language: Cognitive and Functional Approaches To Language Structure. 2003, Mahwah NJ: Lawrence Erlbaum Associates

    Google Scholar 

  18. Heritage J, Maynard D: Communication in Medical Care: Interaction between primary care physicians and patients. 2006, Cambridge: Cambridge University Press

    Book  Google Scholar 

  19. Beach W: Between Dad and Son: initiating, delivering, and assimilating bad cancer news. Health Commun. 2002, 14 (3): 271-298. 10.1207/S15327027HC1403_1.

    Article  PubMed  Google Scholar 

  20. Sterponi L: Account episodes in family discourse: the making of morality in everyday interaction. Discourse Studies. 2003, 5 (1): 79-100.

    Article  Google Scholar 

  21. Jones A: Creating history: documents and patient participation in nurse-patient interviews. Sociol Health Illn. 2009, 31 (6): 907-923. 10.1111/j.1467-9566.2009.01190.x.

    Article  PubMed  Google Scholar 

  22. Peräkylä A, Antaki C, Vehviläinen S, Leudar I: Conversation Analysis and Psychotherapy. 2008, Cambridge: Cambridge University Press

    Book  Google Scholar 

  23. Parry R: Practitioners’ accounts for treatment actions and recommendations in physiotherapy: when do they occur, how are they structured, what do they do?. Sociol Health Illn. 2009, 31 (6): 835-853. 10.1111/j.1467-9566.2009.01187.x.

    Article  PubMed  Google Scholar 

  24. Koschmann T, LeBaron C, Goodwin C, Zemel A, Dunnington G: Formulating the triangle of doom. Gesture. 2007, 7 (1): 97-122. 10.1075/gest.7.1.06kos.

    Article  Google Scholar 

  25. Hindmarsh J, Pilnick A: Knowing bodies: embodiment, teamwork and anaesthetic practice. Organ Stud. 2005, 28 (9): 1395-1416.

    Article  Google Scholar 

  26. Drew P: Contested evidence in courtroom cross-examination: The case of a trial for rape. Talk at Work: Interaction in Instititional Settings. Edited by: Drew P, Heritage J. 1992, Cambridge: Cambridge University Press, 470-520.

    Google Scholar 

  27. Garcia A: Dispute resolution without disputing: How the interactional organization of mediation hearings minimizes argument. Am Sociol Rev. 1991, 56 (6): 818-835. 10.2307/2096258.

    Article  Google Scholar 

  28. Greatbatch D, Dingwall R: Talk and identity in divorce mediation. Identities in Talk. Edited by: Antaki C, Widdecombe S. 1998, London: Sage, 121-132.

    Google Scholar 

  29. Hepburn A, Potter J: Designing the recipient: managing advice resistance in institutional settings. Soc Psychol Q. 2011, 74 (2): 216-241. 10.1177/0190272511408055.

    Article  Google Scholar 

  30. Stokoe E: Overcoming barriers to mediation in intake calls to services: Research-based strategies for mediators. Mediation Digest. 2012, []

    Google Scholar 

  31. Potter J: Discourse analysis. Handbook of Data Analysis. Edited by: Hardy M, Bryman A. 2003, London: Sage, 607-624.

    Google Scholar 

  32. Sidnell J, Stivers T: The Handbook of Conversation Analysis. 2013, Malden MA: Wiley-Blackwell

    Google Scholar 

  33. Edwards D, Potter J: Discursive psychology, mental states and descriptions. Conversation and Cognition. Edited by: te Molder H, Potter J. 2005, Cambridge: Cambridge University Press, 241-259.

    Chapter  Google Scholar 

  34. Heath C: Embodied action: video and the analysis of social interaction. Qualitative Research. Edited by: Silverman D. 2011, London: Sage, 250-270. 3

    Google Scholar 

  35. Peräkylä A: Validity in research on naturally occuring social interaction. Qualitative Research. Edited by: Silverman D. 2011, London: Sage, 365-382. 3

    Google Scholar 

  36. Parry R: Video-based conversation analysis. Sage Handbook of Qualitative Methods in Health Research. Edited by: Bourgeault I, DeVries R, Dingwall R. 2010, London: Sage, 373-396.

    Chapter  Google Scholar 

  37. ten Have P: Doing Conversation Analysis: A Practical Guide. 1999, London: Sage, 47-74. 4

    Google Scholar 

  38. Maynard DW: Bearing bad news in clinical settings. Progress in Communication Sciences. Edited by: Dervin B. 1991, New Jersey: Ablex, 143-172.

    Google Scholar 

  39. Sorjonen M, Raevaara L, Haakana M, Tammi T, Peräkylä A: Lifestyle discussions in medical interviews. Communication in Medical Care: Interaction between primary care physicians and patients. Edited by: Heritage J, Maynard D. 2006, Cambridge: Cambridge University Press, 340-378.

    Chapter  Google Scholar 

  40. Gill VT, Pomerantz A, Denvir P: Pre‒emptive resistance: patients’ participation in diagnostic sense‒making activities. Sociol Health Illn. 2010, 32 (1): 1-20. 10.1111/j.1467-9566.2009.01208.x.

    Article  PubMed  Google Scholar 

  41. West C: Coordinating closings in primary care visits: producing continuity of care. Studies in Interactional Sociolinguistics. 2006, 20: 379-415.

    Google Scholar 

  42. Jefferson G: Glossary of transcript symbols with an introduction. Conversation Analysis: Studies from the First Generation. Edited by: Lerner GH. 2004, Amsterdam: John Benjamins Publishing Company, 13-31.

    Chapter  Google Scholar 

  43. Clayman SE, Gill VT: Conversation analysis. Handbook of Data Analysis. Edited by: Hardy M, Bryman A. 2004, Beverly Hills: Sage, 589-606.

    Google Scholar 

  44. Heritage J: Questioning in medicine. ‘Why do you ask?’: The Function of Questions in Institutional Discourse. Edited by: Freed A, Ehrlich S. 2009, New York: Oxford University Press, 42-68.

    Chapter  Google Scholar 

  45. Roulston K: Close encounters of the ‘CA’ kind: a review of literature analysing talk in research interviews. Qual Res. 2006, 6 (4): 515-534. 10.1177/1468794106068021.

    Article  Google Scholar 

  46. Stivers T, Majid A: Questioning children: Interactional evidence of implicit bias in medical interviews. Soc Psychol Q. 2007, 70 (4): 424-441. 10.1177/019027250707000410.

    Article  Google Scholar 

  47. Stokoe E: Simulated interaction and communication skills training: The ‘Conversation Analytic Role-play Method’. Applied Conversation Analysis: Intervention and change in institutional talk. Edited by: Antaki C. 2011, Basingstoke: Palgrave Macmillan

    Google Scholar 

  48. Sheon N, Lee S, Facente S: From questionnaire to conversation: a structural intervention to improve HIV test counselling. Patient Educ Couns. 2010, 81: 468-475. 10.1016/j.pec.2010.08.011.

    Article  PubMed  PubMed Central  Google Scholar 

  49. Heritage J, Robinson J, Elliott M, Beckett M, Wilkes M: Reducing patients’ unmet concerns in primary care: the difference one word can make. J Gen Intern Med. 2007, 22 (10): 1429-1433. 10.1007/s11606-007-0279-0.

    Article  PubMed  PubMed Central  Google Scholar 

  50. Mangione-Smith R, Stivers T, Elliott M, McDonald L, Heritage J: Online commentary during the physical examination: a communication tool for avoiding inappropriate antibiotic prescribing?. Soc Sci Med. 2003, 56: 313-320. 10.1016/S0277-9536(02)00029-1.

    Article  PubMed  Google Scholar 

  51. Parry R, Land V, Seymour J: Communicating face-to-face about sensitive future matters including end of life: a systematic review of evidence from fine-grained observational research. 2011, PROSPERO, CRD42011001626 []

    Google Scholar 

  52. Seymour J, French J, Richardson E: Dying matters: let’s talk about it. BMJ. 2010, 341: 646-648.

    Article  Google Scholar 

  53. The General Medical Council: Treatment and care towards the end of life: good practice in decision making. 2010, Ethical Guidance. London: General Medical Council, []

    Google Scholar 

  54. Royal College of General Practitioners, Royal College of Nursing: End of Life Care Patient Charter - A charter for the care of people who are nearing the end of their life. 2011, []

    Google Scholar 

  55. Public Health Intelligence North East: ‘A Good Death’ Consultation: Full Research Findings. 2010, []

    Google Scholar 

  56. Raijmakers N, van Zuylen L, Costantini M, Caraceni A, Clark J, De Simone G, Lundquist G, Voltz R, Ellershaw J, ven der Heide A, on behalf of OPCARE9: Issues and needs in end-of-life decision making: an international modified delphi study. Palliat Med. 2011, 26 (7): 947-953.

    Article  PubMed  Google Scholar 

  57. Cherlin E, Fried T, Prigerson H, Schulman-Green D, Johnson-Hurzeler R, Bradley E: Communication between physicians and family caregivers about care at the end of life: When do disucsisons occur and what is said. J Palliat Med. 2005, 8 (6): 1176-1185. 10.1089/jpm.2005.8.1176.

    Article  PubMed  PubMed Central  Google Scholar 

  58. Peräkylä A: AIDS Counselling: Institutional interaction and clinical practice. 1995, Cambridge: Cambridge University Press

    Book  Google Scholar 

  59. Lutfey K, Maynard D: Bad news in oncology: how physician and patient talk about death and dying without using those words. Soc Psychol Q. 1998, 61 (4): 321-341. 10.2307/2787033.

    Article  Google Scholar 

  60. Parry R: Are interventions to enhance communication performance in allied health professionals effective and how should they be delivered? Direct and indirect evidence. Patient Educ Couns. 2008, 73: 186-195. 10.1016/j.pec.2008.05.029.

    Article  PubMed  Google Scholar 

  61. Toerien M, Brookes ST, Metcalfe C, de Salis I, Tomlin Z, Peters TJ, Sterne J, Donovan JL: A review of reporting of participant recruitment and retention in RCTs in six major journals. Trials. 2009, 10 (1): 52-10.1186/1745-6215-10-52.

    Article  PubMed  PubMed Central  Google Scholar 

  62. Whiting P, Toerien M, de Salis I, Sterne JA, Dieppe P, Egger M, Fahey T: A review identifies and classifies reasons for ordering diagnostic tests. J Clin Epidemiol. 2007, 60 (10): 981-989. 10.1016/j.jclinepi.2007.01.012.

    Article  PubMed  Google Scholar 

  63. Bruinsma SM, Rietjens JAC, Seymour JE, Anquinet L, van der Heide A: The experiences of relatives with the practice of palliative sedation: a systematic review. J Pain Symptom Manage. 2012, 44 (3): 431-445. 10.1016/j.jpainsymman.2011.09.006.

    Article  PubMed  Google Scholar 

  64. Cox K, Bird L, Arthur A, Kennedy S, Pollock K, Kumar A, Stanton W, Seymour J: Public attitudes to death and dying in the UK: a review of published literature. BMJ Supportive & Palliative Care. 2013, 3 (1): 37-45. 10.1136/bmjspcare-2012-000203.

    Article  Google Scholar 

  65. Drew P: Conversation analysis. Handbook of Language and Social Interaction. Edited by: Fitch K, Sanders R. 2005, Mahwah NJ: Lawrence Erlbaum

    Google Scholar 

  66. Toerien M, Shaw R, Duncan R, Reuber M: Offering patients choices: A pilot study of interactions in the seizure clinic. Epilepsy and Behaviour. 2011, 20: 312-320. 10.1016/j.yebeh.2010.11.004.

    Article  Google Scholar 

  67. Parry R: Giving reasons for doing something now, or at some other time. Res Lang Soc Interact. 2013, 46 (2): 105-124. 10.1080/08351813.2012.754653.

    Article  Google Scholar 

  68. Dropbox file hosting service.,

  69. Endnote Web Web based reference organiser. []

  70. Levinson SC: Pragmatics (Cambridge textbooks in linguistics). 1983, Cambridge: Cambridge University Press

    Google Scholar 

  71. Heritage J: Conversation analysis and institutional talk. Handbook of Language and Social Interaction. Edited by: Fitch K, Sanders R, Mahwah NJ. 2005, : Lawrence Erlbaum, 103-148.

    Google Scholar 

  72. Greenhalgh T, Peacock R: Effectiveness and efficiency of search methods in systematic reviews of complex evidence: audit of primary sources. BMJ. 2005, 331: 1064-1065. 10.1136/bmj.38636.593461.68.

    Article  PubMed  PubMed Central  Google Scholar 

  73. Ethno/CA News: Information on Ethnomethodology and Conversation Analysis: Bibliographies. []

  74. Languse – Internet forum for the discussion of issues relationg to research in language and social interaction. []

  75. Google Scholar website. []

  76. Drummond A: Personal communication. 2013, Number of reviewers involved in selecting publications at different stages of a review

    Google Scholar 

  77. Dixon-Woods M, Shaw R, Agarwal S, Smith J: The problem of appraising qualitative research. Qual Saf Health Care. 2004, 13: 223-225. 10.1136/qshc.2003.008714.

    Article  CAS  PubMed  PubMed Central  Google Scholar 

  78. Noyes J, Popay J, Pearson A, Hannes K: Qualitative research and Cochrane reviews. Cochrane handbook for systematic reviews of interventions. Edited by: Higgins J, Green S. 2008, London: Wiley Blackwell, 571-592.

    Chapter  Google Scholar 

  79. Peräkylä A: Authority and accountability: the delivery of diagnosis in primary health care. Soc Psychol Q. 1998, 61 (4): 301-320. 10.2307/2787032.

    Article  Google Scholar 

  80. Curl T, Drew P: Contingency and action: a comparison of two forms of requesting. Res Lang Soc Interact. 2008, 41 (2): 129-153. 10.1080/08351810802028613.

    Article  Google Scholar 

  81. Atkins S, Lewin S, Smith H, Engel M, Fretheim A, Vomink J: Conducting a meta-ethnography of qualitative literature: Lessons learnt. BMC Med Res Methodol. 2008, 8: 21-10.1186/1471-2288-8-21.

    Article  PubMed  PubMed Central  Google Scholar 

  82. Morse J: Biased reflections: principles of sampling and analysis in qualitative inquiry. Moving beyond effectiveness in evidence synthesis: Methodological issues in the synthesis of diverse sources of evidence. Edited by: Popay J. 2006, London: National Institute for Health and Clinical Excellence, 53-60.

    Google Scholar 

  83. Denzin N, Lincoln Y: The discipline and practice of qualitative research. The Sage Handbook of Qualitative Research. 2005, Thousand Oaks: Sage, 1-42. 3

    Google Scholar 

  84. Heritage J: Conversation analysis. Garfinkel and Ethnomethodology. 1984, Oxford: Polity Press, 233-292.

    Google Scholar 

  85. Drew P, Chatwin J, Collins S: Conversation analysis: a method for research into interactions between patients and health-care professionals. Health Expect. 2001, 4: 58-70. 10.1046/j.1369-6513.2001.00125.x.

    Article  CAS  PubMed  Google Scholar 

  86. Schegloff EA: Analyzing single episodes of interaction: an exercise in conversation analysis. Soc Psychol Q. 1987, 50 (2): 101-114. 10.2307/2786745.

    Article  Google Scholar 

  87. Liberati A, Altman DG, Tetzlaff J, Mulrow C, Gøtzsche PC, Ioannidis JP, Clarke M, Devereaux P, Kleijnen J, Moher D: The PRISMA statement for reporting systematic reviews and meta-analyses of studies that evaluate health care interventions: explanation and elaboration. Ann Intern Med. 2009, 151 (4): W65-W94.

    Article  PubMed  Google Scholar 

  88. Wong G, Greenhalgh T, Westhorp G, Buckingham J, Pawson R: RAMESES publication standards: realist syntheses. BMC Med. 2013, 11 (1): 21-10.1186/1741-7015-11-21.

    Article  PubMed  PubMed Central  Google Scholar 

  89. Social Care Institute for Excellence: Using systematic reviews to improve social care. 2003, London: Social Care Institute for Excellence, []

    Google Scholar 

Pre-publication history

Download references


This project was supported by the University of Nottingham’s Sue Ryder Care Centre for the Study of Supportive, Palliative, and End of Life Care and by a grant from the University’s School of Nursing, Midwifery and Physiotherapy. We thank the review advisory team for their intelligent contributions: Dr Merran Toerien of the University of York’s Department of Sociology, Professor Paul Drew previously of the University of York, now at Loughborough University, and Professor Jane Seymour of the Sue Ryder Care Centre for the Study of Supportive, Palliative and End of Life Care at the University of Nottingham. Professor Douglas Maynard, Dr Peter Nowak, Professor Tony Arthur, and Professor Elizabeth Stokoe provided very helpful comments on earlier versions of this paper.

Author information

Authors and Affiliations


Corresponding author

Correspondence to Ruth H Parry.

Additional information

Competing interests

RP and VL declare we have no financial or non-financial competing interests.

Authors’ contributions

RP devised and led design of the review approach and the ‘Review of Future Talk’, supervised its execution, and drafted the manuscript. VL participated in the design and execution of the Review – including the bulk of database searching, organising data, designing templates and extracting data, and contributed to the manuscript. Both authors read and approved the final manuscript.

Electronic supplementary material


Additional file 1: Word groups used for electronic database searching in the Review of Future Talk. File containing word group terms used in searching. (PDF 175 KB)


Additional file 2: Template used for recording characteristics of data and analysis for studies in the Review of Future Talk. File containing blank template used in recording the characteristics of each study with regards data and analysis. (PDF 201 KB)


Additional file 3: Template used for data extraction (background details, findings and claims) of papers in the Review of Future Talk. File containing blank template used for data extraction. (PDF 294 KB)

Rights and permissions

Open Access This article is published under license to BioMed Central Ltd. This is an Open Access article is distributed under the terms of the Creative Commons Attribution License ( ), which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.

Reprints and permissions

About this article

Cite this article

Parry, R.H., Land, V. Systematically reviewing and synthesizing evidence from conversation analytic and related discursive research to inform healthcare communication practice and policy: an illustrated guide. BMC Med Res Methodol 13, 69 (2013).

Download citation

  • Received:

  • Accepted:

  • Published:

  • DOI: