Skip to main content
  • Research article
  • Open access
  • Published:

Comparison of psychometric properties between recall methods of interview-based physical activity questionnaires: a systematic review

Abstract

Background

This systematic review examined the methodological quality of studies and assessed the psychometric qualities of interview-administered Past-week and Usual-week Physical Activity Questionnaires (PAQs). Pubmed and Embase were used to retrieve data sources.

Methods

The studies were selected using the following eligibility criteria: 1) psychometric properties of PAQs were assessed in adults; 2) the PAQs either consisted of recall periods of usual 7-days (Usual-week PAQs) within the past 12 months or during the past 7-days (Past-week PAQs); and 3) PAQs were interview-administered. The COSMIN taxonomy was utilised to critically appraise study quality and a previously established psychometric criteria employed to evaluate the overall psychometric qualities.

Results

Following screening, 42 studies were examined to determine the psychometric properties of 20 PAQs, with the majority of studies demonstrating good to excellent ratings for methodological quality. For convergent validity (i.e., the relationship between PAQs and other measures), similar overall associations were found between Past-week PAQs and Usual-week PAQs. However, PAQs were more strongly associated with direct measures of physical activity (e.g., accelerometer) than indirect measures of physical activity (i.e., physical fitness), irrespective of recall methods. Very few psychometric properties were examined for each PAQ, with the majority exhibiting poor ratings in psychometric quality. Only a few interview-administered PAQs exhibited positive ratings for a single psychometric property, although the other properties were either rated as poor or questionable, demonstrating the limitations of current PAQs.

Conclusion

Accordingly, further research is necessary to explore a greater number of psychometric properties, or to develop new PAQs by addressing the psychometric limitations identified in the current review.

Peer Review reports

Background

The prevalence and severity of obesity is continually increasing in most of the Western world, developing into epidemic proportions worldwide [1]. Lack of physical activity reduces physical fitness, and is a major contributor to this global health crisis and is associated with development of chronic diseases and cancer, leading to increased mortality [2]. In contrast, participation in physical activity has been associated with improved health outcomes, lower incidences of health problems and reduced mortality rates [3,4,5]. International guidelines developed by the World Health Organisation (WHO) recommend that children and adults to engage in 60 min of moderate-to-vigorous physical activity each day [6, 7]. To assess whether physical activity is being performed at these recommended levels, adequate monitoring of patient’s lifestyles and behaviours is needed [8]. This enables health professionals to establish disease risks and develop interventions to address physical inactivity.

Questionnaires are typically used to assess physical activity level, as: 1) they are cost-effective and practical; 2) a large amount of information is collectable in a relatively short period of time; and 3) the results are easily quantifiable [9]. However, a number of disadvantages have been proposed, such as: 1) results are influenced by subjective measures; 2) misinterpretation of questions and recall bias due to language/cultural barriers or cognitive impairment; and 3) questionnaires not specifically developed for people with certain physiological/chronic conditions. Nonetheless, physical activity questionnaires are currently the most widely-used and acceptable forms of obtaining information on physical activity characteristics, particularly for larger-scale observational studies and research interventions [9]. There are several types of physical activity questionnaires which are primarily categorised according to recall periods. The two main recall methods currently utilised, measures recent physical activity performed over the past 7 days (i.e., Past-week PAQs) and the average week physical activity performed within the past 1–12 months (i.e., Usual-week PAQs) [10]. Previous research suggested that Past-week PAQs more accurately reflect the actual physical activity characteristics undertaken that week; however, Usual-week PAQs may minimise week-to-week variability [10], seasonal differences [11] and lifestyle factors, such as pregnancy [12]. Accordingly, the two recall methods may have distinct psychometric characteristics. It is therefore important to establish the validity and reliability of both types of PAQs, to ensure the PAQs selected are fit for purpose.

The Consensus-based Standards for the Selection of Health Measurement Instrument (COSMIN) checklist is a critical appraisal tool which evaluates methodological quality of studies that examine the psychometric properties of health related measures [13]. When combined with quality criteria for psychometric properties [14, 15], it provides a contemporary framework to assess overall psychometric quality of PAQs. According to a recent systematic review using the COSMIN checklist [16], when convergent validity was assessed by comparing PAQs with other measures (e.g., accelerometer or other PAQs), Past-week PAQs had higher correlations than Usual-week PAQs. These findings demonstrated that Past-week PAQs may assist clinicians in accessing the same constructs as those of other measures with better precision of PA level recordings. However, Doma and colleagues [16] only reported on studies that examined the psychometric properties of PAQs that were self-administered and excluded studies that administered PAQs via interviews. Whilst the ability to compare these data from self-administered PAQs to interview-administered PAQs are limited, it is currently the best available evidence of a similar construct.

The majority of PAQs can be either self-administered or interview-administered via face-to-face or telephone conducted by trained interviewers. For example, Active Australia Survey (AAS) is a commonly administered PAQ which assesses past-week PA level; its psychometric properties have previously been evaluated via both self-administration [17] and interview-administration [18]. The advantages of self-administered PAQs are that it is cost-effective, particularly when the PAQs distributed via postal mail or online, and minimises interviewer bias [19]. However, self-administered PAQs also risk introducing respondent bias, especially if respondents have literacy and numeracy difficulties [20]. These limitations can be overcome with interview-administered PAQs, although interviewees may overestimate reporting of their PA level due to social desirability [21]. In addition, the interviewee may over- or under-report physical activity level if instructions given by interviewers are not well standardised, or if interviewers are selective with phrasing the PAQs [22].

Although there is evidence that the mode of questionnaire administration may influence the accuracy and quality of the responses [20], to date, systematic reviews have only reported on the psychometric properties of self-administered PAQs [16, 23, 24], with overall findings indicating that only a few self-reported PAQs had reasonable reliability and validity ratings. No systematic reviews have explored the literature to determine the psychometric properties of interview-administered PAQs, particularly when compared between Past-week and Usual-week PAQs.

Therefore, the purpose of this systematic review was to evaluate the methodological quality of studies that have investigated the psychometric properties of interview-administered Past-week and Usual-week PAQs and to determine the overall psychometric quality for each PAQ. The results of this review will aid practitioners and researchers in selecting interview-administered PAQs that are appropriate for their purposes and through identifying the effects of recall differences on psychometric soundness.

Methods

The current systematic review was conducted in accordance with the Preferred Reporting Items for Systematic Reviews and Meta-Analyses (PRISMA) statement [25]. The PRISMA statement is a checklist that consists of 27 items that are used to ensure transparency of reporting for systematic reviews.

Inclusion/exclusion criteria

Studies on the psychometric properties of PAQs were only considered eligible if: 1) published in English; 2) physical activity questionnaires were developed in English; 3) administered to adults (>18yo) in English-speaking countries either with, or without pathological conditions (e.g., cardiovascular disease, musculoskeletal disease, metabolic disease or respiratory disease); 4) questionnaires consisted of recall methods of the past-week (i.e., previous 7 days) and usual-week (i.e., previous 7 days over 1–12 months); 5) questionnaires classified physical activity level based on energy expenditure, step count, distance travelled or duration of physical activity with the corresponding metabolic equivalent of task (MET); and 6) if the questionnaires were administered by trained interviewers. Studies were excluded if: 1) published as abstracts, 2) conference proceedings or dissertations; 3) used questionnaires with recall methods of less than 7 days, or recall over the previous 1–12 months that do not report average physical activity level over a 7-day period (i.e., average physical activity over the past month would be excluded whilst average 7-day physical activity over the past month would be included); 4) conducted using paediatric population or those with known cognitive impairment; 5) used questionnaires were translated into a language other than English; and 6) if the questionnaires were administered to individuals from non-English speaking backgrounds as cross-cultural validation was beyond the scope of this systematic review.

Information sources

A systematic literature search was conducted by two authors in June 2017 using two electronic databases (Embase and Pubmed). Subject headings and free text were used as part of the search for both databases, with date restrictions of the past half year applied for the free text search (refer to Table 1 for all search terms used during each electronic search). Following elimination of duplicates, a total of 7191 abstracts were retrieved from the search. The search process summary in accordance with the PRISMA guidelines is depicted in Fig. 1.

Table 1 Search terms and databases used to obtain abstracts
Fig. 1
figure 1

Literature search conducted based on the PRISMA guidelines

Selection process

Two independent reviewers with a health science background initially screened all abstracts against the pre-established inclusion/exclusion criteria, with each abstract rated as either meeting (“yes”), potentially meeting (“maybe”) or not meeting (“no”) the inclusion criteria. The reviewers were also trained by the primary author (KD) to ensure transparency of the inclusion criteria prior to abstract screening. Upon completion of abstract selection, a random 40% of abstracts were compared between the two independent reviewers and any disagreement between reviewers were consulted by a third reviewer (KD). Our calculation showed a Weighted Kappa calculation of 0.85 (95%CI: 0.81–0.90) which was considered as excellent for inter-rater reliability [26]. Original articles from selected abstracts classified as either “yes” or “maybe” were accessed for further screening by the two reviewers using the same inclusion/exclusion criteria. The first author provided final decisions if any disparity occurred between the two reviewers during the selection process of original articles.

Assess study methodological quality using COSMIN ratings

The COSMIN taxonomy of measurement properties and definitions for health-related patient-reported outcomes were used to evaluate the methodological quality of the included studies [27] (see Table 2). The COSMIN checklist evaluates the methodological quality of studies on psychometric properties and consists of nine domains: internal consistency, reliability (test-retest reliability, inter-rater reliability and intra-rater reliability), measurement error (absolute measures), content validity, structural validity, hypothesis testing, cross-cultural validity, criterion validity and responsiveness [13]. ‘Interpretability’ is not considered to be a psychometric property; thus, it was excluded from this review. Of the nine domains, ‘responsiveness’ was not evaluated as the questionnaire’s ability to detect changes over time was beyond the scope of this review. Furthermore, ‘cross-cultural validity’ was not assessed, as questionnaires either administered in non-English speaking countries or translated into non-English languages did not meet this review’s inclusion criteria. Finally, whilst accelerometry and double-labelled water technique are considered the ‘gold standard’ of assessing physical activity level, there is a risk of under-reporting certain exercise modes (e.g., swimming or resistance training) [28]. Therefore, comparison of physical activity level reported by PAQs and objective measures (i.e., accelerometer and double-labelled water method) was considered as ‘convergent validity’.

Table 2 Definitions for aspects of domains and measurement properties from the COSMIN checklist by Mokkink et al. [27]

Each COSMIN checklist domain consists of between 5 to 18 items which addresses various aspects of study design and statistical methods. Based on a 4-point rating system (i.e., excellent, good, fair and poor, respectively), Terwee and colleagues [13] initially suggested that the overall methodological quality of each domain should mirror the rating of the lowest-rated item (i.e., if four items were rated ‘Good’ and one ‘Poor’, the overall score would be ‘Poor’). However, given that each domain consists of items that assess a variety of methodological qualities, rating the overall methodological quality of a domain solely based on the lowest scoring single item undermines the ability of the checklist to explore subtle differences in psychometric qualities of each questionnaire [29]. Subsequently, a revised scoring method was implemented for this review by reporting the overall methodological quality of each domain as a percentage rating, as per Cordier, Speyer [29]. This revised scoring method has also been utilised successfully in a systematic review that compared Past-week and Usual-week PAQs, with sufficient sensitivity to detect differences between psychometric properties [16]. Specifically, the raw scores of each item were used to calculate a percentage of rating according to the following formula:

$$ Total\ score\ of\ each\ domain=\frac{\left(\mathrm{Total}\ \mathrm{score}\ \mathrm{obtained}-\mathrm{minimum}\ \mathrm{score}\ \mathrm{obtained}\right)}{\left(\mathrm{Highest}\ \mathrm{score}\ \mathrm{possible}-\mathrm{minimum}\ \mathrm{score}\ \mathrm{possible}\right)}\times 100 $$

The final percentage score depicting the overall methodological quality of each domain was then classified as follows: Poor = 0–25.0%, Fair = 25.1–50.0%, Good = 50.1–75.0%, Excellent = 75.1–100.0% [30]. Once the psychometric quality ratings of each paper were completed, ratings from a random 40% of papers were compared between two independent reviewers (KD and LP), resulting in a weighted Kappa of 0.84 (0.62–1.00), indicating excellent agreement.

Quality of the psychometric properties

To account for varying sample sizes of each study when comparing the reliability (i.e., reproducibility) and convergent validity (a form of hypothesis testing that evaluates the correlation between two related measures, for example, physical activity levels measured from the PAQs under investigation and other measures) between PAQs, the weighted mean of correlation coefficient (i.e., r-values) were calculated, using the following formula:

$$ \overline{x}=\frac{\sum \limits_{i=1}^n{w}_i{x}_i}{\sum \limits_{i=1}^n{w}_i} $$

Where w = r-value of the comparison within a study (e.g., PAQ vs. another instrument or PAQ vs. Accelerometer/pedometer) and x = sample size of the comparison.

For the strength of reliability, once the weighted r-values were calculated for each study per PAQ, these measures were then averaged to compare the overall correlation between Past-week and Usual-week PAQs. For the strength of convergent validity, weighted r-values were averaged to compare overall correlations between Past-week and Usual-week PAQs, and between parameters that reported direct measures of PA level (e.g., diaries, other PAQs, accelerometers, pedometers) and indirect measures of PA level (e.g., aerobic fitness, muscular strength). If the sample size between each study was equivalent, then the normal non-weighted r-values were averaged. The strength of correlation was classified according to Cohen’s method, with the following: 0–0.29, 0.3–0.49 and ≥ 0.5 as weak, moderate and strong, respectively [31].

The psychometric quality of each measurement property per PAQ for each study (Table 3) was also classified using the following quality criteria: “positive” (+), “conflicting” (±), “indeterminate” (?), “negative” (−), “not reported” (NR) or “not evaluated” (NE) [15, 30]. Studies that were rated as “poor” based on the COSMIN rating were excluded from further analyses and received “not evaluated” (NE). Finally, an overall quality score of assessments for each psychometric property was calculated based on the levels of evidence by Schellingerhout, Verhagen [14]. These scores were determined by integrating the methodological quality rating of the included studies on psychometric properties using the COSMIN checklist, and the quality criteria for measurement properties of assessment according to Terwee, Bot [15] and Cordier, Chen [30] (see Table 3). Figure 2 depicts a flowchart of the analysis process involved in determining the overall quality score for each assessment.

Table 3 The modified version of the psychometric quality rating set out by (Terwee et al., 2007) and (Cordier et al., [30])
Fig. 2
figure 2

Flow chart of the methodological quality rating based on COSMIN, derivation of the quality of psychometric properties and overall quality score per measurement

Data items and synthesis of results

Domains from the COSMIN checklist and psychometric property qualities were assessed for each included study according to Terwee, Bot [15] and Cordier, Chen [30]. The results were then reported in the following order: 1) the description of the literature search (see Table 1); 2) the characteristics of the interview-based PAQ measures (see Table 4) and studies reporting on the development and validation of the interview-based PAQ measures (see Table 4); 3) the methodological quality according to the COSMIN checklist of each study that have reported on the psychometric properties of PAQs (see Table 6); 4) the comparisons of the average weighted r-values of test-retest reliability and convergent validity between Past-week and Usual-week PAQs (see Table 7); 5) the quality of relevant psychometric properties for each study based on the criteria by Terwee, Bot [15] and Cordier, Chen [30] (Table 8); and 6) the overall quality rating of psychometric properties based on the levels of evidence by Schellingerhout, Verhagen [14] for each PAQ and comparing these results between Past-week and Usual-week PAQs (see Table 9).

Table 4 Characteristics of interview-administered Past-week and Usual-week physical activity questionnaires

Results

Systematic literature search

Following removal of duplicate abstracts, a total of 3447 abstracts were screened according to the inclusion criteria. Upon completion of screening, 75 PAQs and 117 of their corresponding full-text articles were examined for eligibility. Amongst these items, 20 PAQs and 42 of their corresponding articles were included. The remaining 55 PAQs were excluded for the following reasons: non-specified recall periods; recall period was beyond 7 days; recall period was less than 7 days; and various combinations of recall periods.

Included physical activity questionnaires

Table 4 displays the characteristics of the included PAQs, with description of their corresponding studies shown in Table 5. There were 6 PAQs that assessed Usual 7-days of PA level with two PAQs that had a 1-month recall period (Phone FITT and YPAS), one PAQ that had a 3-month recall period (IPEQ-WA) and three PAQs that had a 12-month recall period (CaMos, MAQ and NHS II; Table 4). The remaining 13 PAQs encompassed items that assessed PA level over the Past-7 days. A majority of PAQs had subscales that were separated by the intensity of PA (e.g., light, moderate and vigorous), whereas other PAQs had subscales categorised by the mode of PA (e.g., walking, stairs, occupational and gardening activities).

Table 5 Description of studies for the development and validation of interview-administered Past-week and Usual-week physical activity questionnaires

Psychometric properties of PAQs

Table 6 provides an overview of the methodological quality assessment of studies reporting on psychometric properties of usual-week and past-week physical activity questionnaires using the COSMIN checklist. The most frequently reported psychometric properties based on the COSMIN rating assessment was hypothesis testing (18 of 20 PAQs), ranging from fair to excellent qualities, followed by reliability (13 of 20 PAQs), ranging from good to excellent qualities. The least reported psychometric properties included measurement error (4 of 20 PAQs), ranging from good to excellent qualities, internal consistency (3 of 20 PAQs), ranging from poor to fair qualities and content validity (3 of 20 PAQs), ranging from fair to good qualities. No studies were identified that reported structural validity. When different PAQ recall methods were compared (i.e., Past-week PAQ versus Usual-week PAQ), similar frequencies in psychometric properties were found for Usual 7-day PAQs and Past 7-day PAQs with internal consistency (16.7 and 14.3%, respectively) and content validity (16.7 and 14.3%, respectively). However, notable differences were also shown with Usual 7-day PAQs more frequently reported for reliability (83.3% vs. 57.1%) and Past 7-day PAQs more frequently for measurement error (24.1% vs. 16.7%) and hypothesis testing (92.9% vs. 83.3%).

Table 6 Overview of the methodological quality assessment of studies reporting on psychometric properties of interview-administered Usual-week and Past-week physical activity questionnaires using the COSMIN checklist

Table 7 demonstrates the weighted mean of the r-values for test-retest reliability and convergent validity between the types of PAQ (i.e., Usual-week vs. Past-week) and type of comparator measures (i.e., direct vs. indirect measures). Test-retest reliability data was available for 7 of the 20 PAQs. According to the average weighted mean of the r-values, the reliability of both Usual-week and Past-week PAQs showed strong correlations when assessed across two separate time points, with similar reliability measures for Usual-week (r = 0.63) and Past-week (r = 0.56) PAQs. According to Cohen’s methods, when direct and indirect measures were combined for convergent validity (data was available for 17 of 20 PAQs), Usual-week PAQs exhibited a moderate correlation (r = 0.30), whereas Past-week PAQs shows a weak correlation (r = 0.28). With respect to measurement type for each recall of PAQ, the convergent validity for direct measures had moderate correlations for both Usual-week (r = 0.33) and Past-week PAQs (r = 0.40) compared to weak correlations for indirect measures (r = 0.28 and r = 0.19, respectively). When Usual-week and Past-week PAQs were compared separately between direct and indirect measures, similar correlations were observed for the Past-week PAQs (r = 0.40) and the Usual-week PAQs (r = 0.33) for direct measures with moderate correlations. However, for indirect measures, there was a moderate correlation (r = 0.33) for Usual-week PAQs whereas the Past-week PAQs had a weak correlation (r = 0.19). Finally, when both Past-week and Usual-week PAQs were combined, direct measures had a moderate correlation (r = 0.39) whereas indirect measures had a weak correlation (r = 0.21).

Table 7 The weighted mean of the r-values for reliability testing and convergent validity of Past-week and Usual-week physical activity questionnaires

Table 8 displays the quality of psychometric properties of both types of PAQs (i.e., Usual-week and Past-week) according to the criteria established by Terwee, Bot [15] and Cordier, Chen [30]. Table 9 provides the overall summary rating of the psychometric properties for each PAQ based on the levels of evidence by Schellingerhout, Verhagen [14]. According to Table 9, very few psychometric properties were reported (40 out of 120 possible ratings: 33.3%), with the quality of psychometric properties primarily reported for reliability (13/20: 65%) and hypothesis testing (18/20: 90%). Fewer results were identified for internal consistency (2/20: 10%), content validity (3/20: 15%) and measurement error (4/20: 20%), while structural validity was not rated for any of the PAQs. Of all the psychometric properties rated for psychometric quality [32], the results were mainly negative (17/40: 42.5%), consisting of “strong negative” (10/40: 25%), “moderate negative” (5/40: 12.5%) and “limited negative” (2/40: 5%). Several psychometric properties were reported with “conflicting” (13/40: 32.5%), whilst fewer psychometric properties were reported for “indeterminate” (5/40: 12.5%). One psychometric property had a “not evaluated” rating, due to poor COSMIN scoring.

Table 8 Quality of psychometric properties per study based on the criteria by Terwee, Bot [9] and Cordier, Chen [30]
Table 9 Overall quality score of psychometric properties for each interview-administered Usual-week and Past-week physical activity questionnaire using the levels of evidence by Schellingerhout et al., [14]

The relative number of negative (“strong negative” [6/18: 33.3%] and “moderate negative” [3/18: 16.7%]) and conflicting (8/18: 44.4%) ratings were reported the most for hypothesis testing; only one “strong positive” rating was identified. For reliability, a greater relative number of positive ratings (“strong positive” [1/13: 7.7%] and “moderate positive” [2/13: 15.4%]) were found. However, reliability also exhibited several negative (“strong negative” [4/13: 30.8%] and “moderate negative” [1/13: 7.7%]) and conflicting (5/13: 38.5%) ratings. The relative number of “indeterminate” ratings was greatest for measurement error (4/4: 100%), whilst internal consistency showed only one “indeterminate”, “limited negative” and “not evaluated” ratings (1/3: 33.3%, respectively). There was one “moderate positive” rating (1/3: 33.3%) for content validity, although the rest of the ratings consisted of one “moderate negative” and “limited negative” ratings (1/3: 33.3%, respectively).

When comparing the PAQs, reliability demonstrated positive ratings for Past-Week Modifiable Activity Questionnaire (PWMAQ) (“strong positive”), Phone-FITT (“moderate positive”) and the Veterans Physical Activity Questionnaire (VAPAQ) (“moderate positive”). However, the results for PWMAQ also demonstrated a “conflicting” rating in hypothesis testing, Phone-FITT received a “limited negative” and “moderate negative” in content validity and hypothesis testing, respectively, and VAPAQ received an “indeterminate” rating for measurement error. The AAS (modified) demonstrated a “strong positive” rating for hypothesis testing, although this PAQ also received a “strong negative” and “indeterminate” rating for reliability and measurement error, respectively. While PWMAQ, Phone-FITT, VAPAQ and AAS (modified) received mixed results, these measures have a substantial number of psychometric properties that were not reported.

When compared between the types of PAQs, a similar percentage of negative ratings (limited, moderate or strong) were shown for Past-week (12/28: 42.9%) and Usual-week (5/12: 41.7%) PAQs. Conversely, the relative number of positive ratings (limited, moderate or strong) for the Past-week PAQs (4/28: 14.3%) was greater than Usual-week PAQs (1/12: 8.3%), although the absolute number of “positive” ratings were small. The number of NR ratings [33] were noticeable across all PAQs.

Discussion

This systematic review examined the methodological quality of studies that investigated the psychometric properties of interview-administered, Usual-week and Past-week PAQs, in an adult population. There were 20 PAQs with 42 corresponding articles that reported on the psychometric properties of PAQs, of which 6 were Usual-week and 14 were Past-week PAQs. Amongst the psychometric properties, hypothesis testing was reported most frequently, followed by reliability, whereas measurement error, content validity and internal consistency were the least reported. Furthermore, structural validity was not reported in any of the included studies. The methodological quality of the studies exhibited good to excellent ratings across most of the psychometric properties. As per the average weighted mean of the r-values, both Usual-week PAQs and Past-week PAQs showed moderate correlations for reliability and convergent validity for direct measures, whereas convergent validity for indirect measures exhibited weak correlations irrespective of the type of PAQ. When comparing the weighted mean of the r-values between PAQ types, convergent validity for direct measures indicated moderate correlations for both Past-week and Usual-week PAQs, although convergent validity for indirect measures demonstrated moderate correlations for Usual-week PAQs, while weak correlations were observed for Past-week PAQs. According to the level of evidence (i.e., overall quality), most of the psychometric properties exhibited “moderate negative” to “strong negative” ratings irrespective of PAQ types, highlighting concerns for utilising current interview-administered PAQs.

Quality of studies based on the COSMIN taxonomy

Of the psychometric property reliability, most psychometric studies reported on test-retest reliability with good to excellent COSMIN ratings, whereas measurement error was only reported for four measures (YPAS, AAS [modified], PASE and VAPAQ), also with good to excellent ratings. Measurement error is an essential property of the reliability dimension, as it quantifies the magnitude of systematic and random error of PA levels that is not caused by true changes in the construct being measured; thus allowing practitioners to establish meaningful differences in PA measures [34]. Subsequently, more research is warranted to determine the measurement error of PAQs when administered in an interview setting. Compared to test-retest reliability and measurement error, internal consistency was reported for only three measures (YPAS, PASE and PAR) with poor to fair ratings. This discrepancy was due to included studies consisting of a small sample size and/or examining correlations between different items of the same PAQ without conducting Cronbach alpha statistic and factor analyses. According to Terwee et al. [13], both statistical approaches determine whether all items measure the same construct and checks the uni-dimensionality of the scale. Thus, future studies should consider these limitations when examining the internal consistency of interview-administered PAQs.

With the exception of two PAQs (CaMos and VAPAQ), hypothesis testing was reported for all PAQs with the majority of included studies reporting good to excellent ratings for methodological quality. However, studies only investigated three PAQs (Phone-FITT, PAAQ and PASE) for content validity with fair to good COSMIN ratings. The methodological limitations identified from these studies included lack of description on whether piloting was conducted by investigators, expert practitioners and/or the target population. Thus, future studies should consider these issues when examining the content validity of PAQs. Most alarming is that none of the PAQs investigated structural validity. This means that the underlying constructs of all the PAQs are currently unknown, as appropriate statistical analyses to ascertain the factor structure or dimensionality (e.g., dimensionality and principle component analysis using Rasch analysis and exploratory and/or confirmatory factor analysis) of the measures and associated subscales have not been conducted. Therefore, when assessing structural validity of PAQs, future studies should identify whether their approach is in accordance with a formative (i.e., integrative items forming a construct) or reflective (i.e., items are reflective of the same underlying constructs) model.

Quality of psychometric properties

In the current review, test-retest reliability for the Usual-week PAQs was comparable to the Past-week PAQs based on the average weighted mean of the r-values. These results are in contrast to a previous study by Delbaere, Hauer [35], who compared the reliability of self-administered incidental and planned exercise questionnaire (IPEQ) between Usual-week (i.e., past three months) and Past-week versions. According to their results, the Usual-week IPEQ (ICC = 0.84) version showed greater test-retest reliability compared to the Past-week IPEQ version (ICC = 0.77). Authors speculated that PAQs with usual 7-day recall periods during the past several months exhibit better stability in PA measures across time points, compared with PAQs with past-week recalls given that PA levels may fluctuate from week-to-week [10], or season-to-season [11]. The discrepancies in findings between the current review, and the work by Delbaere, Hauer [35], may be due to distinct acceptable cut-off levels being employed for test-retest reliability. For example, Delbaere, Hauer [35] established acceptable ICC values at ≥0.6, whereas the current review utilised an acceptable ICC criteria of ≥0.7 according to the criteria set out by [15]. Therefore, where ICC values (≥0.6 to < 0.7) were classified as “acceptable” for [35], would have been considered below the acceptable cut-off level in the current review with a “negative” rating. In addition, the test-retest reliability in the current review was compared between PAQs with different recall methods based on average weighted mean of the r-values across multiple studies, whereas [35] compared different recall versions of IPEQ within the same study and population. Subsequently, the variation in study design and the type of PAQs may have diluted potential differences in the weighted mean of the r-values between Past-week and Usual-week PAQs in the current review. This is further supported by a previous systematic review by [16], who also reported comparable test-retest reliability of average weighted mean of the r-values for self-administered Usual-week and Past-week PAQs.

When comparing convergent validity (i.e., hypothesis testing) between PAQ recall types, the average weighted mean of the r-values of Past-week PAQs were comparable with Usual-week PAQs for direct measures. However, the average weighted mean of the r-values was greater for Usual-week PAQs compared with Past-week PAQs within our review. These findings differ to a previous systematic review reported by [16] in self-reported PAQs where convergent validity for direct measures were greater for Past-week PAQs than Usual-week PAQs. In addition, the average weighted mean of the r-values for direct measures of convergent validity for both PAQ recall periods showed a moderate correlation (r = 0.35), whereas Doma, Speyer [16] reported weak correlations for the same measures (r = 0.27) based on self-reported PAQs in their previous review. This trend has also been reported by previous studies that compared convergent validity between interview-administered and self-administered PAQs [36, 37]. For example, Chu, Ng [36] reported stronger associations between Global Physical Activity Questionnaire (GPAQ) for the interview-administered method (r = 0.44–0.52) compared with the self-administered method (r = 0.28–0.38) when compared against accelerometers. Collectively, PAQs administered via interview may allow reporting of PA levels with greater accuracy than by self-administration, possibly due to minimisation of respondent bias [36].

While weighted-mean of the r-values for direct measures of convergent validity were comparable between Past-week and Usual-week PAQs, indirect measures of convergent validity were stronger for Usual-week PAQs (i.e., moderate correlations) than Past-week PAQs (i.e., weak correlations). This suggests that Usual-week PAQs better reflect physical fitness (e.g., VO2max, 6-min walk test) and its associated physiological conditions (e.g., BMI, body fat percentage) than Past-week PAQs when administered via interviews. These results are expected, given that physical fitness measures are stable across several weeks despite exercise termination [38], as opposed to the inherent week-to-week fluctuations observed with PA level [39]. Subsequently, when estimating physical fitness levels based on PA level ascertained from PAQs, we encourage the use of Usual-week PAQs rather than Past-week PAQs, particularly when administered via interviews. However, it should be noted that the current review included studies consisting of older adults with a number of pathological conditions (e.g., cardiovascular disease, musculoskeletal disease and neurological disease), where chronic exercise adaptations and deconditioning may differ in response to apparently healthy, younger individuals [32, 40, 41]. Separating these populations was difficult in the current review as the majority of studies incorporated apparently healthy participants with those who had several pathological conditions in the one study. Thus, future research should systematically compare psychometric properties of PAQs between individuals with pathological conditions and their apparently healthy counterparts.

For the overall Level of Evidence, irrespective of recall methods, there was a substantial number of missing psychometric data (i.e., not reported [NR]), indicating that the psychometric properties pertinent to determining the quality of current PAQs are not being examined effectively. Of the few psychometric properties reported, there were only four that reported “moderate positive” to “strong positive” ratings, with the rest as “strong to limited negative”, “indeterminate” and “conflicting” ratings, which demonstrate the weaknesses of current PAQs. No studies examined structural validity of PAQs, and only three PAQs examined internal consistency with ratings of “not evaluated”, “indeterminate” and “limited negative”. These weak results and lack of reporting is particularly concerning given that both structural validity and internal consistency are based on a reflective model, whereby all items are manifestations of the same underlying construct [27]. In addition, only a very limited number of PAQs reported on content validity (3/20: 15%), with one “positive” rating and two “negative” ratings. These findings further raise the limitations of current interview-administered PAQs, as content validity measures the degree to which the content of a PAQ is an adequate reflection of the construct being measured [27].

When comparing the overall psychometric qualities of PAQs based on Level of Evidence between recall methods, there were minute differences between Usual-week and Past-week PAQs, which are in line with findings by [16]. Additionally, the “moderate negative” to “strong negative” ratings shown for the majority of psychometric properties in the current review are similar to those reported by other systematic reviews [16, 42, 43]. These psychometric properties were rated poorly as the correlations were predominantly below the acceptable levels for test-retest reliability and convergent validity. However, authors from several studies included in the current review reported that the PAQs demonstrated acceptable test-retest reliability and convergent validity, which conflicts with findings from this current review. The discrepancy in these interpretations is because authors in the included studies considered test-retest reliability and convergent validity as acceptable based on level of significance (p ≤ 0.05), rather than the strength of the relationship (i.e., magnitude of the r-values). Accordingly, the strength of the relationship should be accounted for by future studies, as larger sample sizes are likely to generate associations at a statistically significant level, irrespective of the strength of the relationship. While the methodological quality of measurement error was rated as “good” to “excellent”, the four PAQs corresponding to these ratings (i.e., YPAS, AAS [modified], PASE and VAPAQ) were classified as “indeterminate” for psychometric quality. This is because the included studies did not report minimal important change (MIC) with respect to smallest detectable change (SDC), or whether MIC ranged beyond the limits of agreement (LOA). According to Terwee, Roorda [44], SDC and MIC are essential parameters for reliability to allow better interpretation of change scores. Subsequently, more studies need to incorporate measurement error when examining reliability of PAQs and consider calculations of MIC and SDC and/or LOA for this psychometric property.

Limitation

The primary purpose of the current review was to examine the psychometric properties of interview-administered, Past-week and Usual-week PAQs in an adult population. Thus, investigating the psychometric properties of PAQs with recall time-frames beyond, or within, the 7-day period was beyond the scope of the study. In addition, the current review selectively included studies that examined the psychometric properties of PAQs that were published in an English-speaking country because cultural diversity appears to impact on the psychometric properties of PAQs (e.g., errors of translation between languages, interpretation difficulties). Furthermore, the current review specifically selected studies that were conducted in an adult population, given that PAQs for children and adolescents are developed according to their literacy level. Thus, comparing the psychometric qualities of PAQs between studies that were conducted in English-speaking and non-English speaking countries and between age groups (i.e., children, adolescents and adults) may expand our knowledge on the usability of PAQs across different population groups. Whilst we made every effort to exclude studies that included participants with diagnosed cognitive impairment, the majority of the included studies did not screen for cognitive impairment. Therefore, future studies should consider conducting cognitive assessments to ensure that cognitive conditions are not influencing the psychometric properties of PAQs, particularly in older adults. With respect to abstract screening, there were discrepancies between those who conducted the literature search and those who screened the abstracts, which may have introduced selective bias. However, the reviewers were rigorously trained prior to abstract screening to ensure transparency of the inclusion criteria, and any disagreement between reviewers were resolved by the primary author (KD). Finally, examining the responsiveness and cross-cultural validity of PAQs was beyond the scope of this review. Therefore, comparing the psychometric quality of these properties between different PAQ types may allow better understanding of the sensitivity to changes in PA level.

Conclusion

The current review demonstrated that the psychometric quality of the majority of reported psychometric properties exhibited “negative” ratings. In addition, minimal differences were identified in the psychometric quality between Usual-week and Past-week PAQs. These findings suggested that the psychometric qualities of commonly used interview-administered PAQs are weak irrespective of recall methods. Therefore, caution should be used when measuring PA level using the PAQs included in this review. According to the weighted mean of the r-values, test-retest reliability was stronger for Usual-week PAQs compared with Past-week PAQs, although the reverse was identified for convergent validity for direct measures of PA level. These results indicate that Usual-week PAQs may be more suitable when identifying PA levels, and its corresponding association with physical fitness, of a large population for epidemiological studies. Conversely, Past-week PAQs may allow better detection of changes in PA level following an intervention. Finally, the interview-administered PAQs exhibited stronger convergent validity than previously reported for self-administered PAQs [16]. Therefore, whilst interview-administered PAQs may be time-consuming and cumbersome, researchers may opt to utilise this method over self-administered PAQs to obtain greater accuracy in physical activity level. However, irrespective of the strength of correlations, it is important to note that the quality of the measurement properties were either not examined or were quite poor. Subsequently, future studies should investigate the psychometric properties using more robust methodologies based on the COSMIN to better understand the usability of current PAQs, or to develop new PAQs by addressing issues identified in this review.

Abbreviations

AAS:

Active Australia Survey

CaMos:

Canadian Multicentre Osteoporosis Study

CAQ-PAI:

College Alumni Questionnaire – Physical Activity Index

COSMIN:

Consensus-based Standards for the Selection of Health Measurement Instrument

ICC:

Intra-class correlation coefficient

IPAQLF:

International Physical Activity Questionnaire Long Form

IPAQ-SF:

International Physical Activity Questionnaire Short Form

IPEQ-WA:

Incidental and planned exercise questionnaire

LOA:

Limits of agreement

MAQ:

Modified activity questionnaire

MET:

Metabolic equivalent of task

MIC:

Minimal important change

NE:

Not evaluated

NHS II:

Nurses’ Health Study version II

NR:

Not reported

NZPAQ-LF:

New Zealand Physical Activity Questionnaire Long Form

NZPAQ-SF:

New Zealand Physical Activity Questionnaire Short Form

PA:

Physical activity

PAAQ:

Physical activity adult questionnaire

PAQ:

Physical activity questionnaire

PAR:

Physical activity recall questionnaire

PASE:

Physical activity scale for the elderly

Phone-FITT:

Phone fitness

PRISMA:

Preferred reporting items for systematic reviews and meta-analyses

PWMAQ:

Past week modified activity questionnaire

SCD:

Smallest detectable change

VAPAQ:

Veterans Physical Activity Questionnaire

WHO:

World health organisation

YPAS:

Yale Physical Activity Survey

References

  1. Rush E, Simmons D. Physical activity in children: prevention of obesity and type 2 diabetes. Med Sport Sci. 2014;60:113–21.

    Article  PubMed  Google Scholar 

  2. Pedersen BK, Saltin B. Exercise as medicine - evidence for prescribing exercise as therapy in 26 different chronic diseases. Scand J Med Sci Sports. 2015;25(Suppl 3):1–72.

    Article  PubMed  Google Scholar 

  3. Doma K, Grant A, Morris J. The effects of balance training on balance performance and functional outcome measures following Total knee arthroplasty: a systematic review and meta-analysis. Sports Med. 2018;48(10):2367–85.

    Article  PubMed  Google Scholar 

  4. Pierce DR, Doma K, Leicht AS. Acute effects of exercise mode on arterial stiffness and wave reflection in healthy young adults: a systematic review and meta-analysis. Front Physiol. 2018;9:73.

    Article  PubMed  PubMed Central  Google Scholar 

  5. Poirier P, Giles TD, Bray GA, Hong Y, Stern JS, Pi-Sunyer FX, et al. Obesity and cardiovascular disease: pathophysiology, evaluation, and effect of weight loss: an update of the 1997 American Heart Association scientific statement on obesity and heart disease from the obesity Committee of the Council on nutrition, physical activity, and metabolism. Circulation. 2006;113(6):898–918.

    Article  PubMed  Google Scholar 

  6. Fussenich LM, Boddy LM, Green DJ, Graves LE, Foweather L, Dagger RM, et al. Physical activity guidelines and cardiovascular risk in children: a cross sectional analysis to determine whether 60 minutes is enough. BMC Public Health. 2016;16(1):67.

    Article  CAS  PubMed  PubMed Central  Google Scholar 

  7. Oja P, Titze S. Physical activity recommendations for public health: development and policy context. EPMA J. 2011;2(3):253–9.

    Article  PubMed  PubMed Central  Google Scholar 

  8. Babor TF, Sciamanna CN, Pronk NP. Assessing multiple risk behaviors in primary care. Screening issues and related concepts. Am J Prev Med. 2004;27(2 Suppl):42–53.

    Article  PubMed  Google Scholar 

  9. Lee PH, Macfarlane DJ, Lam TH, Stewart SM. Validity of the international physical activity questionnaire short form (IPAQ-SF): a systematic review. Int J Behav Nutr Phys Act. 2011;8:115.

    Article  PubMed  PubMed Central  Google Scholar 

  10. Ainsworth BE, Caspersen CJ, Matthews CE, Masse LC, Baranowski T, Zhu W. Recommendations to improve the accuracy of estimates of physical activity derived from self report. J Phys Act Health. 2012;9(Suppl 1):S76–84.

    Article  PubMed  PubMed Central  Google Scholar 

  11. Shephard RJ, Aoyagi Y. Seasonal variations in physical activity and implications for human health. Eur J Appl Physiol. 2009;107(3):251–71.

    Article  PubMed  Google Scholar 

  12. Merkx A, Ausems M, Bude L, de Vries R, Nieuwenhuijze MJ. Factors affecting perceived change in physical activity in pregnancy. Midwifery. 2017;51:16–23.

    Article  PubMed  Google Scholar 

  13. Terwee CB, Mokkink LB, Knol DL, Ostelo RW, Bouter LM, de Vet HC. Rating the methodological quality in systematic reviews of studies on measurement properties: a scoring system for the COSMIN checklist. Qual Life Res. 2012;21(4):651–7.

    Article  PubMed  Google Scholar 

  14. Schellingerhout JM, Verhagen AP, Heymans MW, Koes BW, de Vet HC, Terwee CB. Measurement properties of disease-specific questionnaires in patients with neck pain: a systematic review. Qual Life Res. 2012;21(4):659–70.

    Article  PubMed  Google Scholar 

  15. Terwee CB, Bot SD, de Boer MR, van der Windt DA, Knol DL, Dekker J, et al. Quality criteria were proposed for measurement properties of health status questionnaires. J Clin Epidemiol. 2007;60(1):34–42.

    Article  PubMed  Google Scholar 

  16. Doma K, Speyer R, Leicht AS, Cordier R. Comparison of psychometric properties between usual-week and past-week self-reported physical activity questionnaires: a systematic review. Int J Behav Nutr Phys Act. 2017;14(1):10.

    Article  PubMed  PubMed Central  Google Scholar 

  17. Brown WJ, Burton NW, Marshall AL, Miller YD. Reliability and validity of a modified self-administered version of the active Australia physical activity survey in a sample of mid-age women. Aust N Z J Public Health. 2008;32(6):535–41.

    Article  PubMed  Google Scholar 

  18. Fjeldsoe BS, Winkler EA, Marshall AL, Eakin EG, Reeves MM. Active adults recall their physical activity differently to less active adults: test-retest reliability and validity of a physical activity survey. Health Promot J Austr. 2013;24(1):26–31.

    Article  PubMed  Google Scholar 

  19. Touvier M, Kesse-Guyot E, Mejean C, Pollet C, Malon A, Castetbon K, et al. Comparison between an interactive web-based self-administered 24 h dietary record and an interview by a dietitian for large-scale epidemiological studies. Br J Nutr. 2011;105(7):1055–64.

    Article  CAS  PubMed  Google Scholar 

  20. Bowling A. Mode of questionnaire administration can have serious effects on data quality. J Public Health (Oxf). 2005;27(3):281–91.

    Article  Google Scholar 

  21. Okamoto K, Ohsuka K, Shiraishi T, Hukazawa E, Wakasugi S, Furuta K. Comparability of epidemiological information between self- and interviewer-administered questionnaires. J Clin Epidemiol. 2002;55(5):505–11.

    Article  PubMed  Google Scholar 

  22. Schnall AH, Wolkin A, Nakata N. In: Horney JA, editor. Chapter 9 Methods: questionnaire development and interviewing techniques. London: Mica Haley; 2018.

    Google Scholar 

  23. Forsen L, Loland NW, Vuillemin A, Chinapaw MJ, van Poppel MN, Mokkink LB, et al. Self-administered physical activity questionnaires for the elderly: a systematic review of measurement properties. Sports Med. 2010;40(7):601–23.

    Article  PubMed  Google Scholar 

  24. van Poppel MN, Chinapaw MJ, Mokkink LB, van Mechelen W, Terwee CB. Physical activity questionnaires for adults: a systematic review of measurement properties. Sports Med. 2010;40(7):565–600.

    Article  PubMed  Google Scholar 

  25. Liberati A, Altman DG, Tetzlaff J, Mulrow C, Gotzsche PC, Ioannidis JP, et al. The PRISMA statement for reporting systematic reviews and meta-analyses of studies that evaluate healthcare interventions: explanation and elaboration. BMJ. 2009;339:b2700.

    Article  PubMed  PubMed Central  Google Scholar 

  26. Cicchetti DV. Guidelines, criteria, and rules of thumb for evaluating normed and standardized assessment instruments in psychology. Psychol Assess. 1994;6(4):284–90.

    Article  Google Scholar 

  27. Mokkink LB, Terwee CB, Patrick DL, Alonso J, Stratford PW, Knol DL, et al. The COSMIN checklist for assessing the methodological quality of studies on measurement properties of health status measurement instruments: an international Delphi study. Qual Life Res. 2010;19(4):539–49.

    Article  PubMed  PubMed Central  Google Scholar 

  28. Aparicio-Ugarriza R, Mielgo-Ayuso J, Benito PJ, Pedrero-Chamizo R, Ara I, Gonzalez-Gross M, et al. Physical activity assessment in the general population; instrumental methods and new technologies. Nutr Hosp. 2015;31(Suppl 3):219–26.

    PubMed  Google Scholar 

  29. Cordier R, Speyer R, Chen YW, Wilkes-Gillan S, Brown T, Bourke-Taylor H, et al. Evaluating the psychometric quality of social skills measures: a systematic review. PLoS One. 2015;10(7):e0132299.

    Article  PubMed  PubMed Central  CAS  Google Scholar 

  30. Cordier R, Chen YW, Speyer R, Totino R, Doma K, Leicht A, et al. Child-report measures of occupational performance: a systematic review. PLoS One. 2016;11(1):e0147751.

    Article  PubMed  PubMed Central  CAS  Google Scholar 

  31. Cohen J. Statistical power analysis for the behavioral sciences. Hillsdale: Lawrence Erlbaum Associates; 1988.

    Google Scholar 

  32. Hakkinen A, Malkia E, Hakkinen K, Jappinen I, Laitinen L, Hannonen P. Effects of detraining subsequent to strength training on neuromuscular function in patients with inflammatory arthritis. Br J Rheumatol. 1997;36(10):1075–81.

    Article  CAS  PubMed  Google Scholar 

  33. Sarkin JA, Johnson SS, Prochaska JO, Prochaska JM. Applying the transtheoretical model to regular moderate exercise in an overweight population: validation of a stages of change measure. Prev Med. 2001;33(5):462–9.

    Article  CAS  PubMed  Google Scholar 

  34. de Vet HC, Terwee CB, Knol DL, Bouter LM. When to use agreement versus reliability measures. J Clin Epidemiol. 2006;59(10):1033–9.

    Article  PubMed  Google Scholar 

  35. Delbaere K, Hauer K, Lord SR. Evaluation of the incidental and planned activity questionnaire (IPEQ) for older people. Br J Sports Med. 2010;44(14):1029–34.

    Article  CAS  PubMed  Google Scholar 

  36. Chu AH, Ng SH, Koh D, Muller-Riemenschneider F. Reliability and validity of the self- and interviewer-administered versions of the global physical activity questionnaire (GPAQ). PLoS One. 2015;10(9):e0136944.

    Article  PubMed  PubMed Central  CAS  Google Scholar 

  37. Vuillemin A, Oppert JM, Guillemin F, Essermeant L, Fontvieille AM, Galan P, et al. Self-administered questionnaire compared with interview to assess past-year physical activity. Med Sci Sports Exerc. 2000;32(6):1119–24.

    Article  CAS  PubMed  Google Scholar 

  38. Mujika I, Padilla S. Muscular characteristics of detraining in humans. Med Sci Sports Exerc. 2001;33(8):1297–303.

    Article  CAS  PubMed  Google Scholar 

  39. Pettee Gabriel K, McClain JJ, High RR, Schmid KK, Whitfield GP, Ainsworth BE. Patterns of accelerometer-derived estimates of inactivity in middle-age women. Med Sci Sports Exerc. 2012;44(1):104–10.

    Article  PubMed  Google Scholar 

  40. Rehn TA, Munkvik M, Lunde PK, Sjaastad I, Sejersted OM. Intrinsic skeletal muscle alterations in chronic heart failure patients: a disease-specific myopathy or a result of deconditioning? Heart Fail Rev. 2012;17(3):421–36.

    Article  CAS  PubMed  Google Scholar 

  41. Bickel CS, Cross JM, Bamman MM. Exercise dosing to retain resistance training adaptations in young and older adults. Med Sci Sports Exerc. 2011;43(7):1177–87.

    Article  PubMed  Google Scholar 

  42. Reimers AK, Mess F, Bucksch J, Jekauc D, Woll A. Systematic review on measurement properties of questionnaires assessing the neighbourhood environment in the context of youth physical activity behaviour. BMC Public Health. 2013;13:461.

    Article  PubMed  PubMed Central  Google Scholar 

  43. Silsbury Z, Goldsmith R, Rushton A. Systematic review of the measurement properties of self-report physical activity questionnaires in healthy adult populations. BMJ Open. 2015;5(9):e008430.

    Article  PubMed  PubMed Central  Google Scholar 

  44. Terwee CB, Roorda LD, Knol DL, De Boer MR, De Vet HC. Linking measurement error to minimal important change of patient-reported outcomes. J Clin Epidemiol. 2009;62(10):1062–7.

    Article  PubMed  Google Scholar 

  45. Nadalin V, Bentvelsen K, Kreiger N. Reliability of self-reports: data from the Canadian multi-Centre osteoporosis study (CaMos). Chronic Dis Canada. 2004;25(2):28–31.

    Google Scholar 

  46. Merom D, Delbaere K, Cumming R, Voukelatos A, Rissel C, Van Der Ploeg HP, et al. Incidental and planned exercise questionnaire for seniors: validity and responsiveness. Med Sci Sports Exerc. 2014;46(5):947–54.

    Article  PubMed  Google Scholar 

  47. Pettee Gabriel K, McClain JJ, Lee CD, Swan PD, Alvar BA, Mitros MR, et al. Evaluation of physical activity measures used in middle-aged women. Med Sci Sports Exerc. 2009;41(7):1403–12.

    Article  PubMed  Google Scholar 

  48. Kriska AM, Knowler WC, LaPorte RE, Drash AL, Wing RR, Blair SN, et al. Development of questionnaire to examine relationship of physical activity and diabetes in Pima Indians. Diabetes Care. 1990;13(4):401–11.

    Article  CAS  PubMed  Google Scholar 

  49. Kriska AM, Edelstein SL, Hamman RF, Otto A, Bray GA, Mayer-Davis EJ, et al. Physical activity in individuals at risk for diabetes: diabetes prevention program. Med Sci Sports Exerc. 2006;38(5):826–32.

    Article  PubMed  PubMed Central  Google Scholar 

  50. Schulz LO, Harper IT, Smith CJ, Kriska AM, Ravussin E. Energy intake and physical activity in Pima Indians: comparison with energy expenditure measured by doubly-labeled water. Obes Res. 1994;2(6):541–8.

    Article  CAS  PubMed  Google Scholar 

  51. Gill DP, Jones GR, Zou GY, Speechley M. The phone-FITT: a brief physical activity interview for older adults. J Aging Phys Act. 2008;16(3):292–315.

    Article  PubMed  Google Scholar 

  52. Colbert LH, Matthews CE, Havighurst TC, Kim K, Schoeller DA. Comparative validity of physical activity measures in older adults. Med Sci Sports Exerc. 2011;43(5):867–76.

    Article  PubMed  PubMed Central  Google Scholar 

  53. Dipietro L, Caspersen CJ, Ostfeld AM, Nadel ER. A survey for assessing physical activity among older adults. Med Sci Sports Exerc. 1993;25(5):628-42.

    Article  PubMed  Google Scholar 

  54. Gennuso KP, Matthews CE, Colbert LH. Reliability and validity of 2 self-report measures to assess sedentary behavior in older adults. J Phys Act Health. 2015;12(5):727–32.

    Article  PubMed  Google Scholar 

  55. Harada ND, Chiu V, King AC, Stewart AL. An evaluation of three self-report physical activity instruments for older adults. Med Sci Sports Exerc. 2001;33(6):962–70.

    Article  CAS  PubMed  Google Scholar 

  56. Kolbe-Alexander TL, Lambert EV, Harkins JB, Ekelund U. Comparison of two methods of measuring physical activity in south African older adults. J Aging Phys Act. 2006;14(1):98–114.

    Article  PubMed  Google Scholar 

  57. Moore DS, Ellis R, Allen PD, Cherry KE, Monroe PA, O'Neil CE, et al. Construct validation of physical activity surveys in culturally diverse older adults: a comparison of four commonly used questionnaires. Res Q Exerc Sport. 2008;79(1):42–50.

    PubMed  Google Scholar 

  58. Brown WJ, Trost SG, Bauman A, Mummery K, Owen N. Test-retest reliability of four physical activity measures used in population surveys. J Sci Med Sport. 2004;7(2):205–15.

    Article  CAS  PubMed  Google Scholar 

  59. Brown W, Bauman A, Chey T, Trost S, Mummery K. Comparison of surveys used to measure physical activity. Aust N Z J Public Health. 2004;28(2):128–34.

    Article  PubMed  Google Scholar 

  60. Creamer M, Bowles HR, von Hofe B, Pettee Gabriel K, Kohl HW 3rd, Bauman A. Utility of computer-assisted approaches for population surveillance of physical activity. J Phys Act Health. 2014;11(6):1111–9.

    Article  PubMed  Google Scholar 

  61. Mahabir S, Baer DJ, Giffen C, Clevidence BA, Campbell WS, Taylor PR, et al. Comparison of energy expenditure estimates from 4 physical activity questionnaires with doubly labeled water estimates in postmenopausal women. Am J Clin Nutr. 2006;84(1):230–6.

    Article  CAS  PubMed  Google Scholar 

  62. Rauh MJ, Hovell MF, Hofstetter CR, Sallis JF, Gleghorn A. Reliability and validity of self-reported physical activity in Latinos. Int J Epidemiol. 1992;21(5):966–71.

    Article  CAS  PubMed  Google Scholar 

  63. Washburn RA, Smith KW, Goldfield SR, McKinlay JB. Reliability and physiologic correlates of the Harvard alumni activity survey in a general population. J Clin Epidemiol. 1991;44(12):1319–26.

    Article  CAS  PubMed  Google Scholar 

  64. Masse LC, Fulton JE, Watson KB, Tortolero S, Kohl HW 3rd, Meyers MC, et al. Comparing the validity of 2 physical activity questionnaire formats in African-American and Hispanic women. J Phys Act Health. 2012;9(2):237–48.

    Article  PubMed  Google Scholar 

  65. Ahn GE, Chmiel JS, Dunlop DD, Helenowski IB, Semanik PA, Song J, et al. Self-reported and objectively measured physical activity in adults with systemic lupus erythematosus. Arthritis Care Res (Hoboken). 2015;67(5):701–7.

    Article  Google Scholar 

  66. Garriguet D, Tremblay S, Colley RC. Comparison of physical activity adult questionnaire results with accelerometer data. Health Rep. 2015;26(7):11–7.

    PubMed  Google Scholar 

  67. Ainsworth BE, Macera CA, Jones DA, Reis JP, Addy CL, Bowles HR, et al. Comparison of the 2001 BRFSS and the IPAQ physical activity questionnaires. Med Sci Sports Exerc. 2006;38(9):1584–92.

    Article  PubMed  Google Scholar 

  68. Moy KL, Scragg RK, McLean G, Carr H. The New Zealand physical activity questionnaires: validation by heart-rate monitoring in a multiethnic population. J Phys Act Health. 2008;5(Suppl 1):S45–61.

    Article  PubMed  Google Scholar 

  69. Dinger MK, Oman RF, Taylor EL, Vesely SK, Able J. Stability and convergent validity of the physical activity scale for the elderly (PASE). J Sports Med Phys Fitness. 2004;44(2):186–92.

    CAS  PubMed  Google Scholar 

  70. Johansen KL, Painter P, Kent-Braun JA, Ng AV, Carey S, Da Silva M, et al. Validation of questionnaires to estimate physical activity and functioning in end-stage renal disease. Kidney Int. 2001;59(3):1121–7.

    Article  CAS  PubMed  Google Scholar 

  71. Washburn RA, Smith KW, Jette AM, Janney CA. The physical activity scale for the elderly (PASE): development and evaluation. J Clin Epidemiol. 1993;46(2):153–62.

    Article  CAS  PubMed  Google Scholar 

  72. Pettee Gabriel K, McClain JJ, Schmid KK, Storti KL, Ainsworth BE. Reliability and convergent validity of the past-week modifiable activity questionnaire. Public Health Nutr. 2011;14(3):435–42.

    Article  PubMed  Google Scholar 

  73. Albanes D, Conway JM, Taylor PR, Moe PW, Judd J. Validation and comparison of eight physical activity questionnaires. Epidemiology (Cambridge, Mass). 1990;1(1):65–71.

    Article  CAS  Google Scholar 

  74. Blair SN, Haskell WL, Ho P, Paffenbarger RS Jr, Vranizan KM, Farquhar JW, et al. Assessment of habitual physical activity by a seven-day recall in a community survey and controlled experiments. Am J Epidemiol. 1985;122(5):794–804.

    Article  CAS  PubMed  Google Scholar 

  75. Conway JM, Seale JL, Jacobs DR Jr, Irwin ML, Ainsworth BE. Comparison of energy expenditure estimates from doubly labeled water, a physical activity questionnaire, and physical activity records. Am J Clin Nutr. 2002;75(3):519–25.

    Article  CAS  PubMed  Google Scholar 

  76. Garfield BE, Canavan JL, Smith CJ, Ingram KA, Fowler RP, Clark AL, et al. Stanford seven-day physical activity recall questionnaire in COPD. Eur Respir J. 2012;40(2):356–62.

    Article  PubMed  Google Scholar 

  77. Gross LD, Sallis JF, Buono MJ, Roby JJ, Nelson JA. Reliability of interviewers using the seven-day physical activity recall. Res Q Exerc Sport. 1990;61(4):321–5.

    Article  CAS  PubMed  Google Scholar 

  78. Irwin ML, Ainsworth BE, Conway JM. Estimation of energy expenditure from physical activity measures: determinants of accuracy. Obes Res. 2001;9(9):517–25.

    Article  CAS  PubMed  Google Scholar 

  79. Sallis JF, Haskell WL, Wood PD, Fortmann SP, Rogers T, Blair SN, et al. Physical activity assessment methodology in the Five-City project. Am J Epidemiol. 1985;121(1):91–106.

    Article  CAS  PubMed  Google Scholar 

  80. Taylor CB, Coffey T, Berra K, Iaffaldano R, Casey K, Haskell WL. Seven-day activity and self-report compared to a direct measure of physical activity. Am J Epidemiol. 1984;120(6):818–24.

    Article  CAS  PubMed  Google Scholar 

  81. Washburn RA, Jacobsen DJ, Sonko BJ, Hill JO, Donnelly JE. The validity of the Stanford seven-day physical activity recall in young adults. Med Sci Sports Exerc. 2003;35(8):1374–80.

    Article  PubMed  Google Scholar 

  82. Williams E, Klesges RC, Hanson CL, Eck LH. A prospective study of the reliability and convergent validity of three physical activity measures in a field research trial. J Clin Epidemiol. 1989;42(12):1161–70.

    Article  CAS  PubMed  Google Scholar 

  83. Betz HH, Myers J, Jaffe A, Smith K, Dalman R. Reproducibility of the veterans physical activity questionnaire in an elderly population. J Phys Act Health. 2015;12(3):376–81.

    Article  PubMed  Google Scholar 

Download references

Acknowledgements

The authors would like to thank Dr. Peter Fowler and Dr. Colette Thomas for their assistance with screening and compilation of full text articles.

Funding

The authors did not receive funding for this systematic review.

Availability of data and materials

All data are present in the tables and figures.

Author information

Authors and Affiliations

Authors

Contributions

KD contributed to the concept, design of work, acquisition of analysis, interpretation of data and drafting of work and approved the submitted versions and agreed both to be personally accountable for the author’s own contributions and to ensure that questions related to the accuracy or integrity of any part of the work, even ones in which the author was not personally involved, are appropriately investigated, resolved, and the resolution documented in the literature. RS contributed to the concept, design of work, interpretation of data and drafting and revision of work and approved the submitted versions and agreed both to be personally accountable for the author’s own contributions and to ensure that questions related to the accuracy or integrity of any part of the work, even ones in which the author was not personally involved, are appropriately investigated, resolved, and the resolution documented in the literature. LP contributed to acquisition of data, analysis of data and drafting of work and approved the submitted versions and agreed both to be personally accountable for the author’s own contributions and to ensure that questions related to the accuracy or integrity of any part of the work, even ones in which the author was not personally involved, are appropriately investigated, resolved, and the resolution documented in the literature. RC contributed to the concept, design of work, interpretation of data and drafting and revision of work and approved the submitted versions and agreed both to be personally accountable for the author’s own contributions and to ensure that questions related to the accuracy or integrity of any part of the work, even ones in which the author was not personally involved, are appropriately investigated, resolved, and the resolution documented in the literature. All authors read and approved the final manuscript.

Corresponding author

Correspondence to Kenji Doma.

Ethics declarations

Ethics approval and consent to participate

Ethics approval was not applicable as this manuscript was a systematic review.

Consent for publication

Not applicable

Competing interests

The authors declare that they have no competing interests.

Publisher’s Note

Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Rights and permissions

Open Access This article is distributed under the terms of the Creative Commons Attribution 4.0 International License (http://creativecommons.org/licenses/by/4.0/), which permits unrestricted use, distribution, and reproduction in any medium, provided you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons license, and indicate if changes were made. The Creative Commons Public Domain Dedication waiver (http://creativecommons.org/publicdomain/zero/1.0/) applies to the data made available in this article, unless otherwise stated.

Reprints and permissions

About this article

Check for updates. Verify currency and authenticity via CrossMark

Cite this article

Doma, K., Speyer, R., Parsons, L.A. et al. Comparison of psychometric properties between recall methods of interview-based physical activity questionnaires: a systematic review. BMC Med Res Methodol 19, 43 (2019). https://doi.org/10.1186/s12874-019-0684-1

Download citation

  • Received:

  • Accepted:

  • Published:

  • DOI: https://doi.org/10.1186/s12874-019-0684-1

Keywords