 Research article
 Open Access
 Open Peer Review
 Published:
Fitting parametric random effects models in very large data sets with application to VHA national data
BMC Medical Research Methodology volume 12, Article number: 163 (2012)
Abstract
Background
With the current focus on personalized medicine, patient/subject level inference is often of key interest in translational research. As a result, random effects models (REM) are becoming popular for patient level inference. However, for very large data sets that are characterized by large sample size, it can be difficult to fit REM using commonly available statistical software such as SAS since they require inordinate amounts of computer time and memory allocations beyond what are available preventing model convergence. For example, in a retrospective cohort study of over 800,000 Veterans with type 2 diabetes with longitudinal data over 5 years, fitting REM via generalized linear mixed modeling using currently available standard procedures in SAS (e.g. PROC GLIMMIX) was very difficult and same problems exist in Stata’s gllamm or R’s lme packages. Thus, this study proposes and assesses the performance of a meta regression approach and makes comparison with methods based on sampling of the full data.
Data
We use both simulated and real data from a national cohort of Veterans with type 2 diabetes (n=890,394) which was created by linking multiple patient and administrative files resulting in a cohort with longitudinal data collected over 5 years.
Methods and results
The outcome of interest was mean annual HbA1c measured over a 5 years period. Using this outcome, we compared parameter estimates from the proposed random effects meta regression (REMR) with estimates based on simple random sampling and VISN (Veterans Integrated Service Networks) based stratified sampling of the full data. Our results indicate that REMR provides parameter estimates that are less likely to be biased with tighter confidence intervals when the VISN level estimates are homogenous.
Conclusion
When the interest is to fit REM in repeated measures data with very large sample size, REMR can be used as a good alternative. It leads to reasonable inference for both Gaussian and nonGaussian responses if parameter estimates are homogeneous across VISNs.
Background
Many translational research projects are generating very large data sets (VLDS) which require fitting complex models to answer questions of public health interest. Datasets can be considered “very large” because of large numbers of study subjects or units of analysis and/or large numbers of variables, and both situations present challenges during the analysis phase, especially when observations are clustered at some level (eg. Longitudinal data). An example of VLDS with large number of observations is a twoyear group randomized trial designed to assess the impact of a quality improvement intervention on colorectal cancer screening in primary care practices. Electronic medical record data were obtained from a sample of 68,150 patients from 32 primary care practices in 19 US states, followed monthly over a 2year time period [1]. Similarly, an example of VLDS with large number of variables as well as units of analysis is an functional magnetic resonance imaging study of neural changes underlying speechperception training [2] in which whole brain images of 40 patients were taken to make functional inference, resulting in hundreds of time series data clustered within thousands of voxels.
Fitting complex models for these types of data sets can be difficult, requiring inordinate amounts of computer time for parameter estimation, requiring memory allocations beyond what are available or containing data structures that prevent model convergence, even within stateoftheart computational infrastructures of medium size research facilities such as ours. For instance, fitting complicated generalized linear mixed models (GLMMs) for data from the examples above using software such as SAS 9.2.2 (Cary, NC), Stata 11 (College Station, TX) or R (R2.11.1) may not be possible using desktop computers typically available to researchers within our institutions (64 bit server with 12GB and 667MHz dual ranked DIMMS and 48GB of RAM). Although a few methods for modeling VLDSs exist, current practice mainly involves data reduction processes, which usually result in loss of information.
Recently, we have been working on a longitudinal study of the trajectory of HbA1c control in patients with type2 diabetes treated within the Veterans Administration (VA) healthcare setting, and we have been faced with the problem of fitting GLMMs on over 890,000 patients, clustered in 23 Veterans Integrated Service Networks (VISNs) and followed over 5 years. Fitting mixed effects logistic regression model with over 30 covariates for making individual level inference resulted in an out of memory error using a 64 bit server with 12GB and 667MHz dual ranked DIMMS and 48GB of RAM.
In SAS procedures such as Proc GLIMMIX, fitting mixed effect models with the recommended standard syntax of including subject ID in a Class statement was not possible. This procedure with the standard syntax ran out of memory when we attempted to fit a model with the simplest scenario of including a random intercept. With adhoc modifications (see discussion section) to the standard syntax, however, we were able to fit the model despite it took longer time. Similar problems were observed in Stata’s gllamm, and R’s lme4 packages.
With the current focus on personalized medicine, patient/subject level inference is often of key interest in translational research. GLMMs are a very rich class of models that are traditionally used to make such individuallevel inference by breaking down the total variation in the observed response into withinsubject and betweensubject variation. These models are also used to incorporate natural heterogeneity in the estimates due to unmeasured explanatory variables [3–5]. In GLMMs, the joint distribution of the vector of responses is fully specified and the withinsubject association among repeated or clustered measures is induced via incorporation of one or more random effects into the model. As a result, interpretation of the regression coefficients for GLMM relies on the induced model for the covariance among the responses. When population level inference is of interest, marginal models (e.g. general linear models) are often used, and withinsubject association among repeated responses is incorporated by directly making assumptions about the covariance (e.g. autoregressive, compound symmetry, etc). While such models may not be as difficult to fit with VLDSs, subjectlevel inference cannot be made using the marginal model framework since the mean response and covariance are modeled separately [3]. Currently, methodology for fitting parametric mixed effect models for VLDSs is underdeveloped.
There are some recent Bayesian methods proposed for fitting parametric random effects models to VLDSs [6–8]. Owen [9] and Huang and Gelman [7] propose a computational strategy, akin to a Bayesian meta regression, based on sampling the data, computing separate posterior distributions based on each sample, and then combining these to get a consensus posterior inference. Their approach reduces the number of parameters as well as sample size for each separate model fit and can lead to efficient inference.
An alternative is a 2stage “data squashing” method [10]. In this method, the complete data is partitioned into compact subregions in the first stage. Then one generates a set of “pseudodata” and weights within each subregion so that the weighted moments on the squashed data match the unweighted moments on the original data. This method is less sensitive to outliers than random sampling, but it has the potential to be computationally intensive. To date, its characteristics are only known in simpler fixed effect and descriptive models. Madigan et al. [11, 12] proposed a data squashing method which first groups subjects based on their contribution to the likelihood and then fits models to the mean of each group. Although this approach may be promising for some models, it is unwieldy under the Dirichlet process prior (DPP) due the complicated structure of the likelihood [12]. In general, the Bayesian approaches which use DPP to automatically cluster individuals into latent classes [13, 14] may not be feasible in very large data sets due to limitations in current Markov chain Monte Carlo (MCMC) algorithms [12, 15].
Motivated by the scarcity of work in this area and the challenge we faced with the analysis of our VLDS, we propose a random effects meta regression (REMR) approach in which VISNspecific estimates are combined via meta regression. We make comparisons with two other approaches, (1) average estimates from analysis of 1000 data sets obtained via simple random sampling (SRS) of the original data with simulated 95% confidence intervals (CIs), (2) weighted average estimates from analysis of 1000 data sets obtained via VISNstratified random sampling (StRS) with simulated 95% CIs. Using simulated data, we also assess biases present within each approach, noting whether they provide equivalent inferences as would be obtained from analysis of the full data. The paper is organized as follows: section 2 presents the motivating example; section 3 describes the details of the statistical methods; section 4 presents the results of the analysis; and section 5 discusses the findings.
Motivating example
A national cohort of Veterans with type 2 diabetes was created by linking patient and administrative files from the Veterans Health Administration (VHA) National Patient Care and Pharmacy Benefits Management (PBM) databases. Veterans were included in the cohort if they had type 2 diabetes defined by two or more International Classification of Diseases, Ninth Revision (ICD9) codes for diabetes (250, 357.2, 362.0, and 366.41) in the previous 24 months (2000 and 2001) and during 2002 from inpatient stays and/or outpatient visits on separate days (excluding codes from lab tests and other nonclinician visits), and prescriptions for insulin or oral hypoglycemic agents (VA classes HS501 or HS502, respectively) in 2002 [16]. Veterans identified as having type 2 diabetes by ICD9 codes were excluded from the cohort if they did not have prescriptions for diabetic medications (HS501 or HS502) in 2002. The datasets were linked using patient scrambled Social Security Numbers and resulted in 890, 394 Veterans, who were followed until death, loss to followup, or through December 2006. The study was approved by our Institutional Review Board and local VA Research Development committee.
Outcome measure
The primary outcome was glycosylated hemoglobin (HbA1c) level. In addition, a binary outcome defined as HbA1c ≥ 8.0% was used.
Primary independent variable
For this project, the primary research question was whether HbA1c differed significantly by race/ethnicity, classified as nonHispanic white (NHW), nonHispanic black (NHB), Hispanic, and other/unknown/missing.
Demographic variables
Age, gender, marital status (i.e., single or married) and percentage serviceconnectedness (i.e., degree of disability due to illness or injury that was aggravated by or incurred in military service) were available and treated as covariates in the model. Location of residence was defined as urban and rural/highly rural, [17] and hospital region was defined by the five geographic regions of the country based on VHA Veteran’s Integrated Service Networks (VISNs): Northeast (VISNs 1, 2, & 3), MidAtlantic (VISNs 4, 5, 6, 9, & 10), South (VISNs 7, 8, 16, & 17), Midwest (VISNs 11, 12, 15, 19, & 23), and West (VISNs 18, 20, 21, & 22) [18].
Comorbidity
Variables included substance abuse, anemia, cancer, cerebrovascular disease, congestive heart failure, cardiovascular disease, depression, hypertension, hypothyroidism, liver disease, lung disease, fluid and electrolyte disorders, obesity, psychoses, peripheral vascular disease, and other (AIDS, rheumatoid arthritis, renal failure, peptic ulcer disease and bleeding, weight loss) and were defined based on ICD9 codes at entry into the cohort. In our final models, we included a categorical summary of count of comorbidities defined as (0=none, 1=one, 2=two 3=three or more), a process which has been shown to be as or more efficient than more complicated algorithms [19].
Methods
Overview of the generalized linear mixed model (GLMM)
To model the relationship between HbA1c (Y) and covariates (X), a GLMM approach was used. For the ith subject (i=1,.,N) with n_{i} (j=1,…,n_{i}) repeated measurements, we considered the model, E(Y_{i} X_{i},Z_{i}) =g^{1}(X_{i}β + Z_{i}b_{i}), where g is a monotone link function and Y_{i} is Nx1 vector of responses, X_{i} is n_{i}xp matrix of covariates, Z_{i} is n_{i}xq matrix of covariates (q≤p), β is a px1 vector of fixed effect parameters, b_{i} is a qx1 vector of random effects. We assume that b_{i}~N(O,G), where G is a qxq covariance matrix for b_{i}. An identity link function results in a linear mixed model for the continuous HbA1c outcome, and a logit link results in logistic mixed effects model for the dichotomous HbA1c outcome. If b_{i} is a vector of random intercept and slope, it results in a 2x2 covariance matrix G which indicate natural heterogeneity among individuals in both their baseline level and changes in the expected outcomes over time. In our models, a personlevel random effect was included in all models to account for withinindividual correlations. This approach accommodates a wide range of distributional assumptions, multilevel data, measurement of subjects at different time points, modeling individual level effects, missing data, and time varying or invariant covariates [20].
A special case is the linear mixed model given by, Y_{i} X_{i},Z_{i} =X_{i}β + Z_{i}b_{i} + e_{i}, where e_{i}~N(0,R_{i}) and independent of b_{i}. Assuming, R_{i}=σ^{2}I_{ni}, the conditional distribution of Y_{i}b_{i} is given by the multivariate Gaussian distribution with mean Z_{i}b_{i} +X_{i}β and variance σ^{2}I_{ni}. In this model, the response for the i^{th} subject is assumed to differ from the population mean, E(Y_{i})=X_{i}β, by a subject specific effect, Z_{i}b_{i}, and a withinsubject measurement error e_{i}. The estimates of the parameters in a mixed model are determined as the values that optimize an objective function which is either the likelihood of the parameters given the observed data (ML) or a related objective function called the restricted ML (REML). In practice REML is often preferred. The loglikelihood based on the observed data assuming that the vector of all variance components in G and R_{i} can be denoted by α can be written as,
where Σ _{ i }(α) = Z _{ i } G(α)Z _{ i } ^{′} + R _{ i }(α) and parameter estimates are obtained via NewtonRaphson.
Weighted Generalized Linear Mixed Effects Model (WGLMM)
The WGLMM is a model wellsuited for analysis of survey sampled data. We use it to analyze our type2 diabetes cohort data in the context of finitely sampled data (e.g. VISNstratified randomly sampled data).
In sample surveys, units are sometimes drawn with unequal selection probabilities, and if the design probabilities are informative (i.e. they are related to the response) [21, 22] the model holding for the sample will be different from the model holding for the finite population. Consequently, the usual estimators will be biased for the finite population quantity [23, 24]. A common design based solution is to use weighted estimators where the contribution of unit i is weighted by the inverse probability of selection into the sample [25, 26]. An alternative is a model based approach where the sampling units enter into the model as random effect terms. In the former, a pseudolikelihood approach for accommodating inverse probability weights is implemented by using adaptive quadrature, and a sandwich estimator is used to obtain standard errors that account for complex sampling, since modelbased standard error estimates may not be valid. These types of models can be fitted using SAS procedures such as Proc GLIMMIX by including weights via the WEIGHT statement. Similarly, xtmixed and gllamm can be used in Stata and the lme4 package in R. In the Stata program gllamm, a full pseudomaximumlikelihood estimation which allows for specification of probability weights, is implemented via adaptive quadrature [27]. The weights enter the logpseudolikelihood as if they were frequency weights, representing the number of times that each unit should be replicated. Adaptive quadrature [27, 28] provides good approximations to the integrals in the pseudolikelihood. However, these are not easy to implement in practice since the logpseudolikelihood cannot simply use one set of weights based on the overall inclusion probabilities but must use separate weights at for the fixed and random effects.
The generalizations of Equation (1) above to the special case of weighted linear mixed model (WLMM) can be described by the change in the conditional distribution (Y_{i}b_{i})~N( Z_{i}b_{i} +X_{i}β ;R_{i}(α)W_{i} ^{1}) and Σ _{ i }(α) = Z _{ i } ^{′} W _{ i } ^{− 1} G(α)Z _{ i } + R _{ i }(α)W _{ i } ^{− 1} where the weights matrix, W, is constant. In our implementation of WLMM, we used a residual subject specific pseudolikelihood (RSPL) to estimate parameters and simulated 95% CIs which are calculated based on 1000 simulations.
Meta regression approach
Another approach to deal with fitting parametric random effects models to VLDS is to do aggregated analysis after estimating the parameters at some level of administrative or sampling based subsets of the data. This can lead to substantial gain in the time required to fit these models and can be adapted to parallel processing, leading to further computational time savings. In the case of likelihood inference, this idea leads to a pseudolikelihood [29–31], where weights are incorporated as if they were frequency weights. The resulting estimator is design consistent and hence model consistent under suitable regularity conditions such as those discussed by Isaki and Fuller [32]. However, this consistency typically comes at a price of reduced efficiency [33].
Since VHA research data are provided at VISN level, models for each VISN can be fitted, and a mechanism to combine these parameter estimates is suggested. After models relating HbA1c and covariates are fitted for each VISN, the next step is to use pooling methods to obtain national estimates. This can be done using fixed effects [34, 35] and random effects meta regression [36–38]. The fixed effects approach [39] is based on weighted regression with the weight being the number of patients in each VISN as fraction of the total population. Covariates can either be VISN level factors or aggregates of individual level covariates. The second approach is random effects metaregression [35]. Generally the regression coefficients are regressed on an intercept and VISNlevel covariates. A random intercept is included in the regression to take into account the betweenVISN variation. This leads to the usual DerSimonian and Laird [40] random effects estimate of the pooled regression coefficient.
Fixed effects meta regression (FEMR)
Let ψ_{i} be an effect of interest to estimate for VISN i. In our study these are the regression coefficients associated with covariates such as race/ethnicity in the GLMM model. Let ϕ_{i} be the corresponding sample estimate. The fixed effects meta regression can be given by ϕ_{i} =τ +ε_{i}, where τ is the pooled mean or FEMR estimate and ε_{i}~N (0, σ_{i} ^{2}) is the random error. This can be adjusted for covariates via weighted regression as:
Random effects meta regression (REMR)
In REMR, a standard onestep DerSimonian and Laird [40] random effects pooled estimate of the regression coefficient is obtained via, ϕ_{i} =τ +ν_{i} +ε_{i} , where ε_{i}~N (0, σ_{i} ^{2}) and ν_{i}~N (0, σ_{v} ^{2}) with v_{i} (VISN level random effect) and ε_{i} (random error) are uncorrelated. In the covariate adjusted version, the model is fitted by regressing the effect estimate on an intercept and VISNlevel covariates:
where τ is the pooled mean or REMR estimate. The adjustment covariates can be VISN level covariates (z) or aggregates of individual level covariates (x) to account for additional causes of heterogeneity [41]. For both fixed effects [34] and random effects meta regression [37, 38], we used restricted maximum likelihood via Proc GLIMMIX and Proc PLM (SAS 9.2.2) to obtain pooled estimates.
Summary of modelling strategies used
In this paper, we study two broad strategies for longitudinal analyses of VLDS: random effects meta regression (REMR) and estimation based on sampling of the full data (SRS and StRS). Within each strategy we model the continuous outcome of HbA1c using a linear mixed model and the binary outcome of HbA1c (<8% vs. ≥ 8%) using mixed effects logistic regression. The primary independent variable is race/ethnicity, and a number of subjectlevel covariates are included.
Test of homogeneity
The main goal of REMR and FEMR is to obtain a single global or pooled effect summarized across VISNs. But, obtaining pooled estimates assumes homogeneity of VISN level effects. According to [42], the main sources of heterogeneity are clinical incomparability or design incomparability. Clinical incomparability can be caused by differences in the VISN level populations, and design incomparability can be caused due to differences, for example, in missing data and measurement error. These are issues, however, that mainly arise in pooling of effects from different studies. In our data, these issues may not arise at all or will have very limited impact on generalizability of results. If heterogeneity is found or suspected to exist, the common approaches used in metaanalysis are (1) to stratify the studies into homogeneous subgroups and then fit a separate fixed effects estimate [43], (2) construct a random effects estimate [40] across all VISNs (a random effects approach incorporates both withinstudy and betweenstudy variability: if heterogeneity exists among VISNs, a summary measure across those VISNs may not be provided), or (3) fit a metaregression model that explains the heterogeneity in terms of VISNlevel covariates. We implemented an approach of removing outliers, as suggested by Draper and Smith [44] in conjunction with approach 2 above.
Model selection
Although the purpose of this project was not to “select” an optimal model, model fit assessment was facilitated using maximum likelihood (or pseudolikelihood) information criteria, factors typically used in model selection. Two common approaches in the literature include Akaike information criterion (AIC) [45] and Bayesian information criterion (BIC) [46]. Across competing models the lowest value on each criterion indicates the best fitting model. These statistics account for both model fit (deviance) and model complexity by penalizing models with a larger number of parameters. For GLMM, pseudoAIC and pseudoBIC, which are adjusted for fixed effects and covariance parameters, are used. However these pseudo information criteria are not useful for comparing models that differ in their pseudo data [3]. Thus in our example, they are used to compare AIC/BIC values of models estimated using the sampled data and to the full data models. AIC/BIC values from REMR are not comparable with those from the full data.
Bootstrap simulation study design
Simulation studies based on 1000 repeated resamplings of (sample size: 1%, 5%, 10% and 25%) the full data are used to asses and compare the methods discussed above. This is implemented via a nonparametric bootstrapping approach [47] with a repeated sample of the observed data with replacement with the study subject as the sampling unit to form 1000 simulated datasets of each sample size. Traditionally, MonteCarlo simulation studies based on data generated from statistical models have been used for this kind of comparative study. Resampling has the advantage that the data in resampled datasets are based on observations from real datasets and thus reflect the appropriate level of diversity and variability found in realistic populations [47, 48]. Our datasets are large enough to permit numerous samples of reasonable size to arrive at stable conclusions within the resampled data [47]. Performance of methods in terms of bias and efficiency will be benchmarked against results from the original full data. The full data estimates are used as true values to judge bias in estimates from the sampling and meta regression approaches.
Hardware
All analyses for this investigation were run on a Dell PowerEdge 2900 III server with two dual core Intel Xeon X5260 processors with 6 megabyte cache, with a clock speed of 3.33 gigahertz, and a frontside bus of 1333 megahertz. The server has been configured with 12 four gigabyte (GB), 667MHz dual ranked dual inline memory modules for a total of 48GB of RAM. Data are stored on six one terabyte (TB) 7200 revolution per minute nearline serial attached small computer system interface, 3GB per second 3 ½ inch HotPlug hard drives forming a 3TB redundant array of independent disk level 5 storage system. This server runs a 64bit version of Windows 2003 R2 Enterprise X64 Edition Service Pack 2 operating system.
Software
Datasets were organized for this study using SAS version 9.2.2 (Cary, NC) and SAS transport data sets created. Data were read into a 64bit version of R for Windows 2.11.1 (R Development Core Team 2010) using the “Hmisc” [49] and “foreign” [50] packages. Plots were done using R, and regression analysis was accomplished using the lme4 package [51].
Results
The full cohort consisted of 890,394 Veterans with diabetes followed from 2002 through 2006. The cohort is characterized based on demographics, HbA1c and comorbidities in Table 1. The full cohort is 62% NHW, 12% NHB, 14% Hispanic, and 12% other race/ethnicity. Characteristics of the 1%, 5%, 10% and 25% random samples estimates as well as REMR reflect that of the full cohort, with prevalence estimates across the four sampling scenarios (1%, 5%, 10% and 25%) and the full cohort differing by no more than 1%. The mean age in the full cohort is 66.2 years and ranges from 66.2 years in the 10% and 25% SRS to 66.5 years in the 1% SRS.
Linear mixed models were used to model continuous HbA1c levels in both the full cohort and each of the SRS and StRSs (using the weighted approach described in section 3.2). Parameter estimates associated with betas representing the different race/ethnic and comorbidity groupings, standard errors of the betas, and 95% confidence intervals are reported in Tables 2. For all variables examined with the exception of comorbidity, the direction of the association between each independent variable and HbA1c was consistent in both SRS and StRS with the full cohort. In fact, the beta estimates for all the 1% to 5% SRSs or StRS scenarios were within the 95% confidence limits of the full cohort estimates (1% estimates are removed from Tables to save space). For example, the beta estimate for NHB indicated that HbA1c levels were 0.46 units higher in NHB than NHW in the full cohort, while estimates in the SRSs ranged from 0.39 in the 1% SRS to 0.46 in the 25% SRS. The estimates for NHB in the StRS also ranged from 0.44 in the 1% sample to 0.47 in the 25% sample. However, for the number of comorbidities variable, while the direction of the beta estimate remained similar between the full cohort and the SRS or StRS, the magnitude and 95% CI for the 1% sample was higher, and the association was not statistically significant (Table 2). For example, the β estimate (95% confidence intervals) for three or more comorbidities was 0.11 (0.11, 0.13) for the full cohort and it ranged between 0.04 (0.05, 0.16) in the 1% SRS to 0.10 (0.08, 0.13) in the 25% SRS and 0.18 (0.08, 0.30) in the 1% StRS to 0.11 (0.1, 0.15) in the 25% StRS. The smaller standard errors and tighter confidence intervals observed in the full cohort are a direct result of the larger sample size.
The REMR estimates on the other hand are very close to the full sample estimates. For example, the beta estimate for NHB indicated that HbA1c levels were 0.45 (0.41, 0.49) higher in NHB than NHW in REMR which is comparable to 0.46 (0.45, 0.46) in the full cohort. Similarly, for three comorbidities the full cohort results were 0.11 (0.11, 0.13) while REMR resulted in 0.09 (0.05, 0.13). In all these models, the intercept was very well approximated even in the 1% sampled data. It should be noted that, REMR can be highly affected by outliers in the estimates that are aggregated to get the final estimates. In our case, VISN 13 and 14 exhibited extreme values and hence were removed to maintain the homogeneity assumption required by REMR in order to get unbiased estimates. In Table 2, the REMR estimates as well as their 95% CI are very similar to the full sample estimates reported in the first row of the table except that the full data 95% CI estimates are much tighter as expected. REMR models that included VISN 13 and 14 are summarized in Additional file 1: Appendix Tables 1 and 2. These appendix tables show that REMR can lead to biased estimates when the assumption of homogeneity of VISN level estimates is violated.
Table 3 compares results from GLMM for binary HbA1c for the three different approaches considered. While it was not possible using standard GLMMIX syntax to obtain estimates for the full sample or for samples of size 10% or more (due to well known computational convergence and memory limitation errors), we were able to obtain estimates for the full sample using tricks such as sorting by ID and time and removing them from the class statement. However, it was not possible to obtain estimates using R’s lme4 package. Compared to NHW, the REMR estimate (and 95% CIs) for the log odds of HbA1c>8% was 0.58(0.52, 0.64) for NHB, 0.11(0.05,0.17) for Hispanic and 0.32(0.26,0.38) for Other race/ethnic groups. These mimic the full sample estimates reported in the first row of Table 3 showing that REMR estimates (when homogeneity assumption is met) lead to unbiased estimates. On the other hand, the estimates from both SRS and StRS were very biased with bias increasing inversely with sample size. The corresponding estimates (and 95% CIs) of the logodds ratio in the 5% SRS were 1.69(1.40,1.99), 1.10(0.70,1.50) and 1.05(0.75,1.35) for NHB, Hispanic, and Other respectively. The results from StRS were similar to SRS. On the other hand, while the REMR logodds ratio estimate for having three or more comorbidities was 0.25(0.19,0.31), the corresponding estimates from 5% SRS and StRS were 0.33(0.24,0.90) and 0.80(0.22,1.39) respectively. The estimates with VISNs 13 and 14 included are in Additional file 1: Appendix Tables 1 and 2, indicating even worse performance by these sampling approaches.
Figure 1 depicts the beta estimates for race by VISN and the aggregated REMR estimate and their corresponding 95% CI in the association study of HbA1c and race after adjusting for covariates. Figure 2 shows corresponding estimates for binary HbA1c>8% using GLMM. Both figures indicate homogeneity in the estimates across VISNs except in VISNs 13 and 14 (which were removed to meet the homogeneity assumption required by REMR since both VISNs exhibited extreme estimated values). The outlier estimates from VISNs 13 and 14 are clearly depicted in Additional file 1: Appendix Figures S1 and S2. Figure 3 shows the goodness of fit statistics for making comparison among the different models. As expected the AIC and BIC values decreased as sample size increases. Also, AIC/BIC estimates for comparable models in StRS were smaller than those from SRS indicating better fit. While not comparable with the others, the weighted AIC/BIC values for REMR were the smallest indicating better fit.
Additional results corresponding to the analysis of the original full data such as the distribution of subjects in each VISN (Additional file 1: Appendix Table S3), goodness of fit statistics (Additional file 1: Appendix Table S4), how long each model took to fit (Additional file 1: Appendix Table S5) and full covariate models (Additional file 1: Appendix Tables S6 and S7) are in the appendix. Additional file 1: Appendix Table S5 shows that fitting LMM in R’s lme4 package required about 4 times longer time than SAS’s Proc MIXED. Also, while it was not possible to fit GLMM in R, fitting GLMM in Proc GLIMMIX took longer time than obtaining pooled estimates of parameters via REMR.
Discussion and conclusion
Models with random effects are useful for patient level inference just as marginal models are useful for population level inference. However, for very large data sets, it can be difficult to fit models with random effects using commonly available statistical software such as SAS. There are very few papers on this topic and the most recent work involves a 2stage Bayesian algorithm [12]. While their method has advantages, there are problems when one needs to adjust for multiple covariates, and it is not clear whether their approach will work in VLDS settings as large as ours.
This study assesses and compares REMR to two sampling based approaches using bootstrap simulation studies. Our results indicate that REMR provides parameter estimates that are less likely to be biased with smaller standard errors when the VISN level estimates are homogenous. The sampling approaches also provide parameter estimates that were equivalent to the full data estimates except when the outcome variable was binary. Thus, when the interest is to fit random effect models in repeated measures data with very large sample size, REMR may be used as a good alternative.
Some adhoc approaches can also be considered to ameliorate the challenges with the double optimization required when fitting GLMM to VLDS. For example, SAS Proc HPMIXED is developed to fit LMM to VLDS and provides computational advantages over Proc Mixed in certain situations. Also, sorting the data by variables that need to be in the CLASS statement of Proc MIXED or GLIMMIX, sorting by random effect subject identifiers, may also alleviate the computational burden. However, all of these methods can often not overcome the computational challenges with very large data sets, like those mentioned in the introduction, which makes REMR attractive.
One of the key problems with REMR is handling situations involving heterogeneous parameter estimates. For example, Additional file 1: Appendix Figures S4 and S5 show the estimates for VISNs 13 and 14 which are clearly outliers in the opposite direction. One approach is to remove these outliers, as we did, and obtain unbiased and efficient estimates. The estimates after removing the VISNs with outliers are in Figures 1 and 2. Another approach, illustrated in an extensive simulation study by Morton et al. [41, 41] is to incorporate important covariates at either the study or person level. However, despite the importance of including covariates, a model that includes a covariate that is an aggregate of a personlevel characteristic rather than a study characteristic may also produce biased results. The tradeoff between the biases of incorporating an aggregated covariate versus excluding it requires further exploration. While Bayesian random effects meta regression [10, 50, 51] may be is an alternative, it is not clear how these methods will work for VLDSs and is a topic of future work.
Our work demonstrates a variety of approaches that may be used in analyses of VLDSs, especially when observations are clustered such as in a longitudinal setting. Our simulation results show that SRS and StRS approaches appear to lead to reasonable parameter estimates with Gaussian responses but may be biased when responses are nonGaussian (eg. Binary). REMR may be an optimal strategy for both Gaussian and nonGaussian responses, especially when parameter estimates are homogeneous across clusters.
Abbreviations
 CI:

Confidence interval
 FEMR:

Fixed effects meta regression
 GLMM:

Generalized linear mixed model
 LMM:

Linear mixed model
 NHB:

Nonhispanic black
 NHW:

Nonhispanic white
 REM:

Random effect model
 REMR:

Random effects meta regression
 SRS:

Simple random sample
 StRS:

Stratified random sample
 VISN:

Veteran’s integrated service network
 VLDS:

Very large data sets
 VHA:

Veteran’s health administration
 WGLMM:

Weighted GLMM
 WLMM:

Weighted LMM.
References
 1.
Ornstein S, Nemeth LS, Jenkins RG, Nietert PJ: Colorectal cancer screening in primary care: translating research into practice. Medical care. 2010, 48 (10): 900906.
 2.
Eckert MA, Keren NI, Roberts DR, Calhoun VD, Harris KC: Agerelated changes in processing speed: unique contributions of cerebellar and prefrontal cortex. Front Hum Neurosci. 2010, 4: 10
 3.
Fitzmaurice GM, Laird NM, Ware JH: Applied Longitudinal Analysis. 2004, New York: John Wiley & Sons
 4.
Aitkin M, Anderson D, Francis B, Hinde J: Statistical modeling in GLIM. 1989
 5.
Breslow NE, Clayton DG: Approximate inference in generalized linear mixed models. J Am Stat Assoc. 1993, 88: 925.
 6.
Guha S, Ryan L: GaussSeidel estimation of generalized linear mixed models with application to Poisson modeling of spatially varying disease rates. 2006, Boston: MA: Harvard School of Public Health
 7.
Huang Z, Gelman A: Sampling for Bayesian computation with large datasets. SSRN eLibrary. 2005
 8.
Tao H, Palta M, Yandell BS, Newton MA: An estimation method for the semiparametric mixed effects model. Biometrics. 1999, 55 (1): 102110.
 9.
Owen A: Data squashing by empirical likelihood. Data Mining and Knowledge Discovery. 2003, 7: 101113.
 10.
DuMouchel WH: Bayesian metaanalysis. Statistical Methodology in the Pharmaceutical Sciences. Edited by: Berry DA. 1999, New York: Marcel Dekker
 11.
Madigan D, Raghavan N, DuMouchel W, Nason M, Posse C, Ridgeway G: Likelihoodbased data squashing: a modeling approach to instance construction. Data Mining and Knowledge Discovery. 2002, 6: 173190.
 12.
Pennell ML, Dunson DB: Fitting semiparametric random effects models to large data sets. Biostatistics. 2007, 8 (4): 821834.
 13.
Bush CA, Maceachern SN: A semiparametric Bayesian model for randomised block designs. Biometrika. 1996, 83 (2): 275285.
 14.
Kleinman KP, Ibrahim JG: A semiparametric Bayesian approach to generalized linear mixed models. Stat Med. 1998, 17 (22): 25792596.
 15.
Ishwaran H, James LF: Gibbs Sampling Methods for StickBreaking Priors. J Am Stat Assoc. 2001, 96 (453): 161173.
 16.
Miller DR, Safford MM, Pogach LM: Who has diabetes? Best estimates of diabetes prevalence in the Department of Veterans Affairs based on computerized patient data. Diabetes Care. 2004, 27 (Suppl 2): B1021.
 17.
West AN, Lee RE, ShambaughMiller MD, Bair BD, Mueller KJ, Lilly RS, Kaboli PJ, Hawthorne K: Defining “Rural” for Veterans’ Health Care Planning. J Rural Health. 2011, 26 (4): 301309.
 18.
ORD: Veterans Health Administration Field Research Advisory Committee Operating Procedure. 2004, Office of Research and Development (ORD)
 19.
Quan H, Sundararajan V, Halfon P, Fong A, Burnand B, Luthi JC, Saunders LD, Beck CA, Feasby TE, Ghali WA: Coding algorithms for defining comorbidities in ICD9CM and ICD10 administrative data. Medical care. 2005, 43 (11): 11301139.
 20.
Diggle PJ, Heagerty P, Liang KY, Zeger SL: Analysis of Longitudunal Data. 2002, Oxford, England: Oxford University Press, 2nd
 21.
Rubin DB: Inference and missing data. Biometrika. 1976, 63: 581592.
 22.
Little RJA: Models for nonresponse in sample surveys. J Am Stat Assoc. 1982, 77: 237250.
 23.
Pfeffermann D: The use of sampling weights for survey data analysis. Stat Methods Med Res. 1996, 5: 239261.
 24.
Pfeffermann D, Skinner CJ, Holmes DJ, Goldstein H, Rasbash J: Weighting for unequal selection probabilities in multilevel models. J R Stat Soc: Series B. 1998, 60: 2340.
 25.
Kish L: Survey Sampling. 1965, London: John Wiley & Sons
 26.
Cochran WG: Sampling Techniques. 1977, New York: John Wiley & Sons, 3rd
 27.
RabeHesketh S, Skrondal A, Pickles A: Maximum likelihood estiamtion of limited and discrete dependent variable models with nested random effects. J Econometrics. 2005, 128: 301323.
 28.
RabeHesketh S, Skrondal A, Pickles A: Reliable estimation of generalized linear mixed models using adaptive quadrature. The Stata Journal. 2002, 2 (1): 121.
 29.
Binder DA: On the variances of asymptotically normal estimators from complex surveys. Int Stat Rev. 1983, 51: 279292.
 30.
Chambers RL, Skinner CJ: Analysis of Survey Data. 2003, Chichester: John Wiley & Sons
 31.
Skinner CJ: Domain means, regression and multivariate analysis. Analysis of Complex Surveys. Edited by: Skinner CJ, Holt D, Smith TMF. 1989, Chichester: John Wiley & Sons, Inc
 32.
Isaki CT, Fuller WA: Survey design under the regression superpopulation model. J Am Stat Assoc. 1982, 77: 8996.
 33.
Binder DA, Roberts GR: Designbased and modelbased methods for estimating model parameters. Analysis of Survey Data. Edited by: Chambers RL, Skinner CJ. 2003, Chichester: John Wiley & Sons
 34.
Normand SL: Metaanalysis: formulating, evaluating, combining, and reporting. Stat Med. 1999, 18 (3): 321359.
 35.
Berkey CS, Hoaglin DC, Mosteller F, Colditz GA: A randomeffects regression model for metaanalysis. Stat Med. 1995, 14 (4): 395411.
 36.
Hartung J, Knapp G, Sinha BK: Statistical metaanalysis with applications. 2008, New York: John Wiley & Sons
 37.
Jackson C, Best N, Richardson S: Hierarchical related regression for combining aggregate and individual data in studies of socioeconomic disease risk factors. J R Stat Soc, Series A. 2008, 171: 159178.
 38.
van Houwelingen HC, Arends LR, Stijnen T: Advanced methods in metaanalysis: multivariate approach and metaregression. Stat Med. 2002, 21 (4): 589624.
 39.
Stuck AE, Siu AL, Wieland GD, Adams J, Rubenstein LZ: Comprehensive geriatric assessment: a metaanalysis of controlled trials. Lancet. 1993, 342 (8878): 10321036.
 40.
DerSimonian R, Laird N: Metaanalysis in clinical trials. Control Clin Trials. 1986, 7 (3): 177188.
 41.
Morton SC, Adams JL, Suttorp MJ, Shekelle PG: Metaregression approaches: What, Why, When, and How?. AHRQ Publication No 040033. 2004, Agency for Healthcare Research and Quality, Rockville (MD)
 42.
Thompson SG: Controversies in metaanalysis: the case of the trials of serum cholesterol reduction. Stat Methods Med Res. 1993, 2 (2): 173192.
 43.
Hardy RJ, Thompson SG: Detecting and describing heterogeneity in metaanalysis. Stat Med. 1998, 17 (8): 841856.
 44.
Draper NR, Smith H: Applied regression analysis. 1998, New York, NY: John Wiley & Sons, Inc
 45.
Akaike H: Information theory and an extension of the maximum likelihood principle. Second International Symposium on Information Theory: 1973. 1973, Budapest: Akademiai Kiado, 267281.
 46.
Schwarz GE: Estimating the dimension of a model. Annals of Statistics. 1978, 6 (2): 461464.
 47.
Harrell FE: Hmisc: Harrell Miscellaneous. R package version 3.83. 2010
 48.
DebRoy S, Bivand R: foreign: Read Data Stored by Minitab, S, SAS, SPSS, Stata, Systat, dBase, … R package version 0.841. 2010
 49.
Bates D, Maechler M, Bolker B: lme4: Linear mixedeffects models using S4 classes. R package version 0.99937533. 2010
 50.
Louis TA, Zelterman D: Bayesian approaches to research synthesis. The Handbook of Research Synthesis. Edited by: Cooper H, Hedges LV. 2000, New York: Russel Sage Foundation, 411422.
 51.
Smith TC, Spiegelhalter DJ, Thomas A: Bayesian approaches to randomeffects metaanalysis: a comparative study. Stat Med. 1995, 14 (24): 26852699.
Prepublication history
The prepublication history for this paper can be accessed here:http://www.biomedcentral.com/14712288/12/163/prepub
Acknowledgments
1. The manuscript represents the views of the authors and not those of the Department of Veterans Affairs, the United States Government
2. All authors had access to the data and contributed to the manuscript.
Funding
This work was supported, by the Veterans Health Administration Health Services Research and Development (HSR&D) program [grant #REA 08261, Center for Disease Prevention and Health Interventions for Diverse Populations]. The funding agency did not participate in the design and conduct of the study; collection, management, analysis, and interpretation of the data; or preparation, review, and approval of the manuscript.
Author information
Additional information
Competing interests
None of the authors have any financial disclosure or conflict of interest to report.
Authors' contributions
Study concept and design: MG, Acquisition of data: PM and LE. Analysis and interpretation of data: MG, LE, PM, KH, GG, PN, Drafting of the manuscript: MG, KH, PM, GG, PN, and LE. Critical revision of the manuscript for important intellectual content: MG, LE, PM, PN, and KH. Study supervision: MG and LE. All authors read and approved the final manuscript.
Authors’ original submitted files for images
Below are the links to the authors’ original submitted files for images.
Rights and permissions
This article is published under license to BioMed Central Ltd. This is an Open Access article distributed under the terms of the Creative Commons Attribution License (http://creativecommons.org/licenses/by/2.0), which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.
About this article
Cite this article
Gebregziabher, M., Egede, L., Gilbert, G.E. et al. Fitting parametric random effects models in very large data sets with application to VHA national data. BMC Med Res Methodol 12, 163 (2012) doi:10.1186/1471228812163
Received
Accepted
Published
DOI
Keywords
 Generalized linear mixed model
 Homogeneity
 Random effect meta regression
 Longitudinal data
 Very large dataset