Skip to main content

Imputation methods for missing data for polygenic models

Abstract

Methods to handle missing data have been an area of statistical research for many years. Little has been done within the context of pedigree analysis. In this paper we present two methods for imputing missing data for polygenic models using family data. The imputation schemes take into account familial relationships and use the observed familial information for the imputation. A traditional multiple imputation approach and multiple imputation or data augmentation approach within a Gibbs sampler for the handling of missing data for a polygenic model are presented.

We used both the Genetic Analysis Workshop 13 simulated missing phenotype and the complete phenotype data sets as the means to illustrate the two methods. We looked at the phenotypic trait systolic blood pressure and the covariate gender at time point 11 (1970) for Cohort 1 and time point 1 (1971) for Cohort 2. Comparing the results for three replicates of complete and missing data incorporating multiple imputation, we find that multiple imputation via a Gibbs sampler produces more accurate results. Thus, we recommend the Gibbs sampler for imputation purposes because of the ease with which it can be extended to more complicated models, the consistency of the results, and the accountability of the variation due to imputation.

Background

Methods to handle missing data have been a statistical area of research for many years [1]. Little has been done within the context of pedigree analysis. The goals of this paper are: 1) to present two imputation methods for missing phenotype information, and 2) to compare estimates of the additive polygenic effect using variance components or mixed models between the Genetic Analysis Workshop 13 (GAW13) simulated missing phenotype and the complete phenotype data sets for each imputation method [2–4]. In narrowing the focus of our topic, we only looked at the phenotypic trait systolic blood pressure and the covariate gender at time point 11 (1970) for Cohort 1 and time point 1 (1971) for Cohort 2. The methods for imputation described herein include traditional multiple imputation and multiple imputation (data augmentation) via a Gibbs sampler, with both methods accounting for the familial information in the imputation. In fitting the polygenic model to produce estimates of the overall mean effect, gender effect, additive genetic variance, and the residual error variance, we used the expectation-maximization (EM) algorithm program PolyEM [5] and a Bayesian analysis involving use of a Gibbs sampler.

Methods

Traditional multiple imputation method

Multiple imputation is carried out by using the conditional distribution of the missing values given the observed values. Let Yi = (YTmis, YTobs)T be the quantitative phenotype values for the ith family with Ymis representing the vector of missing values and Yobs representing the vector of observed values. Similarly, we can partition the mean and covariance matrix. Thus, for the polygenic model (including an overall mean and a gender effect), the distribution of Ymis given Yobs is

Ymis | Yobs ~ N(μ1, Σ1),

with μ1 = (μ + X1β) + (σ2 D12)(σ2 D22 + τ2 I22)-1(Yobs - (μ + X2 β))

and Σ1 = (σ2 D11 + τ2 I11) - (σ2 D12)(σ2D22 + τ2 I22)-1 (σ2 D21).

The imputation is carried out by generating missing values from the conditional multivariate normal distribution, taking into account the family structure. This imputation is completed m times to produce m complete data sets to analyze. From these m analyses, the final point estimate would be the mean of the m estimates. The computation of the standard error can be done by first computing the between-imputation variation, B m , and the within-imputation variation, W m . Then, the total variability is T m = W m + B m (m + 1)/m. Confidence intervals for parameters of interest use the t-distribution with degrees of freedom (m - 1) (1 + 1/ (m + 1)*W m /B m )2 [6].

The one problem with this imputation scheme is that values for μ, β, σ2, and τ2 are required for the imputation. To address this issue, we ran the analysis on the observed data and used the resulting estimates to create k complete data sets, from which k sets of point estimates are found. Then, the average of the k sets of estimates is used for the m multiple imputations. Point estimates were found using the EM algorithm program PolyEM with k = 5 and m = 25 [5, 7]. Variance estimates were found using the Fisher Information matrix and the large sample properties of maximum likelihood estimates (MLEs) [4, 7, 8].

Multiple imputation via Gibbs sampler

Another approach for the imputation of missing data is through a Bayesian analysis via a Gibbs sampler. The Gibbs sampler is a particular Markov chain algorithm that is useful when working with high dimensional problems. In addition to the traditional use of the Gibbs sampler, an imputation step can also be added to impute missing values. A multiple imputation scheme can be implemented by having an imputation step at the beginning of the Gibbs sampler. For each iteration, a new imputation is done, giving multiple imputations for each missing value [9–15]. For a detailed proof of the data augmentation methodology, see Tanner and Wong [12].

A Bayesian polygenic model with non-informative prior distributions for the ith family is Yi = Xiβ + ai + εi, where Yi is a vector containing the individual responses in family i, Xi is a design matrix containing covariate information, β is a vector of covariate effects, ai is a vector of random family effects where ai ~ MVN(0, σ2Di) and Di is a known coefficient of relationship matrix, and εi ~ MVN(0, τ2I). Non-informative priors were then placed on all other parameters in the model, i.e., p(β) proportional to 1, p(σ2) proportional to 1/σ2, and p(τ2) proportional to 1/τ2 [6].

The steps for implementing the Gibbs sampler with an imputation step for the Bayesian polygenic model follow.

  1. 1.

    Set starting values for β(0), σ2(0), τ2(0), ai(0) for all i = 1,..., k, and set m = 1 (iteration).

  2. 2.

    If yij is missing, impute yij by simulating an observation from N(Xij β (m-1) + aij(m-1), τ2(m-1)).

  3. 3.

    Generate β(m) from MVN((XTX)-1XT(y - a(m-1)), τ2(m-1)(XTX)-1).

  4. 4.

    Generate τ2(m) from INGAM(N/2, 1/2 Σ(yi-Xiβ(m) - ai(m-1))T(yi-Xiβ(m) - ai(m-1))).

  5. 5.

    Generate σ2(m) from INGAM(N/2, 1/2 Σ ai(m-1)T Di-1ai(m-1)).

  6. 6.

    Generate ai(m) from MVN(μa(m), Va(m)), where μa(m) = (1/σ2(m) * Di-1 + 1/τ2(m)* Ii)-1 * (1/ τ2(m) * Ii (yi - Xiβ(m))) and Va(m) = (1/σ2(m) * Di-1 + 1/τ2(m) * Ii)-1.

  7. 7.

    After one iteration of the algorithm, you have (β(m), σ2(m), τ2(m), a(m)). Set m = m + 1 and repeat steps 1 through 6.

Approximate (1 - α)% posterior confidence intervals are found by taking the α/2 and the 1-α/2 percentiles of the simulated marginal posterior distributions for parameters of interest. The simulated posterior distributions will reflect the uncertainty in the estimation of the parameter along with the uncertainty due to the imputation of the missing data.

Results

The two imputation methods were run for three replicates of the GAW13 simulated complete and missing data sets for the phenotypic trait of systolic blood pressure and the covariate of gender at time point 11 (1970) for Cohort 1 and time point 1 (1971) for Cohort 2. We limited our analysis to two odd-numbered replicates and one even-numbered replicate to demonstrate the methods due to time and computational restrictions. Table 1 displays the results for the GAW13 missing and complete data sets, in which a likelihood analysis using an EM algorithm using the traditional imputation approach for the missing data. Table 2 displays the results for the GAW13 missing and complete data sets, in which a Bayesian model was used involving a data augmentation step for missing data.

Table 1 95% Confidence intervals using "traditional" multiple imputation within a likelihood analysis for the complete and missing simulated data
Table 2 95% Approximate posterior intervals using data augmentation within a Bayesian/Model for the complete and missing simulated data

The confidence intervals show little difference between the complete and missing data set analysis within each type of imputation method with regards to the estimates for the error variance component. The data augmentation within a Gibbs sampler gives intervals similar to those of the complete data set, while the traditional imputation approach for missing data produced differing intervals when compared with the complete data set analysis. Not only are the confidence intervals narrower after the imputation of missing data, but the point estimates for the polygenic variance component are much smaller (underestimated) as compared to the results of the complete data sets for the traditional imputation method. These intervals show that the traditional imputation method produced inaccurate intervals as compared with the intervals for the complete data set. The inaccuracy may be due to the choice of the parameter values for the imputation or the degrees of freedom used for the intervals [16].

Conclusions

Missing data complicates statistical analysis. One way to deal with missing data is through the use of imputation. In the case of family data, the information provided by the dependence structure can be utilized in the imputation of missing data. We have discussed two methods of imputation and illustrated their application through the GAW13 simulated data sets. We assumed the missing data mechanism in the GAW13 data set was ignorable. The Gibbs sampler and the traditional multiple imputation method can be easily applied to non-ignorable missing data, such as in cases involving censored phenotype data. In the case of non-ignorable missingness, not adjusting for the missing/censored values will lead to biased estimates. Based on results from the three replicates, the Gibbs sampler approach seems to give more accurate confidence intervals, as opposed to the traditional multiple imputation approach. In addition, the traditional imputation method has the drawback of needing parameter values for the completion of the imputation. The Gibbs sampler approach does not encounter this problem, because any set of estimates may be used as starting values for the algorithm. In conclusion, we recommend the Gibbs sampler for imputation purposes because of the ease with which it can be extended to more complicated models, the consistency of the results, and the accountability of the variation due to imputation. Future work is planned to investigate the use of data augmentation for missing phenotype data in longitudinal studies and quantitative trait locus analysis.

References

  1. Little RJA, Rubin D: Statistical Analysis with Missing Data. New York, Wiley. 2002, 2

    Google Scholar 

  2. Hopper J, Mathews J: Extensions to multivariate normal models for family analysis. Ann Hum Genet. 1982, 46: 373-383.

    Article  CAS  PubMed  Google Scholar 

  3. de Andrade M, Amos C, Thiel T: Methods to estimate genetic components of variance for quantitative traits in family studies. Genet Epidemiol. 1999, 17: 64-76. 10.1002/(SICI)1098-2272(1999)17:1<64::AID-GEPI5>3.0.CO;2-M.

    Article  CAS  PubMed  Google Scholar 

  4. Lange K, Westlake J, Spence M: Extensions to pedigree analysis. III. Variance components by scoring method. Ann Hum Genet. 1976, 39: 485-491.

    Article  CAS  PubMed  Google Scholar 

  5. Thompson E, Shaw R: Pedigree analysis for quantitative traits: variance components without matrix inversion. Biometrics. 1990, 46: 399-413. 10.2307/2531445.

    Article  CAS  PubMed  Google Scholar 

  6. Gelman A, Carlin J, Stern H, Rubin D: Bayesian Data Analysis. New York, Chapman & Hall. 1995

    Google Scholar 

  7. Dempster A, Laird N, Rubin D: Maximum likelihood from incomplete data via the EM algorithm (with discussion). J R Stat Soc B. 1977, 39: 1-38.

    Google Scholar 

  8. Rao C: Linear Statistical Inference and Its Applications. New York, Wiley. 1973, 2

    Google Scholar 

  9. Hopke P, Liu C, Rubin D: Multiple imputation for multivariate data with missing and below-threshold measurements: time-series concentrations of pollutants in the arctic. Biometrics. 2001, 57: 22-33. 10.1111/j.0006-341X.2001.00022.x.

    Article  CAS  PubMed  Google Scholar 

  10. Gelfand A, Smith A: Sampling-based approaches to calculating marginal densities. J Am Stat Assoc. 1990, 85: 398-409. 10.2307/2289776.

    Article  Google Scholar 

  11. Geman S, Geman D: Stochastic relaxation, Gibbs distributions, and the Bayesian restoration of images. IEEE Trans Pattern Analysis Machine Intelligence. 1984, 6: 721-741.

    Article  CAS  Google Scholar 

  12. Tanner M, Wong W: The calculation of posterior distributions by data augmentation. J Am Stat Assoc. 1987, 82: 528-540. 10.2307/2289457.

    Article  Google Scholar 

  13. Li K: Imputation using Markov chains. J Stat Comput Simul. 1988, 30: 57-79.

    Article  Google Scholar 

  14. Van Dyk D, Meng X: The art of data augmentation. J Comput Graph Stat. 2001, 10: 1-50. 10.1198/10618600152418584.

    Article  Google Scholar 

  15. Schafer J: Analysis of Incomplete Multivariate Data. New York, Chapman & Hall. 1997

    Google Scholar 

  16. Barnard J, Rubin D: Small-sample degrees of freedom with multiple imputation. Biometrika. 1999, 86: 948-955. 10.1093/biomet/86.4.948.

    Article  Google Scholar 

Download references

Acknowledgments

We thank Curtis Olswold for his help. This research was partially founded by NIH grant R01HL71917.

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Brooke Fridley.

Rights and permissions

Reprints and permissions

About this article

Cite this article

Fridley, B., Rabe, K. & Andrade, M.d. Imputation methods for missing data for polygenic models. BMC Genet 4 (Suppl 1), S42 (2003). https://doi.org/10.1186/1471-2156-4-S1-S42

Download citation

  • Published:

  • DOI: https://doi.org/10.1186/1471-2156-4-S1-S42

Keywords