Namespaces
Variants
Actions

Difference between revisions of "Linear mixed model"

From Encyclopedia of Mathematics
Jump to: navigation, search
(Created page with " {| class="wikitable" !Copyright notice <!-- don't remove! --> |- | This article ''Linear Mixed Model'' was adapted from an original article by Geert Molenberghs, which a...")
 
Line 53: Line 53:
 
$\yi$ of repeated measurements for each unit separately, where some of the
 
$\yi$ of repeated measurements for each unit separately, where some of the
 
regression parameters are specific (random effects, $\bi$), while others are
 
regression parameters are specific (random effects, $\bi$), while others are
not (fixed effects, $\bm\beta$). The distributional assumptions in (\ref{lin
+
not (fixed effects, $\bm\beta$). The distributional assumptions in \ref{lin mix eff model 2} with respect to the random effects can be motivated as
mix eff model 2}) with respect to the random effects can be motivated as
 
 
follows. First, $\mbox{E}(\bi)=\bm{0}$ implies that the mean of $\yi$ still
 
follows. First, $\mbox{E}(\bi)=\bm{0}$ implies that the mean of $\yi$ still
 
equals $X_i\bm\beta$, such that the fixed effects in the random-effects
 
equals $X_i\bm\beta$, such that the fixed effects in the random-effects
model (\ref{lin mix eff model 1}) can also be interpreted marginally. Not only
+
model \ref{lin mix eff model 1} can also be interpreted marginally. Not only
 
do they reflect the effect of changing covariates within specific units, they
 
do they reflect the effect of changing covariates within specific units, they
 
also measure the marginal effect in the population of changing the same
 
also measure the marginal effect in the population of changing the same
Line 65: Line 64:
 
matrix $V_i = Z_i D Z_i' + \Sigma_i$.
 
matrix $V_i = Z_i D Z_i' + \Sigma_i$.
  
Note that the random effects in (\ref{lin mix eff model 1}) implicitly imply
+
Note that the random effects in \ref{lin mix eff model 1} implicitly imply
 
the marginal covariance matrix $V_i$ of $\yi$ to be of the very specific form
 
the marginal covariance matrix $V_i$ of $\yi$ to be of the very specific form
 
$V_i=Z_i D Z_i' + \Sigma_i$. Let us consider two examples under the
 
$V_i=Z_i D Z_i' + \Sigma_i$. Let us consider two examples under the
Line 73: Line 72:
 
are $n_i$-dimensional vectors containing only ones.
 
are $n_i$-dimensional vectors containing only ones.
  
The marginal model implied by expressions (\ref{lin mix eff model 1}) and
+
The marginal model implied by expressions \ref{lin mix eff model 1} and
(\ref{lin mix eff model 2}) is
+
\ref{lin mix eff model 2} is
 
\begin{eqnarray*}
 
\begin{eqnarray*}
 
\yi & \sim & N(X_i \bm\beta, V_i), \quad V_i = Z_i D Z_i' + \Sigma_i
 
\yi & \sim & N(X_i \bm\beta, V_i), \quad V_i = Z_i D Z_i' + \Sigma_i
Line 112: Line 111:
 
\end{eqnarray*}
 
\end{eqnarray*}
  
\
+
 
 
====References====
 
====References====
 
{|
 
{|
Line 125: Line 124:
 
|-
 
|-
 
|}
 
|}
*{Acknowledgment}
+
*Acknowledgment
  
 
Based on an article from Lovric, Miodrag (2011), International Encyclopedia of Statistical Science. Heidelberg: Springer Science +Business Media, LLC
 
Based on an article from Lovric, Miodrag (2011), International Encyclopedia of Statistical Science. Heidelberg: Springer Science +Business Media, LLC

Revision as of 11:59, 31 March 2016

Copyright notice
This article Linear Mixed Model was adapted from an original article by Geert Molenberghs, which appeared in StatProb: The Encyclopedia Sponsored by Statistics and Probability Societies. The original article ([http://statprob.com/encyclopedia/LinearMixedModel.html StatProb Source], Local Files: pdf | tex) is copyrighted by the author(s), the article has been donated to Encyclopedia of Mathematics, and its further issues are under Creative Commons Attribution Share-Alike License'. All pages from StatProb are contained in the Category StatProb.

2020 Mathematics Subject Classification: Primary: 62-XX [MSN][ZBL]


$ \newcommand{\yi}{\bm{y_i}} $ $ \def\bm#1{\mathbf{#1}} $ $ \newcommand{\bi}{\bm{b_i}} $

Linear Mixed Model

Geert Molenberghs


I-BioStat, Universiteit Hasselt & Katholieke Universiteit Leuven, Belgium.

In many studies, data are collected hierarchically. Not always do such data follow balanced, multivariate designs. For example, in repeated measurements may be taken at almost arbitrary time points, resulting in an extremely large number of time points at which only one or only a few measurements have been taken. Many of the parametric covariance models described so far may then contain too many parameters to make them useful in practice, while other, more parsimonious, models may be based on assumptions which are too simplistic to be realistic. A general, and very flexible, class of parametric models for continuous longitudinal data is formulated as follows: \begin{eqnarray} \yi | \bi & \sim & N(X_i \bm\beta + Z_i \bm{b_i}, \Sigma_i), \label{lin mix eff model 1} \\ \bi & \sim & N(\bm{0}, D), \label{lin mix eff model 2} \end{eqnarray} where $X_i$ and $Z_i$ are $(n_i \times p)$ and $(n_i \times q)$ dimensional matrices of known covariates, $\bm\beta$ is a $p$-dimensional vector of regression parameters, called the fixed effects, $D$ is a general $(q \times q)$ covariance matrix, and $\Sigma_i$ is a $(n_i \times n_i)$ covariance matrix which depends on $i$ only through its dimension $n_i$, i.e., the set of unknown parameters in $\Sigma_i$ will not depend upon $i$. Finally, $\bm{b_i}$ is a vector of subject-specific or random effects.

The above model can be interpreted as a linear regression model for the vector $\yi$ of repeated measurements for each unit separately, where some of the regression parameters are specific (random effects, $\bi$), while others are not (fixed effects, $\bm\beta$). The distributional assumptions in \ref{lin mix eff model 2} with respect to the random effects can be motivated as follows. First, $\mbox{E}(\bi)=\bm{0}$ implies that the mean of $\yi$ still equals $X_i\bm\beta$, such that the fixed effects in the random-effects model \ref{lin mix eff model 1} can also be interpreted marginally. Not only do they reflect the effect of changing covariates within specific units, they also measure the marginal effect in the population of changing the same covariates. Second, the normality assumption immediately implies that, marginally, $\yi$ also follows a normal distribution with mean vector $X_i \bm\beta$ and with covariance matrix $V_i = Z_i D Z_i' + \Sigma_i$.

Note that the random effects in \ref{lin mix eff model 1} implicitly imply the marginal covariance matrix $V_i$ of $\yi$ to be of the very specific form $V_i=Z_i D Z_i' + \Sigma_i$. Let us consider two examples under the assumption of conditional independence, i.e., assuming $\Sigma_i=\sigma^2 I_{n_i}$. First, consider the case where the random effects are univariate and represent unit-specific intercepts. This corresponds to covariates $Z_i$ which are $n_i$-dimensional vectors containing only ones.

The marginal model implied by expressions \ref{lin mix eff model 1} and \ref{lin mix eff model 2} is \begin{eqnarray*} \yi & \sim & N(X_i \bm\beta, V_i), \quad V_i = Z_i D Z_i' + \Sigma_i \end{eqnarray*} which can be viewed as a multivariate linear regression model, with a very particular parameterization of the covariance matrix $V_i$.

With respect to the estimation of unit-specific parameters $\bi$, the posterior distribution of $\bi$ given the observed data $\yi$ can be shown to be (multivariate) normal with mean vector equal to $D Z_i' V_i^{-1}(\bm\alpha) (\yi - X_i \bm\beta)$. Replacing $\bm\beta$ and $\bm\alpha$ by their maximum likelihood estimates, we obtain the so-called empirical Bayes estimates $\widehat{\bi}$ for the $\bi$. A key property of these EB estimates is shrinkage, which is best illustrated by considering the prediction $\widehat{\yi} \equiv X_i \widehat{\bm\beta} + Z_i \widehat{\bi}$ of the $i$th profile. It can easily be shown that \begin{eqnarray*} \widehat{\yi} & = & \Sigma_i V_i^{-1} X_i \widehat{\bm\beta} + \ \left(I_{n_i} - \Sigma_i V_i^{-1} \right) \yi, \end{eqnarray*} which can be interpreted as a weighted average of the population-averaged profile $X_i \widehat{\bm\beta}$ and the observed data $\yi$, with weights $\Sigma_i V_i^{-1}$ and $I_{n_i} - \Sigma_i V_i^{-1}$, respectively. Note that the "numerator" of $\Sigma_i V_i^{-1}$ represents within-unit variability and the "denominator" is the overall covariance matrix $V_i$. Hence, much weight will be given to the overall average profile if the within-unit variability is large in comparison to the between-unit variability (modeled by the random effects), whereas much weight will be given to the observed data if the opposite is true. This phenomenon is referred to as shrinkage toward the average profile $X_i \widehat{\bm\beta}$. An immediate consequence of shrinkage is that the EB estimates show less variability than actually present in the random-effects distribution, i.e., for any linear combination $\bm\lambda$ of the random effects, \begin{eqnarray*} \mbox{var}(\bm\lambda' \widehat{\bi}) & \leq & \mbox{var}(\bm\lambda' \bi) = \bm\lambda' D \bm\lambda. \end{eqnarray*}


References

[1] Fitzmaurice, G.M., Davidian, M., Verbeke, G., and Molenberghs, G.(2009). Longitudinal Data Analysis. Handbook. Hoboken, NJ: John Wiley & Sons.
[2] Fitzmaurice, G.M., Laird, N.M., and Ware, J.H. (2004). Applied Longitudinal Analysis. New York: John Wiley & Sons.
[3] Henderson, C.R. (1984) Applications of Linear Models in Animal Breeding. Guelph, Canada: University of Guelph Press.
[4] Verbeke, G. and Molenberghs, G. (2000) Linear Mixed Models for Longitudinal Data. New York: Springer.
  • Acknowledgment

Based on an article from Lovric, Miodrag (2011), International Encyclopedia of Statistical Science. Heidelberg: Springer Science +Business Media, LLC


How to Cite This Entry:
Linear mixed model. Encyclopedia of Mathematics. URL: http://encyclopediaofmath.org/index.php?title=Linear_mixed_model&oldid=38521