knitr::opts_chunk$set( collapse = TRUE, comment = "#>" ) options("digits"=4)
library(caracas) library(Matrix) ##packageVersion("caracas")
if (!has_sympy()) { # SymPy not available, so the chunks shall not be evaluated knitr::opts_chunk$set(eval = FALSE) inline_code <- function(x) { deparse(substitute(x)) } }
This vignette is based on caracas
version
r packageVersion("caracas")
. caracas
is avavailable on CRAN at
[https://cran.r-project.org/package=caracas] and on github at
[https://github.com/r-cas/caracas].
A linear mixed model can be written in general form as
$$ y = Xb + Zu + e $$
Here, $y$ is a vector of observables, $X$ is a model matrix and $b$ is a corresponding vector of regression coefficients. Also $Z$ is a model matrix and $u$ is a corresponding vector of random effects. Lastly, $e$ is also a vector of random errors. Because there are two random effects ($u$ and $e$) such models are often called mixed models. qIt is assumed that $u$ and $e$ are independent and that $u \sim N(0, G)$ and $e \sim N(0, R)$. Typically $G$ and $R$ depend on unknown parameters $\theta$, so we may write $G(\theta)$ and $R(\theta)$ instead.
Consequently
$$ E(y)=\mu=Xb, \quad Var(y) = V(\theta) =ZG(\theta)Z'+R(\theta). $$
We illustrate fitting a mixed model based on a subset of the shoes data available, e.g. in the MASS and doBy packages. We also compare the results with the output from lmer() if the lme4 package is installed and with gls() if the nlme package is installed.
The log-likelihood is
$$ logL(b, \theta) = -\frac 1 2 (\log|V(\theta)| + (y-Xb)'V(\theta)^{-1}(y-Xb)). $$
For any value of $\theta$, the MLE for $b$ is
$$ \hat b =\hat b(\theta) = (X'V(\theta)^{-1}X)^{-1}X'V(\theta)^{-1}y. $$
Plugging this estimate into the log-likelihood gives the profile log-likelihood which is a function of $\theta$ only:
$$ plogL(\theta) = - \frac 1 2 (\log|V(\theta)| + (y-X\hat b(\theta))'V(\theta)^{-1}(y-X\hat b(\theta))). $$
shoes_long <- data.frame( stringsAsFactors = FALSE, type = c("A", "B", "A", "B", "A", "B", "A", "B"), wear = c(13.2, 14, 8.2, 8.8, 10.9, 11.2, 14.3, 14.2), boy = as.factor(c("1", "1", "2", "2", "3", "3", "4", "4"))) y <- shoes_long$wear X <- model.matrix(~type, data=shoes_long) |> head(10) Z <- model.matrix(~-1+boy, data=shoes_long) |> head(10) y |> head() X |> head() Z |> head()
caracas
We define the following symbols in caracas. All caracas symbols are postfixed with an underscore.
y_ <- as_sym(y) X_ <- as_sym(X) Z_ <- as_sym(Z) b_ <- vector_sym(ncol(X_), "b") def_sym(tau2, sigma2)
## Covariance matrices for random effects G_ <- diag_("tau^2", ncol(Z)) R_ <- diag_("sigma^2", nrow(Z))
So for this example, $\theta=(\tau, \sigma)$. The variance of $y$ is
## Variance etc of y V_ <- Z_ %*% G_ %*% t(Z_) + R_ #Vi_ <- inv_block(V_) #Vi_ <- inv(V_) Vi_ <- inv_woodbury(R_, Z_, G_) detV_ <- determinant(V_, log=FALSE)
$$
r tex_list("G=", G_, ", R=", R_, zero_as_dot=T)
$$
The first blocks of $V$ and $V^{-1}$ are:
$$
r tex_list("V=", V_[1:4,1:4], zero_as_dot=T)
$$
$$
r tex_list("V^{-1}=", Vi_[1:4,1:4], zero_as_dot=T)
$$
A programmatic approach is to define a function returning the log-likelihood and the profile log-likelihood as a caracas symbol. Notice that we supply the both $V$, $V^{-1}$ and $\log|V|$ as arguments to the log-likelihood function. There is redundancy in this, but it is convenient to be able to supply alternative forms for $V$.
get_logL0 <- function(y, X, b, V, Vi=inv_block(V), detV=determinant(V, log=FALSE)) { res <- function(y, X, b) { return(y - X %*% b) } qform <- function(res, Vi) { return(t(res) %*% Vi %*% res) } res <- res(y, X, b) Q <- qform(res, Vi) return(-0.5 * (log(detV) + Q)) } get_bhat <- function(y, X, V, Vi=inv_block(V)) { return(solve(t(X) %*% Vi %*% X, t(X) %*% Vi) %*% y) } get_V <- function(Z, G, R) { return(Z %*% G %*% t(Z) + R) } get_hessian <- function(logL, b) { return(-der2(logL, b)) }
V_ <- get_V(Z_, G_, R_) bhat_ <- get_bhat(y_, X_, V_) |> simplify() b_ <- vector_sym(ncol(X_), "b")
$$
r tex_list("b=", b_, ", \\hat b=", bhat_, zero_as_dot=FALSE)
$$
Notice that $\hat b$ is a function of $V$ and hence of $\theta$. If $\hat b$ is substituted into the log-likelihood, we get the profile log-likelihood. On the other hand, $b$ is a parameter of fixed effect and the log-likelihood is a function of both $b$ and $\theta$.
plogL_ <- get_logL0(y_, X_, bhat_, V_) |> simplify() out <- optim_sym(c(.1, .1), plogL_, method="L-BFGS-B", control=list(fnscale=-1), hessian=TRUE) var_par <- out$par var_par bhat2_ <- subs(bhat_, as.list(var_par)) bhat2_ as_expr(bhat2_)
A technicality: A caracas symbol can be coerced to an R function. Therefor $\hat b$ can be evaluated at the MLE as a function of $\theta$.
as_func(bhat_, vec_arg=TRUE)(var_par)
To obtain the asymptotic variance of $\hat b$, we need to compute the Hessian of the log-likelihood at the MLE. This is done by differentiating the log-likelihood with respect to $b$ and then evaluating at the MLE. The Hessian is then inverted to obtain the asymptotic variance of $\hat b$.
logL_ <- get_logL0(y_, X_, b_, V_) |> simplify() J_ <- get_hessian(logL_, b_) J2_ <- subs(J_, as.list(var_par)) Vb_ <- solve(J2_) as_expr(bhat2_) as_expr(Vb_)
An alternative is to maximize the full likelihood, we need to maximize with respect to both the variance parameters and the fixed effects. For this to work we often need to impose restrictions on the parameter space (not necessary for this specific example, though)
eps <- 1e-6 out <- optim_sym(c(0, 0, .1, .1), logL_, method="L-BFGS-B", lower=c(-Inf, -Inf, eps, eps), upper=c(Inf, Inf, Inf, Inf), control=list(fnscale=-1), hessian=TRUE) out$par solve(-out$hessian)[1:2, 1:2] |> round(4)
It is often useful to reparametrize the model to avoid constraints on the parameters. Three typical cases are:
If $\alpha$ must be positive (e.g. a variance), reparametrize $\alpha = \exp(w)$ and optimize with respect to $w$ which is unconstrained.
If $\alpha$ must be in $[0, 1]$, (e.g. a probability) reparametrize $\alpha = \frac{\exp(w)}{1+\exp(w)}$ and optimize with respect to $w$ which is unconstrained.
If $\alpha$ must be in $[-1, 1]$, (e.g. a correlation) reparametrize $\alpha = \frac{\exp(w)-1}{\exp(w)+1} = \tanh(w)$ and optimize with respect to $w$ which is unconstrained.
V2_ <- subs(V_, list(sigma="exp(log_sigma)", tau="exp(log_tau)")) |> simplify() bhat2_ <- get_bhat(y_, X_, V2_)
$$
r tex_list("V2=", V2_[1:4,1:4], zero_as_dot=T)
$$
With this parameterization, the profile log-likelihood is maximized as:
plogL2_ <- get_logL0(y_, X_, bhat2_, V2_) out <- optim_sym(c(.1, .1), plogL2_, method="L-BFGS-B", control=list(fnscale=-1), hessian=TRUE) var_par <- out$par var_par |> exp()
Moreover, with this parameterization, the full log-likelihood can be maximized without constraints, and we can obtain the Hessian (and hence the asymptotic variance of the MLE) directly:
logL2_ <- get_logL0(y_, X_, b_, V2_) out <- optim_sym(c(0, 0, .1, .1), logL2_, method="L-BFGS-B", control=list(fnscale=-1), hessian=TRUE) var_par <- out$par var_par solve(-out$hessian)[1:2, 1:2]
An alternative is to write V in terms of a variance and a correlation parameter
def_sym(sigma) B <- as_sym(toeplitz(c(1, "rho"))) V3_ <- sigma^2 * kronecker(diag_(1, ncol(Z)), B)
$$
r tex_list("V3=", V3_[1:4,1:4], zero_as_dot=T)
$$
bhat3_ <- get_bhat(y_, X_, V3_) plogL3_ <- get_logL0(y_, X_, bhat3_, V3_) logL3_ <- get_logL0(y_, X_, b_, V3_) out <- optim_sym(c(.1, .1), plogL3_, method="L-BFGS-B", lower=c(-1+eps, eps), upper=c(1-eps, Inf), control=list(fnscale=-1), hessian=TRUE) var_par <- out$par var_par subs(bhat3_, as.list(var_par)) J_ <- get_hessian(logL3_, b_) J2_ <- subs(J_, as.list(var_par)) Vb_ <- solve(J2_) Vb_
Just for illustration, we note that maximizing the full likelihood fails in this case in the sense that we get wrong results.
out <- optim_sym(c(0, 0, .1, .1), logL3_, method="L-BFGS-B", lower=c(-Inf, -Inf, -1+eps, eps), upper=c(Inf, Inf, 1-eps, Inf), control=list(fnscale=-1), hessian=TRUE) out$par solve(-out$hessian)[1:2, 1:2]
To maximize the full likelihood, we need to reparametrize the correlation parameter. We can do this by writing $\rho = \frac{\exp(2w)-1}{\exp(2w)+1}$ and optimize with respect to $w$ which is unconstrained.
V4_ <- subs(V3_, list(rho="(exp(2*w)-1)/(exp(2*w)+1)"))
$$
r tex_list("V4=", V4_[1:4,1:4], zero_as_dot=T)
$$
logL4_ <- get_logL0(y_, X_, b_, V4_) out <- optim_sym(c(0, 0, .1, .1), logL4_, method="L-BFGS-B", control=list(fnscale=-1), hessian=TRUE) var_par <- out$par var_par tanh(var_par["w"])
if (require(nlme)){ model <- gls(wear ~ type, correlation = corCompSymm(form = ~ 1 | boy), method="ML", data = shoes_long) sigma2 <- model$sigma^2 C <- corMatrix(model$modelStruct$corStruct) V <- sigma2 * as.matrix(bdiag(C)) as(V, "sparseMatrix") }
if (require(lme4)){ lmm_fit <- lmer(wear ~ type + (1|boy), data=shoes_long, REML=FALSE) print(VarCorr(lmm_fit)) print(fixef(lmm_fit)) print(vcov(lmm_fit)) }
Add the following code to your website.
For more information on customizing the embed code, read Embedding Snippets.