Nothing
#### =========== ######
## PREDICT FUNCTION #
#### =========== ######
# include is used for aggregating
# averaged is used to be included in the prediction
# ignored is not used included in the prediction
"predict.mmer" <- function(object, Dtable=NULL, D, ...){
if(is.character(D)){classify <- D}else{classify="id"} # save a copy before D is overwriten
# complete the Dtable withnumber of effects in each term
xEffectN <- object$xEffectsN
zEffectsN <- object$zEffectsN
effectsN <- c(xEffectN,zEffectsN)
start = end = numeric(); add <- 1
for(i in 1:length(effectsN)){
start[i] = add
end[i] = start[i] + effectsN[i] - 1
add = end[i] + 1
}
# fill the Dt table
if(is.null(Dtable) & is.character(D)){ # if user didn't provide the Dtable
Dtable <- object$Dtable # we extract it from the model
termsInDtable <- apply(data.frame(Dtable$term),1,function(xx){all.vars(as.formula(paste0("~",xx)))})
termsInDtable <- lapply(termsInDtable, function(x){intersect(x,colnames(object$data))})
termsInDtable <- lapply(termsInDtable, function(x){return(unique(c(x, paste(x, collapse=":"))))})
# term identified
termsInDtableN <- unlist(lapply(termsInDtable,length))
pickTerm <- which( unlist(lapply(termsInDtable, function(xxx){ifelse(length(which(xxx == D)) > 0, 1, 0)})) > 0)
if(length(pickTerm) == 0){
isInDf <- which(colnames(object$data) %in% D)
if(length(isInDf) > 0){ # is in data frame but not in model
stop(paste("Predict:",classify,"not in the model but present in the original dataset. You may need to provide the
Dtable argument to know how to predict", classify), call. = FALSE)
}else{
stop(paste("Predict:",classify,"not in the model and not present in the original dataset. Please correct D."), call. = FALSE)
}
}
# check if the term to predict is fixed or random
pickTermIsFixed = ifelse("fixed" %in% Dtable[pickTerm,"type"], TRUE,FALSE)
## 1) when we predict a random effect, fixed effects are purely "average"
if(!pickTermIsFixed){Dtable[which(Dtable$type %in% "fixed"),"average"]=TRUE; Dtable[pickTerm,"include"]=TRUE}
## 2) when we predict a fixed effect, random effects are ignored and the fixed effect is purely "include"
if(pickTermIsFixed){Dtable[pickTerm,"include"]=TRUE}
## 3) for predicting a random effect, the interactions are ignored and only main effect is "included", then we follow 1)
## 4) for a model with pure interaction trying to predict a main effect of the interaction we "include" and "average" the interaction and follow 1)
if(length(pickTerm) == 1){ # only one effect identified
if(termsInDtableN[pickTerm] > 1){ # we are in #4 (is a pure interaction model)
Dtable[pickTerm,"average"]=TRUE
}
}else{# more than 1, there's main effect and interactions, situation #3
main <- which(termsInDtableN[pickTerm] == min(termsInDtableN[pickTerm])[1])
Dtable[pickTerm,"include"]=FALSE; Dtable[pickTerm,"average"]=FALSE # reset
Dtable[pickTerm[main],"include"]=TRUE
}
}
## if user has provided D as a classify then we create the D matrix
if(is.character(D)){
interceptColumn <- grep("Intercept",object$Beta$Effect )
Dtable$start <- start
Dtable$end <- end
# create model matrices to form D
P <- sparse.model.matrix(as.formula(paste0("~",D,"-1")), data=object$data)
colnames(P) <- gsub(D,"",colnames(P))
tP <- t(P)
W <- object$W
D = tP %*% W
toRemove <- names(object$U)
for(ii in 1:length(toRemove)){
names(object$U[[ii]][[1]]) <- gsub(toRemove[ii],"",names(object$U[[ii]][[1]]))
}
colnames(D) <- c(as.character(object$Beta$Effect),unlist(lapply(object$U,function(y){names(y[[1]])})))
rd <- rownames(D)
cd <- colnames(D)
for(jRow in 1:nrow(D)){ # for each effect add 1's where missing
myMatch <- which(cd == rd[jRow])
if(length(myMatch) > 0){D[jRow,myMatch]=1}
}
# apply rules in Dtable
for(iRow in 1:nrow(Dtable)){
s <- Dtable[iRow,"start"]; e <- Dtable[iRow,"end"]
# include/exclude rule
if(Dtable[iRow,"include"]){ # set to 1
subD <- D[,s:e,drop=FALSE]
subD[which(subD > 0, arr.ind = TRUE)] = 1
D[,s:e] <- subD
# average rule
if(Dtable[iRow,"average"]){ # set to 1
# average the include set
for(o in 1:nrow(subD)){
v <- which(subD[o,] > 0); subD[o,v] <- subD[o,v]/(length(v) + length(interceptColumn))
}
D[,s:e] <- subD
}
}else{ # set to zero
if(Dtable[iRow,"average"]){ # set to 1
subD <- D[,s:e,drop=FALSE] + 1
subD <- subD/subD
subD[which(subD > 0, arr.ind = TRUE)] = subD[which(subD > 0, arr.ind = TRUE)]/(ncol(subD) + length(interceptColumn))
D[,s:e] <- subD
}else{
D[,s:e] <- D[,s:e] * 0
}
}
}
if(length(interceptColumn) > 0){D[,interceptColumn] = 1}
if(length(which(Dtable$term == "1")) > 0){Dtable[which(Dtable$term == "1"),"include"]=TRUE}
}else{ }# user has provided D as a matrix to do direct multiplication
## calculate predictions and standard errors
bu <- c(object$Beta$Estimate,unlist(object$U)) #object$bu
predicted.value <- D %*% bu
## standard error
Vi <- object$Vi
ViW = Vi %*% W; # ViW (nxp)
WtViW = t(W) %*% ViW; # W'ViW (pxp)
# inverse of W'ViW
WtViWi <- try(
solve(WtViW),
silent = TRUE
)
if(inherits(WtViWi,"try-error") ){
WtViW = WtViW + diag(mean(diag(Vi)), nrow(WtViW),nrow(WtViW))
WtViWi <- try(
solve(WtViW),
silent = TRUE
)
}
# get G (not sure if this is the right G) var(u) = Z' G [Vi - (VX*tXVXVX)] G Z'
# H <- do.call(adiag1, lapply(object$VarU, function(x){do.call(adiag1,x)}))
# H <- adiag1(object$VarBeta*0,H)
# G <- H
# Gs <- lapply(as.list(1:length(object$K)), function(x){object$K[[x]]*as.vector(object$sigma[[x]])})
# Gs2 <- adiag1(object$VarBeta*0,do.call(adiag1,Gs))
# tWG <- W %*% Gs2
# G <- t(tWG) %*% object$P %*% tWG
PEV <- do.call(adiag1, lapply(object$PevU, function(x){do.call(adiag1,x)}))
PEV <- adiag1(object$VarBeta,PEV)
G <- PEV # it should ideally be G but we don't extract that
# build the projection matrix
P <- object$P # Vi - (Vi%*%X%*%(XtViX)%*%t(X)%*%Vi)
# project W using P
WtPW <- t(W) %*% (P) %*% W
# add G to complete the coefficient matrix
Ci <- WtPW + G # G because is the inverse of Gi which is what goes into the M matrix
vcov <- D %*% Ci %*% t(D)
std.error <- sqrt(diag(vcov))
pvals <- data.frame(id=rownames(D),predicted.value=predicted.value[,1], std.error=std.error)
if(is.character(classify)){colnames(pvals)[1] <- classify}
return(list(pvals=pvals,D=D,vcov=vcov, Dtable=Dtable))
}
"print.predict.mmer"<- function(x, digits = max(3, getOption("digits") - 3), ...) {
cat(blue(paste("
The predictions are obtained by averaging/aggregating across
the hypertable calculated from model terms constructed solely
from factors in the include sets. You can customize the model
terms used with the 'hypertable' argument. Current model terms used:\n")
))
# print(x$hypertable)
cat(blue(paste("\n Head of predictions:\n")
))
head(x$pvals,...)
}
# "predict.mmer" <- function(object,classify=NULL,hypertable=NULL,...){
#
# getHypertable <- function(object,classify=NULL,...){
#
# classify<- unique(unlist(strsplit(classify,":")))
#
# if(is.null(classify)){
# stop("Please provide the classify argument. For fitted values use the fitted() function.",call. = FALSE)
# }
#
# oto <- oto2 <- object$terms
# oto2$fixed[[1]] <- setdiff(oto2$fixed[[1]],c("1","-1"))
# oto2$fixed <- lapply(oto2$fixed,function(x){paste(x,collapse = ":")})
# oto2$random <- lapply(oto2$random,function(x){paste(x,collapse = ":")}) # paste to present as the form A:B:C
#
# for(u in 3:length(object$terms)){ # change random terms to split by ":"
# prov <- object$terms[[u]]
# if(length(prov) > 0){
# for(v in 1:length(prov)){
# object$terms[[u]][[v]] <- unlist(strsplit(prov[[v]],":"))
# }
# }
# }
#
# # This doesn't quite give identical results. Does it matter?
# include <- unique(c(all.vars(object$call$fixed)[-1], # variables in the fixed formula (remove the response)
# all.vars(object$call$random))) # variables in the random formula
# # include <- unique(c(attr(terms.formula(object$call$fixed), "term.labels"),attr(terms.formula(object$call$random), "term.labels"))) # paste to present as A:B:C
# include <- setdiff(include, c("1","-1"))
# ##################################################
# # step 0. find all variables used in the modeling
# allTermsUsed <- unique(c(unlist(object$terms$fixed), unlist(object$terms$random)))
# # Remove 1 and -1 terms
# allTermsUsed <- allTermsUsed[which(!allTermsUsed %in% c("1", "-1"))]
# # reformulate turns a character vector into a formula object so that the terms can be pulled out
# allTermsUsed <- unique(attr(terms.formula(reformulate(allTermsUsed)), "term.labels")) # paste to present as A:B:C
# # Remove terms that include a : (i.e. interaction terms)
# allTermsUsed <- allTermsUsed[!grepl(":", allTermsUsed)]
#
# # all.vars(form)
# toAgg <- unique(unlist(strsplit(include,":")))
# toAgg <- intersect(toAgg, colnames(object$dataOriginal))
# ignored <- setdiff(allTermsUsed,toAgg)
#
# # print(head(object$dataOriginal))
# # print(toAgg)
# levelsOfTerms <- lapply(as.list(toAgg),function(x){(unique(object$dataOriginal[,x]))})
#
# DTX <- expand.grid(levelsOfTerms);
#
# colnames(DTX) <- toAgg
#
# toMerge <- unique(object$dataOriginal[,c(colnames(DTX),ignored)])
# if(!is.data.frame(toMerge)){
# toMerge <- data.frame(toMerge); colnames(toMerge) <- colnames(DTX)
# }
# toMerge[,object$terms$response[[1]]] <- 1
#
# DTX <-merge(toMerge, DTX, all.y = TRUE)
#
# if(length(object$terms$response[[1]]) < 2){
# YY = data.frame(DTX[,object$terms$response[[1]]]); colnames(YY) <- object$terms$response[[1]]
# }else{YY = DTX[,object$terms$response[[1]]]}
#
# DTX[,object$terms$response[[1]]] <- apply(YY,2,imputev)
# if(length(ignored) > 0){ # if there's ignored columns
# if(length(ignored) == 1){
# # v=which(colnames(DTX) == ignored)
# DTX[,ignored] <- imputev(DTX[,ignored])
# }else{
# for(o in 1:length(ignored)){
# DTX[,ignored[o]] <- imputev(DTX[,ignored[o]])
# }
# }
# }
# # ##################################################
# # ##################################################
# # ##################################################
#
# if(is.null(object$call$random)){
# originalModelForMatricesSE <- mmer(fixed=object$call$fixed,
# # random=object$call$random,
# rcov=object$call$rcov,
# data=object$dataOriginal, returnParam = TRUE,#reshapeOutput =FALSE,
# init = object$sigma_scaled, constraints = object$constraints,
# naMethodY = object$call$naMethodY,
# naMethodX = object$call$naMethodX,...)
# }else{
# originalModelForMatricesSE <- mmer(fixed=object$call$fixed,
# random=object$call$random,
# rcov=object$call$rcov,
# data=object$dataOriginal, returnParam = TRUE,#reshapeOutput =FALSE,
# init = object$sigma_scaled, constraints = object$constraints,
# naMethodY = object$call$naMethodY,
# naMethodX = object$call$naMethodX,...)
# }
# # ##################################################
# ## hypertable summary
# nLevels <- c(unlist(lapply(originalModelForMatricesSE$X,ncol)), unlist(lapply(originalModelForMatricesSE$Z,ncol)))
# namesLevels <- c(unlist(oto$fixed),names(object$U))
# namesLevelsO <- data.frame( x=c(unlist(oto2$fixed),unlist(oto2$random)), y=c(object$termsN$fixed, object$termsN$random))
# namesLevelsO <- as.vector(unlist(apply(namesLevelsO,1,function(x){rep(x[1],x[2])})))
# formLevels <- c(rep("fixed",length(unlist(oto$fixed))),rep("random",length(names(object$U))))
# id <- 1:length(formLevels)
# ignored <- rep(FALSE,length(id)) # all ignored by default
# include <- rep(TRUE,length(id)) # none include by default
# average <- rep(FALSE, length(id))
# predictSummary <- data.frame(namesLevelsO,namesLevels,formLevels,nLevels,id,ignored,include, average)
# colnames(predictSummary) <- c("termHL","term","type","nLevels","id","ignored","include","average")
# # ##################################################
#
# return(predictSummary)
# }
#
# if(is.null(hypertable)){
# # if user doesn't provide hypertable, we build one that :
# # 1) 'ignores' all random effects that don't match with classify
# # 2) 'averages' all fixed effects that don't match with classify
# hyp <- getHypertable(object=object, classify=classify)
# hypterm2 <- unlist(lapply(as.list(as.character(hyp$term)), function(x){x2 <- strsplit(x, split=":")[[1]];x3<-x2[length(x2)];return(x3)}))
# # print(hyp)
# weWillIgnoreRandom <- list()
# weWillAverageFixed <- list()
# for(ic in 1:length(classify)){
# weWillIgnoreRandom[[ic]]<- setdiff(which(hyp$type == "random"), grep(classify[ic],hypterm2))
# weWillAverageFixed[[ic]]<- setdiff(which(hyp$type == "fixed"), grep(classify[ic],hypterm2))
# }
# weWillIgnoreRandom <- Reduce(intersect,weWillIgnoreRandom) # ignore the ones that don't match with any classify terms
# weWillAverageFixed <- setdiff(Reduce(intersect,weWillAverageFixed),1) # ignore the ones that don't match with any classify terms
# if(length(weWillIgnoreRandom) > 0){ hyp[weWillIgnoreRandom,"ignored"]=TRUE; hyp[weWillIgnoreRandom,"include"]=FALSE}
# if(length(weWillAverageFixed) > 0){hyp[weWillAverageFixed,"average"]=TRUE}
# hypertable <- hyp
# }
# # print(hypertable)
#
# fToAverage=NULL
# classify<- unique(unlist(strsplit(classify,":")))
#
# if(is.null(classify)){
# stop("Please provide the classify argument. For fitted values use the fitted() function.",call. = FALSE)
# }
#
# oto <- oto2 <- object$terms
# oto2$fixed[[1]] <- setdiff(oto2$fixed[[1]],c("1","-1")) # get fixed factors plus intercept
# oto2$fixed <- lapply(oto2$fixed,function(x){paste(x,collapse = ":")}) # paste the fixed factors
# oto2$random <- lapply(oto2$random,function(x){paste(x,collapse = ":")}) # paste the random terms
#
# for(u in 3:length(object$terms)){ # change random terms to split by ":"
# prov <- object$terms[[u]]
# if(length(prov) > 0){
# for(v in 1:length(prov)){
# object$terms[[u]][[v]] <- unlist(strsplit(prov[[v]],":"))
# }
# }
# }
#
# # initially assume that we will include all the factors in the prediction
# include <- setdiff(unique(c(unlist(object$terms$fixed),unlist(object$terms$random))),c("1","-1"))
#
# # now use the input from the user
# if(!is.null(hypertable)){ # if user provides a hypertable, use the customization instead
# include <- setdiff(hypertable[which(hypertable$include),"termHL"],c("1","-1"))
# }
#
# include <- unique(unlist(lapply(include,function(x){y <- regmatches(x, gregexpr("(?<=\\().*?(?=\\))", x, perl=T))[[1]]; if(length(y) > 0){return(y)}else{return(x)}}))) # paste to present as A:B:C
#
# ##################################################
# # step 0. find all variables used in the modeling
# allTermsUsed <- unique(c(unlist(object$terms$fixed), unlist(object$terms$random)))
# allTermsUsed<- allTermsUsed[which(allTermsUsed!= "1")]
# allTermsUsed<- allTermsUsed[which(allTermsUsed!= "-1")]
# allTermsUsed <- unique(unlist(strsplit(allTermsUsed,":")))
# allTermsUsed <- unique(unlist(lapply(allTermsUsed,function(x){y <- regmatches(x, gregexpr("(?<=\\().*?(?=\\))", x, perl=T))[[1]]; if(length(y) > 0){return(y)}else{return(x)}}))) # paste to present as A:B:C
#
# toAgg <- unique(unlist(strsplit(include,":")))
# ignored <- setdiff(allTermsUsed,toAgg)
# # print(ignored)
#
# ## impute variables to avoid issues in dimensions
# for(i in 1:length(allTermsUsed)){
# if(is.numeric(object$dataOriginal[,allTermsUsed[i]]) | is.factor(object$dataOriginal[,allTermsUsed[i]]) ){
# object$dataOriginal[,allTermsUsed[i]] <- imputev(object$dataOriginal[,allTermsUsed[i]])
# }
# }
#
#
# levelsOfTerms <- lapply(as.list(toAgg),function(x){unique(object$dataOriginal[,x])})
# DTX <- expand.grid(levelsOfTerms);
# colnames(DTX) <- toAgg
# # toMerge <- object$dataOriginal[,c(colnames(DTX),ignored,object$terms$response[[1]])]
#
# toMerge <- unique(object$dataOriginal[,c(colnames(DTX),ignored)])
# if(!is.data.frame(toMerge)){
# toMerge <- data.frame(toMerge); colnames(toMerge) <- colnames(DTX)
# }
# toMerge[,object$terms$response[[1]]] <- 1
# # print(toMerge)
# # print(DTX)
# DTX <-merge(toMerge, DTX, all.y = TRUE) # by=intersect(colnames(DTX),colnames(toMerge)),
# # print(str(DTX))
# # DTX <-merge(DTX, object$dataOriginal[,c(colnames(DTX),ignored)], all.x = TRUE)
#
# if(length(object$terms$response[[1]]) < 2){
# YY = data.frame(DTX[,object$terms$response[[1]]]); colnames(YY) <- object$terms$response[[1]]
# }else{YY = DTX[,object$terms$response[[1]]]}
#
# DTX[,object$terms$response[[1]]] <- apply(YY,2,imputev)
# if(length(ignored) > 0){ # if there's ignored columns
# if(length(ignored) == 1){
# # v=which(colnames(DTX) == ignored)
# DTX[,ignored] <- imputev(DTX[,ignored])
# }else{
# for(o in 1:length(ignored)){
# DTX[,ignored[o]] <- imputev(DTX[,ignored[o]])
# }
# }
# }
# # print(DTX)
# # print(dim(DTX))
# # print(str(DTX))
# ##################################################
# # step 1 create all models
# # 1. extended data model (just get matrices)
# # 2. original model without reshaped output
# # 3. original model (just get matrices)
# if(is.null(object$call$random)){
# modelForMatrices <- mmer(fixed=object$call$fixed,
# # random=object$call$random,
# rcov=object$call$rcov,
# data=DTX, returnParam = TRUE,#reshape.results=TRUE,
# naMethodY = object$call$naMethodY,
# naMethodX = object$call$naMethodX,...)
# originalModel <- mmer(fixed=object$call$fixed,
# # random=object$call$random,
# rcov=object$call$rcov,
# data=object$dataOriginal, returnParam = FALSE,reshapeOutput =FALSE,
# init = object$sigma_scaled, constraints = object$constraints,
# naMethodY = object$call$naMethodY,
# naMethodX = object$call$naMethodX,...)
# originalModelForMatricesSE <- mmer(fixed=object$call$fixed,
# # random=object$call$random,
# rcov=object$call$rcov,
# data=object$dataOriginal, returnParam = TRUE,#reshapeOutput =FALSE,
# init = object$sigma_scaled, constraints = object$constraints,
# naMethodY = object$call$naMethodY,
# naMethodX = object$call$naMethodX,...)
# }else{
# modelForMatrices <- mmer(fixed=object$call$fixed,
# random=object$call$random,
# rcov=object$call$rcov,
# data=DTX, returnParam = TRUE,#reshape.results=TRUE,
# naMethodY = object$call$naMethodY,
# naMethodX = object$call$naMethodX, ...)
# originalModel <- mmer(fixed=object$call$fixed,
# random=object$call$random,
# rcov=object$call$rcov,
# data=object$dataOriginal, returnParam = FALSE,reshapeOutput =FALSE,
# init = object$sigma_scaled, constraints = object$constraints,
# naMethodY = object$call$naMethodY,
# naMethodX = object$call$naMethodX,...)
# originalModelForMatricesSE <- mmer(fixed=object$call$fixed,
# random=object$call$random,
# rcov=object$call$rcov,
# data=object$dataOriginal, returnParam = TRUE,#reshapeOutput =FALSE,
# init = object$sigma_scaled, constraints = object$constraints,
# naMethodY = object$call$naMethodY,
# naMethodX = object$call$naMethodX)
# }
# modelForMatrices$U <- originalModel$U
# modelForMatrices$PevU <- originalModel$PevU
# modelForMatrices$VarU <- originalModel$VarU
# modelForMatrices$Beta <- originalModel$Beta
# modelForMatrices$VarBeta <- originalModel$VarBeta
# modelForMatrices$Vi <- originalModel$Vi
# modelForMatrices$P <- originalModel$P
# modelForMatrices$sigma <- originalModel$sigma
# ##################################################
# # calculate Xb and get X multivariate matrices from extended and original model
# ys <- object$terms$response[[1]]
# nt <- length(ys) # number of traits
# TT <- diag(nt) # diagonal matrix
# # which fixed effects to include
# # print(object$terms$fixed[[1]])
# fToUse <- list()
# for(i in 1:length(toAgg)){
# fToUse[[i]]<-grep(toAgg[i],object$terms$fixed[[1]])
# }
# fToUse = sort(c(1,unique(unlist(fToUse))))
#
#
# # print(fToUse)
# # print(length(modelForMatrices$X))
# if(!is.null(hypertable)){ # if user provides a hypertable, use the customization instead
# fToUse <- which(hypertable$type == "fixed" & hypertable$include == TRUE)
# fNotToUse <- which(hypertable$type == "fixed" & hypertable$ignored == TRUE)
# fToAverage <- which(hypertable$type == "fixed" & hypertable$average == TRUE)
# if(length(fNotToUse) > 0){fToUse <- setdiff(fToUse,fNotToUse)}
# }
# # fToUse<- 1:3
# if(length(fToUse) > 0){ # if there's fixed effects to add
# # print(hypertable)
# if(length(fToAverage) > 0){# if there's terms to average
# for(xi in 1:length(fToAverage)){
# # we add a 1 to those terms and then divide over the number of factor levels (columns)
# if(length(which(hypertable$term == "1")) > 0){fact1<-1}else{fact1<-0}
# averFactor <- hypertable[fToAverage[xi],"nLevels"]+fact1
# modelForMatrices$X[[fToAverage[xi]]] <- ((modelForMatrices$X[[fToAverage[xi]]]+1)/(modelForMatrices$X[[fToAverage[xi]]]+1)) / averFactor
# }
# }
# # # correction for intercept if exists
# # if(hypertable$term[1] == "1"){
# # modelForMatrices$X[[1]] <- modelForMatrices$X[[1]]/averFactor
# # }
#
# # print(str(modelForMatrices$X))
# X <- do.call(cbind,modelForMatrices$X[fToUse]) # build X cbinding the ones required
# #
# # find the betas to use
# ncolsX <- unlist(lapply(modelForMatrices$X,ncol))
# # print(ncolsX)
# start=1; betas0 <- list()
# for(i in 1:length(ncolsX)){
# if(ncolsX[i] > 0){
# betas0[[i]] <- start:(start+(ncolsX[i]*nt)-1)
# start= max(betas0[[i]])+1
# }
# }
# # print(betas0)
# # print(start)
# # print(head(X))
# # print(fToUse)
# # print(start)
# X.mv.extended <- kronecker(X,TT)
# Xb=X.mv.extended%*%modelForMatrices$Beta[unlist(betas0[fToUse]),1] # calculate Xb
# # X'ViX
# XtViX = modelForMatrices$VarBeta[unlist(betas0[fToUse]),unlist(betas0[fToUse])]
# # build X multivariate from original model
# Xo <- do.call(cbind,originalModelForMatricesSE$X[fToUse])
# X.mv.original <- kronecker(Xo,TT)
# }
# ##################################################
# # calculate Zu
# pev=NULL
# cov.b.pev=NULL
# zToUse=NULL
# Z.extended=NULL
# if(!is.null(object$call$random)){
# nre <- length(object$terms$random) # number of random effects
# # identify which random terms in the model should be added
# reUsed0 <- list()
# for(i in 1:length(toAgg)){
# reUsed <- numeric()
# for(j in 1:nre){
# result <- grep(toAgg[i],object$terms$random[[j]])
# if(length(result) >0){reUsed[j] =1}else{reUsed[j] =0}
# }
# reUsed0[[i]] <- reUsed
# };
# reUsed0 <- Reduce("+",reUsed0); reUsed0[which(reUsed0 > 0)]=reUsed0[which(reUsed0 > 0)]/reUsed0[which(reUsed0 > 0)]
# # print(reUsed0)
#
# # identify which effects estimated correspond to each random term in the model
# zToUse <- list(); start=1
# for(i in 1:nre){
# zToUse[[i]] <- start:(start+object$termsN$random[i]-1)
# start <- max(zToUse[[i]])+1
# }
# zToUse <- unique(unlist(zToUse[which(reUsed0 == 1)]))
# # print(zToUse)
#
# if(!is.null(hypertable)){ # if user provides a hypertable, use the customization instead
# nFixed <- max(which(hypertable$type == "fixed"))
# # which ones the user wants to force to be included
# rForce <- which(hypertable$type == "random" & hypertable$include == TRUE)
# if(length(rForce) > 0){
# rForce <- rForce - nFixed
# zToUse <- rForce
# }
# # which ones the user wants to force to be excluded
# rNotForce <- which(hypertable$type == "random" & hypertable$ignored == TRUE)
# if(length(rNotForce) > 0){
# rNotForce <- rNotForce - nFixed
# zToUse <- setdiff(zToUse,rNotForce)
# }
# #########################
# # print(str(modelForMatrices))
# rToAverage <- which(hypertable$type == "random" & hypertable$average == TRUE) - length(which(hypertable$type == "fixed"))
# if(length(rToAverage) > 0){# if there's random terms to average
# for(zi in 1:length(rToAverage)){
# # we add a 1 to those terms and then divide over the number of factor levels (columns)
# averFactor <- hypertable[rToAverage[zi],"nLevels"]
# modelForMatrices$Z[[rToAverage[zi]]] <- ((modelForMatrices$Z[[rToAverage[zi]]]+1)/(modelForMatrices$Z[[rToAverage[zi]]]+1)) / averFactor
# }
# }
#
# }
#
# if(length(zToUse) == 0){zToUse=NULL}
# # print(zToUse)
# nz <- length(modelForMatrices$Z)
# Zu <- vector(mode = "list", length = nz) # list for Zu
# for(ir in 1:nz){ # for each random effect
# Z <- modelForMatrices$Z[[ir]] # provisional Z
# Zu[[ir]] <- kronecker(Z,TT) %*% modelForMatrices$U[[ir]] # calculate Zu
# }
# ###########################################
# ## SE step
# # cov(b,u - u.hat), xvxi, pev (C12,C11,C22)
# ## build the multivariate K and Z from the original odel to build
# # ZKfv, Xo, cov.b.pev, pev
# if(!is.null(zToUse)){ # if not only there's random terms but they are relevant for predict
# # print(originalModel$sigma)
# G.mv.original.List <- lapply(as.list((zToUse)),function(x){
# kronecker(as.matrix(originalModelForMatricesSE$K[[x]]),originalModel$sigma[,,x])
# }); G.mv.original <- do.call(adiag1,G.mv.original.List)
#
# tZ.mv.original.List <- lapply(as.list((zToUse)),function(x){
# kronecker(as.matrix(t(originalModelForMatricesSE$Z[[x]])),TT)
# }); tZ.mv.original <- do.call(rbind,tZ.mv.original.List)
#
# G.tZ.mv.original <- G.mv.original %*% tZ.mv.original # GZ' as many rows as obs, as many cols as levels
#
# if(length(fToUse) > 0){ # Cov(b,u) = 0 - (X'ViX)-X' Vi GZ
# cov.b.pev <- 0 - ( XtViX %*% t(X.mv.original) %*% originalModel$Vi %*% t(G.tZ.mv.original) )
# GZtViZG <- G.tZ.mv.original%*%originalModel$Vi%*%t(G.tZ.mv.original)
# GZtViXPXViZG <- G.tZ.mv.original %*% originalModel$Vi %*% X.mv.original %*% XtViX %*% t(X.mv.original) %*% originalModel$Vi %*% t(G.tZ.mv.original)
# VarU <- GZtViZG - GZtViXPXViZG
# pev <- G.mv.original - VarU
# }
#
# # (185 x 3)' (185 x 185) = (3 x 185) (185 x 185) (levs164 x obs185)' = (3 x 164)
# # bring the design matrices for extended model to get standard errors for the extended model
# Z.extended <- lapply(as.list((zToUse)),function(x){
# kronecker(as.matrix(t(modelForMatrices$Z[[x]])),TT)
# }); Z.extended <- do.call(rbind,Z.extended)
# tZ.extended <- t(Z.extended)
# }
# }
# ##################################################
# # add them up
# # print(zToUse)
# if(length(fToUse) > 0 & !is.null(zToUse)){ # fixed and random effects included
# y.hat <- Xb + Reduce("+",Zu[zToUse]) # y.hat = Xb + Zu.1 + ... + Zu.n
# standard.errors <- sqrt(abs(
# rowSums((X.mv.extended%*%XtViX)*X.mv.extended) +
# rowSums(2*(X.mv.extended%*%cov.b.pev)*tZ.extended) +
# rowSums((tZ.extended%*%pev)*tZ.extended)
# )
# )
# }else if( length(fToUse) == 0 & !is.null(zToUse) ){ # only random effects included
# y.hat <- Reduce("+",Zu[zToUse]) # y.hat = Xb + Zu.1 + ... + Zu.n
# standard.errors <- sqrt(abs(
# rowSums((tZ.extended%*%pev)*tZ.extended)
# )
# )
# XtViX=NULL
# }else if( length(fToUse) > 0 & is.null(zToUse) ){ # only fixed effects included
# y.hat <- Xb # y.hat = Xb
# standard.errors <- sqrt(abs(
# rowSums((X.mv.extended%*%XtViX)*X.mv.extended)
# )
# )
# }
# ##################################################
# # add y.hat to the grid
#
# if(length(allTermsUsed) == 1){
# DTX2 <- data.frame(DTX[,-which(colnames(DTX) %in% ys)])
# colnames(DTX2) <- allTermsUsed
# }else{
# DTX2 <- DTX[,-which(colnames(DTX) %in% ys)]
# }
# DTX2L <- rep(list(DTX2),nt)
# for(it in 1:nt){DTX2L[[it]]$trait <- ys[it]}
# DTX2 <- do.call(rbind, DTX2L)
# realOrder <- list()
# for(i in 1:nt){
# realOrder[[i]] <- seq(i,length(y.hat[,1]),nt)
# }
# realOrder <- unlist(realOrder)
# DTX2$predicted.value <- y.hat[realOrder,1]
# DTX2$standard.error <- standard.errors[realOrder]
# DTX2$status <- "Estimable"
# DTX2$status[which(is.nan(DTX2$standard.error))] <- "Issues"
# DTX2$standard.error[which(is.nan(DTX2$standard.error))] <- 0
#
# ##################################################
# # aggregate to desired shape
#
# myForm <- paste0("predicted.value~",paste(classify, collapse = "+"), "+trait")
# pvals <- aggregate(as.formula(myForm), FUN=mean, data=DTX2)
# myFormSE <- paste0("standard.error~",paste(classify, collapse = "+"), "+trait")
# pvalsSE <- aggregate(as.formula(myFormSE), FUN=mean, data=DTX2)
# # pvalsSE <- aggregate(as.formula(myForm), FUN=function(x){1.96 * (sd(x)/sqrt(length(x)))}, data=DTX2)
# colnames(pvalsSE)[ncol(pvalsSE)] <- "standard.error"
# # print(classify)
# pvals <- merge(pvals,pvalsSE, by=c("trait",classify))
#
# # ##################################################
# ## hypertable summary
#
# nLevels <- c(unlist(lapply(originalModelForMatricesSE$X,ncol)), unlist(lapply(originalModelForMatricesSE$Z,ncol)))
# namesLevels <- c(unlist(oto$fixed),names(object$U))
#
# namesLevelsO <- data.frame( x=c(unlist(oto2$fixed),unlist(oto2$random)), y=c(object$termsN$fixed, object$termsN$random))
# namesLevelsO <- unlist(apply(namesLevelsO,1,function(x){rep(x[1],x[2])}))
#
# formLevels <- c(rep("fixed",length(unlist(oto$fixed))),rep("random",length(names(object$U))))
# id <- 1:length(formLevels)
# ignored <- rep(TRUE,length(id)) # all ignored by default
# include <- rep(FALSE,length(id)) # none include by default
# average <- rep(FALSE, length(id))
#
# include[fToUse]=TRUE # specify which fixed effects where used
# if(!is.null(fToAverage)){
# average[fToAverage]=TRUE # specify which fixed effects where used
# }
# if(!is.null(zToUse)){include[length(modelForMatrices$X)+zToUse]=TRUE} # specify which random effects where used
#
# ignored[fToUse]=FALSE # specify which fixed effects ARE NOT ignored
# if(!is.null(zToUse)){ignored[length(modelForMatrices$X)+zToUse]=FALSE} # specify which random effects ARE NOT ignored
#
# predictSummary <- data.frame(namesLevelsO,namesLevels,formLevels,nLevels,id,ignored,include, average)
# colnames(predictSummary) <- c("termHL","term","type","nLevels","id","ignored","include","average")
#
# toreturn2 <- list(pvals=pvals,
# hypertable=predictSummary,
# model=modelForMatrices,
# C11=XtViX,
# C12=cov.b.pev,
# C22=pev,
# Xextended=X.mv.extended,
# Zextended=Z.extended
#
# )
# attr(toreturn2, "class")<-c("predict.mmer", "list")
#
# return(toreturn2)
# }
Any scripts or data that you put into this service are public.
Add the following code to your website.
For more information on customizing the embed code, read Embedding Snippets.