knitr::opts_chunk$set( collapse = TRUE, comment = "#>", fig.width = 12, fig.height = 6 ) options(tibble.print_min = 6L, tibble.print.max = 6L, digits = 3)
In this vignette a PARAFAC model is created for the Shao2019
data. This is done by first processing the count data. Subsequently, the appropriate number of components are determined. Then the PARAFAC model is created and visualized.
library(parafac4microbiome) library(dplyr) library(ggplot2) library(ggpubr)
The data cube in Shao2019$data contains unprocessed counts. The function processDataCube()
performs the processing of these counts with the following steps:
compositions::clr()
function with a pseudo-count of one (on all features, prior to selection based on sparsity).The outcome of processing is a new version of the dataset called processedShao
. Please refer to the documentation of processDataCube()
for more information.
processedShao = processDataCube(Shao2019, sparsityThreshold=0.9, considerGroups=TRUE, groupVariable="Delivery_mode", CLR=TRUE, centerMode=1, scaleMode=2)
A critical aspect of PARAFAC modelling is to determine the correct number of components. We have developed the functions assessModelQuality()
and assessModelStability()
for this purpose. First, we will assess the model quality and specify the minimum and maximum number of components to investigate and the number of randomly initialized models to try for each number of components.
Note: this vignette reflects a minimum working example for analyzing this dataset due to computational limitations in automatic vignette rendering. Hence, we only look at 1-4 components with 5 random initializations each. These settings are not ideal for real datasets. Please refer to the documentation of assessModelQuality()
for more information.
# Setup # For computational purposes we deviate from the default settings minNumComponents = 1 maxNumComponents = 4 numRepetitions = 3 # number of randomly initialized models numFolds = 5 # number of jack-knifed models maxit = 200 ctol= 1e-5 # this is a really bad setting but is needed to save computational time numCores = 1 colourCols = c("Delivery_mode", "phylum", "") legendTitles = c("Delivery mode", "Phylum", "") xLabels = c("Subject index", "Feature index", "Time index") legendColNums = c(3,5,0) arrangeModes = c(TRUE, TRUE, FALSE) continuousModes = c(FALSE,FALSE,TRUE) # Assess the metrics to determine the correct number of components qualityAssessment = assessModelQuality(processedShao$data, minNumComponents, maxNumComponents, numRepetitions, ctol=ctol, maxit=maxit, numCores=numCores)
The overview plot showcases the number of iterations, the sum-of-squared error, the CORCONDIA and the variance explained for 1-4 components.
qualityAssessment$plots$overview
The overview plots shows that we can explain 8-10% of the variation in a three-component model. That is quite low. The CORCONDIA for that number of components is well above the minimum requirement of 60. However, a four-component model yields much lower CORCONDIA values.
Next, we investigate the stability of the models when jack-knifing out samples using assessModelStability()
. This will give us more information to choose between 3 or 4 components.
stabilityAssessment = assessModelStability(processedShao, minNumComponents=1, maxNumComponents=4, numFolds=numFolds, considerGroups=TRUE, groupVariable="Delivery_mode", colourCols, legendTitles, xLabels, legendColNums, arrangeModes, ctol=ctol, maxit=maxit, numCores=numCores) stabilityAssessment$modelPlots[[1]] stabilityAssessment$modelPlots[[2]] stabilityAssessment$modelPlots[[3]] stabilityAssessment$modelPlots[[4]]
The model is stable for 1-4 components. Hence a three-component is the appropriate number of components based on the CORCONDIA score.
We have decided that a three-component model is the most appropriate for the Shao2019
dataset. We can now select one of the random initializations from the assessNumComponents()
output as our final model. We're going to select the random initialisation that corresponded the maximum amount of variation explained for three components.
numComponents = 3 modelChoice = which(qualityAssessment$metrics$varExp[,numComponents] == max(qualityAssessment$metrics$varExp[,numComponents])) finalModel = qualityAssessment$models[[numComponents]][[modelChoice]]
Finally, we visualize the model using plotPARAFACmodel()
.
plotPARAFACmodel(finalModel$Fac, processedShao, 3, colourCols, legendTitles, xLabels, legendColNums, arrangeModes, continuousModes = c(FALSE,FALSE,TRUE), overallTitle = "Shao PARAFAC model")
You will observe that the loadings for some modes in some components are all negative. This is due to sign flipping: two modes having negative loadings cancel out but describe the same thing as two positive loadings. The flipLoadings()
function automatically performs this procedure and also sorts the components by how much variation they describe.
finalModel = flipLoadings(finalModel, processedShao$data) plotPARAFACmodel(finalModel$Fac, processedShao, 3, colourCols, legendTitles, xLabels, legendColNums, arrangeModes, continuousModes = c(FALSE,FALSE,TRUE), overallTitle = "Shao PARAFAC model")
Any scripts or data that you put into this service are public.
Add the following code to your website.
For more information on customizing the embed code, read Embedding Snippets.