```{=html}
For a more detailed explanation of `SPOTlight` consider looking at our manuscript: > Elosua-Bayes M, Nieto P, Mereu E, Gut I, Heyn H. SPOTlight: seeded NMF regression to deconvolute spatial transcriptomics spots with single-cell transcriptomes. *Nucleic Acids Res.* **2021;49(9):e50**. doi: [10.1093](10.1093/nar/gkab043) # Load packages {.unnumbered} ```r library(ggplot2) library(SPOTlight) library(SingleCellExperiment) library(SpatialExperiment) library(scater) library(scran)
SPOTlight
?SPOTlight
is a tool that enables the deconvolution of cell types and cell type
proportions present within each capture location comprising mixtures of cells.
Originally developed for 10X's Visium - spatial transcriptomics - technology, it
can be used for all technologies returning mixtures of cells.
SPOTlight
is based on learning topic profile signatures, by means of an NMFreg
model, for each cell type and finding which combination of cell types fits best
the spot we want to deconvolute. Find below a graphical abstract visually
summarizing the key steps.
The minimal unit of data required to run SPOTlight
are:
Data inputs can also be objects of class r Biocpkg("SpatialExperiment")
(SE),
or r Biocpkg("SingleCellExperiment")
(SCE).
For this vignette, we will use a SE put out by 10X Genomics containing a Visium kidney slide. The raw data can be accessed here.
SCE data comes from the The Tabula Muris Consortium which contains >350,000 cells from from male and female mice belonging to six age groups, ranging from 1 to 30 months. From this dataset we will only load the kidney subset to map it to the Visium slide.
Both datasets are available through Biocondcutor packages and can be loaded into R as follows. ` Load the spatial data:
library(TENxVisiumData) spe <- MouseKidneyCoronal() # Use symbols instead of Ensembl IDs as feature names rownames(spe) <- rowData(spe)$symbol
Load the single cell data. Since our data comes from the Tabula Muris Sensis dataset, we can directly load the SCE object as follows:
library(TabulaMurisSenisData) sce <- TabulaMurisSenisDroplet(tissues = "Kidney")$Kidney
Quick data exploration:
table(sce$free_annotation, sce$age)
We see how there is a good representation of all the cell types across ages except at 24m. In order to reduce the potential noise introduced by age and batch effects we are going to select cells all coming from the same age.
# Keep cells from 18m mice sce <- sce[, sce$age == "18m"] # Keep cells with clear cell type annotations sce <- sce[, !sce$free_annotation %in% c("nan", "CD45")]
If the dataset is very large we want to downsample it to train the model, both in of number of cells and number of genes. To do this, we want to keep a representative amount of cells per cluster and the most biologically relevant genes.
In the paper we show how downsampling the number of cells per cell identity to \~100 doesn't affect the performance of the model. Including >100 cells per cell identity provides marginal improvement while greatly increasing computational time and resources. Furthermore, restricting the gene set to the marker genes for each cell type along with up to 3.000 highly variable genes further optimizes performance and computational resources. You can find a more detailed explanation in the original paper.
Our first step is to get the marker genes for each cell type. We follow the Normalization procedure as described in OSCA. We first carry out library size normalization to correct for cell-specific biases:
sce <- logNormCounts(sce)
We aim to identify highly variable genes that drive biological heterogeneity. By feeding these genes to the model we improve the resolution of the biological structure and reduce the technical noise.
# Get vector indicating which genes are neither ribosomal or mitochondrial genes <- !grepl(pattern = "^Rp[l|s]|Mt", x = rownames(sce)) dec <- modelGeneVar(sce, subset.row = genes) plot(dec$mean, dec$total, xlab = "Mean log-expression", ylab = "Variance") curve(metadata(dec)$trend(x), col = "blue", add = TRUE) # Get the top 3000 genes. hvg <- getTopHVGs(dec, n = 3000)
Next we obtain the marker genes for each cell identity. You can use whichever
method you want as long as it returns a weight indicating the importance of that
gene for that cell type. Examples include avgLogFC
, AUC
, pct.expressed
,
p-value
...
colLabels(sce) <- colData(sce)$free_annotation # Compute marker genes mgs <- scoreMarkers(sce, subset.row = genes)
Then we want to keep only those genes that are relevant for each cell identity:
mgs_fil <- lapply(names(mgs), function(i) { x <- mgs[[i]] # Filter and keep relevant marker genes, those with AUC > 0.8 x <- x[x$mean.AUC > 0.8, ] # Sort the genes from highest to lowest weight x <- x[order(x$mean.AUC, decreasing = TRUE), ] # Add gene and cluster id to the dataframe x$gene <- rownames(x) x$cluster <- i data.frame(x) }) mgs_df <- do.call(rbind, mgs_fil)
Next, we randomly select at most 100 cells per cell identity. If a cell type is comprised of \<100 cells, all the cells will be used. If we have very biologically different cell identities (B cells vs. T cells vs. Macrophages vs. Epithelial) we can use fewer cells since their transcriptional profiles will be very different. In cases when we have more transcriptionally similar cell identities we need to increase our N to capture the biological heterogeneity between them.
In our experience we have found that for this step it is better to select the cells from each cell type from the same batch if you have a joint dataset from multiple runs. This will ensure that the model removes as much signal from the batch as possible and actually learns the biological signal.
For the purpose of this vignette and to speed up the analysis, we are going to use 20 cells per cell identity:
# split cell indices by identity idx <- split(seq(ncol(sce)), sce$free_annotation) # downsample to at most 20 per identity & subset # We are using 5 here to speed up the process but set to 75-100 for your real # life analysis n_cells <- 5 cs_keep <- lapply(idx, function(i) { n <- length(i) if (n < n_cells) n_cells <- n sample(i, n_cells) }) sce <- sce[, unlist(cs_keep)]
You are now set to run SPOTlight
to deconvolute the spots!
Briefly, here is how it works:
NMF is used to factorize a matrix into two lower dimensionality matrices
without negative elements. We first have an initial matrix V (SCE count matrix),
which is factored into W and H. Unit variance normalization by gene is performed
in V and in order to standardize discretized gene expression levels,
‘counts-umi’. Factorization is then carried out using the non-smooth NMF method,
implemented in the R package NMF r CRANpkg("NMF")
(14). This method is
intended to return sparser results during the factorization in W and H, thus
promoting cell-type-specific topic profile and reducing overfitting during
training. Before running factorization, we initialize each topic, column,
of W with the unique marker genes for each cell type with weights. In turn, each
topic of H in SPOTlight
is initialized with the corresponding membership of each
cell for each topic, 1 or 0. This way, we seed the model with prior information,
thus guiding it towards a biologically relevant result. This initialization also
aims at reducing variability and improving the consistency between runs. \
NNLS regression is used to map each capture location's transcriptome in V’ (SE count matrix) to H’ using W as the basis. We obtain a topic profile distribution over each capture location which we can use to determine its composition. \
we obtain Q, cell-type specific topic profiles, from H. We select all cells from the same cell type and compute the median of each topic for a consensus cell-type-specific topic signature. We then use NNLS to find the weights of each cell type that best fit H’ minimizing the residuals.
You can visualize the above explanation in the following workflow scheme:
res <- SPOTlight( x = sce, y = spe, groups = as.character(sce$free_annotation), mgs = mgs_df, hvg = hvg, weight_id = "mean.AUC", group_id = "cluster", gene_id = "gene")
Alternatively you can run SPOTlight
in two steps so that you can have the trained model. Having the trained model allows you to reuse with other datasets you also want to deconvolute with the same reference. This allows you to skip the training step, the most time consuming and computationally expensive.
mod_ls <- trainNMF( x = sce, y = spe, groups = sce$type, mgs = mgs, weight_id = "weight", group_id = "type", gene_id = "gene") # Run deconvolution res <- runDeconvolution( x = spe, mod = mod_ls[["mod"]], ref = mod_ls[["topic"]])
Extract data from SPOTlight
:
# Extract deconvolution matrix head(mat <- res$mat)[, seq_len(3)] # Extract NMF model fit mod <- res$NMF
In the next section we show how to visualize the data and interpret
SPOTlight
's results.
We first take a look at the Topic profiles. By setting facet = FALSE
we want to
evaluate how specific each topic signature is for each cell identity.
Ideally each cell identity will have a unique topic profile associated to it as
seen below.
plotTopicProfiles( x = mod, y = sce$free_annotation, facet = FALSE, min_prop = 0.01, ncol = 1) + theme(aspect.ratio = 1)
Next we also want to ensure that all the cells from the same cell identity share
a similar topic profile since this will mean that SPOTlight
has learned a
consistent signature for all the cells from the same cell identity.
plotTopicProfiles( x = mod, y = sce$free_annotation, facet = TRUE, min_prop = 0.01, ncol = 6)
Lastly we can take a look at which genes the model learned for each topic.
Higher values indicate that the gene is more relevant for that topic.
In the below table we can see how the top genes for Topic1
are characteristic
for B cells (i.e. Cd79a, Cd79b, Ms4a1...).
library(NMF) sign <- basis(mod) colnames(sign) <- paste0("Topic", seq_len(ncol(sign))) head(sign) # This can be dynamically visualized with DT as shown below # DT::datatable(sign, fillContainer = TRUE, filter = "top")
See here for additional graphical parameters.
plotCorrelationMatrix(mat)
Now that we know which cell types are found within each spot we can make a graph representing spatial interactions where cell types will have stronger edges between them the more often we find them within the same spot.
See here for additional graphical parameters.
plotInteractions(mat, which = "heatmap", metric = "prop") plotInteractions(mat, which = "heatmap", metric = "jaccard") plotInteractions(mat, which = "network")
We can also visualize the cell type proportions as sections of a pie chart for
each spot. You can modify the colors as you would a standard r CRANpkg("ggplot2")
.
ct <- colnames(mat) mat[mat < 0.1] <- 0 # Define color palette # (here we use 'paletteMartin' from the 'colorBlindness' package) paletteMartin <- c( "#000000", "#004949", "#009292", "#ff6db6", "#ffb6db", "#490092", "#006ddb", "#b66dff", "#6db6ff", "#b6dbff", "#920000", "#924900", "#db6d00", "#24ff24", "#ffff6d") pal <- colorRampPalette(paletteMartin)(length(ct)) names(pal) <- ct plotSpatialScatterpie( x = spe, y = mat, cell_types = colnames(mat), img = FALSE, scatterpie_alpha = 1, pie_scale = 0.4) + scale_fill_manual( values = pal, breaks = names(pal))
With the image underneath - we are rotating it 90 degrees counterclockwise and mirroring across the horizontal axis to show how to align if the spots don't overlay the image.
plotSpatialScatterpie( x = spe, y = mat, cell_types = colnames(mat), img = FALSE, scatterpie_alpha = 1, pie_scale = 0.4, # Rotate the image 90 degrees counterclockwise degrees = -90, # Pivot the image on its x axis axis = "h") + scale_fill_manual( values = pal, breaks = names(pal))
Lastly we can also take a look at how well the model predicted the proportions for each spot. We do this by looking at the residuals of the sum of squares for each spot which indicates the amount of biological signal not explained by the model.
spe$res_ss <- res[[2]][colnames(spe)] xy <- spatialCoords(spe) spe$x <- xy[, 1] spe$y <- xy[, 2] ggcells(spe, aes(x, y, color = res_ss)) + geom_point() + scale_color_viridis_c() + coord_fixed() + theme_bw()
sessionInfo()
Add the following code to your website.
For more information on customizing the embed code, read Embedding Snippets.