Nothing
#' @title Kennard-Stone algorithm for calibration sampling
#' @description
#' \loadmathjax
#' Select calibration samples from a large multivariate data using the
#' Kennard-Stone algorithm
#' @usage
#' kenStone(X, k, metric = "mahal", pc, group,
#' .center = TRUE, .scale = FALSE, init = NULL)
#' @param X a numeric matrix.
#' @param k number of calibration samples to be selected.
#' @param metric distance metric to be used: 'euclid' (Euclidean distance) or
#' 'mahal' (Mahalanobis distance, default).
#' @param pc optional. If not specified, distance are computed in the Euclidean
#' space. Alternatively, distance are computed
#' in the principal component score space and `pc` is the number of principal
#' components retained.
#' If `pc < 1`, the number of principal components kept corresponds to the
#' number of components explaining at least (`pc * 100`) percent of the total
#' variance.
#' @param group An optional `factor` (or vector that can be coerced to a factor
#' by \code{\link{as.factor}}) of length equal to nrow(X), giving the identifier
#' of related observations (e.g. samples of the same batch of measurements,
#' samples of the same origin, or of the same soil profile). When one observation
#' is selected by the procedure all observations of the same group are removed
#' together and assigned to the calibration set. This allows to select calibration
#' points that are independent from the remaining points.
#' @param .center logical value indicating whether the input matrix should be
#' centered before Principal Component Analysis. Default set to \code{TRUE}.
#' @param .scale logical value indicating whether the input matrix should be
#' scaled before Principal Component
#' Analysis. Default set to \code{FALSE}.
#' @param init (optional) a vector of integers indicating the indices of the
#' observations/rows that are to be used as observations that must be included
#' at the first iteration of the search process. Default is \code{NULL}, i.e. no
#' fixed initialization. The function will take by default the two most distant
#' observations. If the \code{group} argument is used, then all the observations
#' in the groups covered by the \code{init} observations will be also included
#' in the \code{init} subset.
#' @return a list with the following components:
#' \itemize{
#' \item{`model`:}{ numeric vector giving the row indices of the input data
#' selected for calibration}
#' \item{`test`:}{ numeric vector giving the row indices of the remaining
#' observations}
#' \item{`pc`:}{ if the `pc` argument is specified, a numeric matrix of the
#' scaled pc scores}
#' }
#' @references
#' Kennard, R.W., and Stone, L.A., 1969. Computer aided design of experiments.
#' Technometrics 11, 137-148.
#' @examples
#' data(NIRsoil)
#' sel <- kenStone(NIRsoil$spc, k = 30, pc = .99)
#' plot(sel$pc[, 1:2], xlab = "PC1", ylab = "PC2")
#' # points selected for calibration
#' points(sel$pc[sel$model, 1:2], pch = 19, col = 2)
#' # Test on artificial data
#' X <- expand.grid(1:20, 1:20) + rnorm(1e5, 0, .1)
#' plot(X, xlab = "VAR1", ylab = "VAR2")
#' sel <- kenStone(X, k = 25, metric = "euclid")
#' points(X[sel$model, ], pch = 19, col = 2)
#' @author Antoine Stevens &
#' \href{https://orcid.org/0000-0002-5369-5120}{Leonardo Ramirez-Lopez} with
#' contributions from Thorsten Behrens and Philipp Baumann
#' @details
#' The Kennard--Stone algorithm allows to select samples with a uniform
#' distribution over the predictor space (Kennard and Stone, 1969).
#' It starts by selecting the pair of points that are the farthest apart.
#' They are assigned to the calibration set and removed from the list of points.
#' Then, the procedure assigns remaining points to the calibration set
#' by computing the distance between each unassigned points
#' \mjeqn{i_0}{i_0} and selected points \mjeqn{i}{i}
#' and finding the point for which:
#'
#' \mjdeqn{d_{selected} = \max\limits_{i_0}(\min\limits_{i}(d_{i,i_{0}}))}{d_{sel ected} = \max_{i_0}(\min_{i}(d_{i,i0}))}
#'
#' This essentially selects point \mjeqn{i_0}{i_0} which is the farthest apart from its
#' closest neighbors \mjeqn{i}{i} in the calibration set.
#' The algorithm uses the Euclidean distance to select the points. However,
#' the Mahalanobis distance can also be used. This can be achieved by performing
#' a PCA on the input data and computing the Euclidean distance on the truncated
#' score matrix according to the following definition of the Mahalanobis \mjeqn{H}{H}
#' distance:
#'
#' \mjdeqn{H_{ij}^2 = \sum_{a=1}^A (\hat t_{ia} - \hat t_{ja})^{2} / \hat \lambda_a}{H_{ij}^2 = sum_{a=1}^A (hat t_{ia} - hat t_{ja})^{2} / hat lambda_a}
#'
#' where \mjeqn{\hat t_{ia}}{hatt_{ia}} is the \mjeqn{a^{th}}{a^{th}} principal component
#' score of point \mjeqn{i}{i}, \mjeqn{\hat t_{ja}}{hatt_{ja}} is the
#' corresponding value for point \mjeqn{j}{j},
#' \mjeqn{\hat \lambda_a}{hat lambda_a} is the eigenvalue of principal
#' component \mjeqn{a}{a} and \mjeqn{A}{A} is the number of principal components
#' included in the computation.
#'
#' @seealso \code{\link{duplex}}, \code{\link{shenkWest}}, \code{\link{naes}},
#' \code{\link{honigs}}
#' @export
kenStone <- function(X,
k,
metric = "mahal",
pc,
group,
.center = TRUE,
.scale = FALSE,
init = NULL) {
if (missing(k)) {
stop("'k' must be specified")
}
if (ncol(X) < 2) {
stop("'X' must have at least 2 columns")
}
if (k < 2) {
stop("Invalid argument: 'k' must be larger than 2")
}
metric <- match.arg(metric, c("mahal", "euclid"))
if (is.data.frame(X)) {
x <- X <- as.matrix(X)
}
if (!missing(pc)) {
pca <- prcomp(X, center = .center, scale = .scale)
if (pc < 1) {
pvar <- pca$sdev^2 / sum(pca$sdev^2)
pcsum <- cumsum(pvar) < pc
if (any(pcsum)) {
pc <- max(which(pcsum)) + 1
} else {
pc <- 1
}
}
scores <- X <- pca$x[, 1:pc, drop = FALSE]
}
if (metric == "mahal") {
# Project in the Mahalanobis distance space
X <- e2m(X, sm.method = "svd")
if (!missing(pc)) {
scores <- X
}
}
m <- nrow(X)
n <- 1:m
if (k >= m) {
k <- m - 1
}
if (!missing(group)) {
if (length(group) != nrow(X)) {
stop("length(group) should be equal to nrow(X)")
}
if (!is.factor(group)) {
group <- as.factor(group)
warning("group has been coerced to a factor")
}
if (k > nlevels(group)) {
stop("k is larger the the number of groups/levels in 'group'")
}
}
distance_mat <- fastDist(X, X, "euclid")
rm(X)
gc()
if (!is.null(init)) {
# init is used to initialize the model set
is_integer <- function(x, tol = .Machine$double.eps^0.5) abs(x - round(x)) < tol
## sanity checks for init
if (!all(is_integer(init))) {
stop("Seems like you are tryning to use non-integers in the init argument")
}
init <- unique(init)
if (k < length(init)) {
stop("Invalid argument: 'k' must be larger than the length of init")
}
id <- init
if (length(init) == 1) {
id <- c(id, which.max(distance_mat[, id]))
}
} else {
# Fist two most distant points to initialize the model set
id <- c(arrayInd(which.max(distance_mat), rep(m, 2)))
}
if (!missing(group)) {
id <- which(group %in% group[id])
group <- group[-id]
}
model <- n[id]
n <- n[-id]
ini <- i <- length(model)
while (i < k) {
if (i == ini) {
distance_sub <- distance_mat[model, -model]
} # first loop
else {
distance_sub <- rbind(mins, distance_mat[nid, -model])
}
mins <- do.call(pmin.int, lapply(1:nrow(distance_sub), function(i) distance_sub[i, ]))
id <- which.max(mins)
if (!missing(group)) {
id <- which(group %in% group[id])
group <- group[-id]
}
nid <- n[id]
model <- c(model, nid)
n <- n[-id]
mins <- mins[-id]
i <- length(model)
}
if (missing(pc)) {
return(list(model = model, test = (1:m)[-model]))
} else {
return(list(model = model, test = (1:m)[-model], pc = scores))
}
}
Any scripts or data that you put into this service are public.
Add the following code to your website.
For more information on customizing the embed code, read Embedding Snippets.