spatAOI = function(AOI){
if(inherits(AOI, c("sf", "sfc", "sfg"))){
terra::vect(AOI)
} else {
AOI
}
}
omit.na <- function(x) { x[!is.na(x)] }
#' Merge List of SpatRaster's across time
#' @description Given a list of SpatRasters with possibly shared names, merge across time
#' @param data list of names SpatRasters
#' @return data.frame with (varname, X_name, Y_name, T_name)
#' @family dap
#' @export
merge_across_time = function(data){
ll = list()
g = unique(names(data))
for (v in unique(g)) {
g_tmp = g[g == v]
ind = grepl(v, names(data))
tmp = data[ind]
n = unlist(lapply(1:length(tmp), function(x) { names(tmp[[x]]) }))
o = order(n)
tmp = rast(tmp)
tmp = tmp[[o]]
names(tmp) = n
ll[[v]] = tmp
}
ll
}
#' Get XYTV data from DAP URL
#' @param obj an OpenDap URL or NetCDF object
#' @param varname name of variable to extract. If NULL, then get all
#' @param varmeta should variable metadata be appended?
#' @return data.frame with (varname, X_name, Y_name, T_name)
#' @family dap
#' @export
dap_xyzv <- function(obj, varname = NULL, varmeta = FALSE) {
if (!inherits(obj, "NetCDF")) {
obj <- open.nc(obj)
on.exit(close.nc(obj))
}
if(is.null(varname)){
raw = suppressWarnings({
tryCatch({
nc_coord_var(obj, variable = NULL)[, c("variable", "X", "Y", "T", "Z")]
}, error = function(e){
vars = nc_vars(obj)$name
lapply(1:length(vars), FUN = function(j){
tryCatch({
nc_coord_var(obj, variable = vars[j])[, c("variable", "X", "Y", "T", "Z")]
},
error = function(x){
NULL
})
})
}) %>%
bind_rows()
})
} else {
raw = suppressWarnings({
tryCatch({
nc_coord_var(obj, variable = varname)[, c("variable", "X", "Y", "T", "Z")]
}, error = function(e){
NULL
})
})
}
raw <- raw[!apply(raw, 1, function(x) {
sum(!is.na(x)) <= 3
}), ]
raw$dim_order = NA
raw$nX = NA
raw$nY = NA
raw$nZ = NA
for(i in 1:nrow(raw)){
o = rev(var.inq.nc(obj, raw$variable[i])$dimids)
dims = nc_dims(obj, varname)$name[o + 1]
length = nc_dims(obj, varname)$length[o + 1]
if(is.na(raw$Z[i]) & length(dims) == 4){
raw$Z[i] = dims[!dims %in% as.vector(raw[i,-1])]
}
o = names(raw)[match(dims, raw[i,])]
raw$dim_order[i] = paste(o, collapse = "")
raw$nX[i] = length[which(dims == raw$X[i])]
raw$nY[i] = length[which(dims == raw$Y[i])]
#raw$nT[i] = length[which(dims == raw$T[i])]
if(is.na(raw$Z)[i]){
raw$nZ[i] = NA
} else {
raw$nZ[i] = length[which(dims == raw$Z[i])]
}
}
names(raw) <- c("varname", "X_name", "Y_name", "T_name", "Z_name", "dim_order",
"nX", "nY", "nZ")
ll <- list()
if (varmeta) {
for (i in 1:nrow(raw)) {
if (unique(nc_var(obj, raw$varname[i])$ndims) > 4) {
ll[[i]] <- NULL
warning("We do not support 5D datasets:", raw$varname[i])
} else {
ll[[i]] <- data.frame(
varname = raw$varname[i],
units = try_att(obj, raw$varname[i], "units"),
description = try_att(obj, raw$varname[i], "long_name")
)
if(is.na(ll[[i]]$description)){
ll[[i]]$description = try_att(obj, raw$varname[i], "LongName")
}
ll[[i]]$description = gsub("\\s+"," ", ll[[i]]$description)
}
}
merge(raw, do.call(rbind, ll), by = "varname")
} else {
raw
}
}
#' TryCatch around RNetCDF::att.get.nc()
#' @param nc "NetCDF" object which points to the NetCDF dataset. Found with RNetCDF::open.nc.
#' @param variable ID or name of the variable from which the attribute will be read, or "NC_GLOBAL" for a global attribute.
#' @param attribute Attribute name or ID.
#' @return Vector with a data type that depends on the NetCDF variable. For NetCDF variables of type NC_CHAR, the R type is either character or raw, as specified by argument rawchar. For NC_STRING, the R type is character. Numeric variables are read as double precision by default, but the smallest R type that exactly represents each external type is used if fitnum is TRUE.
#' @importFrom RNetCDF att.get.nc
#' @family dap
try_att <- function(nc, variable, attribute) {
tryCatch(
{
att.get.nc(nc, variable, attribute)
},
error = function(e) {
NA
}
)
}
#' Extract grid metadata from NC Pointer
#' @param URL location of data to process
#' @param X_name Name of X diminsion. If NULL it is found
#' @param Y_name Name of Y diminsion. If NULL it is found
#' @param stopIfNotEqualSpaced stop if not equal space grid
#' @return list
#' @family dap
#' @export
.resource_grid <- function(URL, X_name = NULL, Y_name = NULL, stopIfNotEqualSpaced = TRUE) {
nc = open.nc(URL)
if (is.null(X_name) | is.null(Y_name)) {
atts <- dap_xyzv(nc)
X_name <- omit.na(unique(atts$X_name))
Y_name <- omit.na(unique(atts$Y_name))
}
nc_grid_mapping <- suppressWarnings(nc_grid_mapping_atts(nc))
degree <- grepl("degree", try_att(nc, X_name, "units"), ignore.case = TRUE)
if (nrow(nc_grid_mapping) == 0) {
if (degree) {
message(paste(
"No projection information found. \n",
"Coordinate variable units are degrees so, \n",
"assuming EPSG:4326"
))
crs <- "EPSG:4326"
} else {
warning("No projection information found in nc file.")
crs <- NA
}
} else {
crs <- try(nc_gm_to_prj(nc_grid_mapping))
if (inherits(crs, "try-error")) {
crs <- NA
} else {
crs
}
}
ncols <- dim.inq.nc(nc, X_name)$len
nrows <- dim.inq.nc(nc, Y_name)$len
close.nc(nc)
if(file.exists(URL)){
xnc = open.nc(URL)
xx <- try(var.get.nc(xnc, X_name))
} else {
xurl = gsub("#fillmismatch", "", URL)
xurl = glue("{xurl}?{X_name}")
xnc = open.nc(xurl)
xx <- try(var.get.nc(xnc, X_name))
}
if (inherits(xx, "try-error")) { xx <- seq_len(ncols) }
rs <- xx[-length(xx)] - xx[-1]
if (!isTRUE(all.equal(min(rs), max(rs), tolerance = 0.025, scale = abs(min(rs))))) {
if (is.na(stopIfNotEqualSpaced)) {
warning("cells are not equally spaced; you should extract values as points")
} else if (stopIfNotEqualSpaced) {
stop("cells are not equally spaced; you should extract values as points")
}
}
if (any(xx > 180) & degree) { xx <- xx - 360 }
xrange <- c(min(xx), max(xx))
resx <- (xrange[2] - xrange[1]) / (ncols - 1)
X1 <- xx[1]
Xn <- xx[length(xx)]
close.nc(xnc)
if(file.exists(URL)){
ync = open.nc(URL)
yy <- try(var.get.nc(ync, Y_name))
} else {
yurl = gsub("#fillmismatch", "", URL)
yurl = glue("{yurl}?{Y_name}")
ync = open.nc(yurl)
yy <- try(var.get.nc(ync, Y_name))
}
if (inherits(yy, "try-error")) { yy <- seq_len(nrows) }
Y1 <- yy[1]
Yn <- yy[length(yy)]
rs <- yy[-length(yy)] - yy[-1]
if (!isTRUE(all.equal(min(rs), max(rs), tolerance = 0.025, scale = abs(min(rs))))) {
if (is.na(stopIfNotEqualSpaced)) {
warning("cells are not equally spaced; you should extract values as points")
} else if (stopIfNotEqualSpaced) {
stop("cells are not equally spaced; you should extract values as points")
}
}
yrange <- c(min(yy), max(yy))
resy <- (yrange[2] - yrange[1]) / (nrows - 1)
if (yy[1] > yy[length(yy)]) {
toptobottom <- FALSE
} else {
toptobottom <- TRUE
}
data.frame(
crs = crs,
# xmin, xmax, ymin, ymax
X1 = X1,
Xn = Xn,
Y1 = Y1,
Yn = Yn,
resX = resx,
resY = resy,
ncols = ncols,
nrows = nrows,
toptobottom = toptobottom
)
}
#' Extract time metadata from NC Pointer
#' @param URL location of data to process
#' @param T_name Name of T dimension. If NULL it is found
#' @return list
#' @family dap
#' @export
.resource_time <- function(URL, T_name = NULL) {
if (is.null(T_name)) {
nc = open.nc(URL)
atts <- dap_xyzv(nc)
T_name <- omit.na(unique(atts$T_name))
close.nc(nc)
}
if(file.exists(URL)){
nc = open.nc(URL)
time_steps <- utcal.nc(
unitstring = att.get.nc(nc, T_name, "units"),
value = var.get.nc(nc, T_name, unpack = TRUE),
type = "c"
)
} else {
url = gsub("#fillmismatch", "", URL)
nc = open.nc(glue("{url}?{T_name}"))
time_steps <- utcal.nc(
unitstring = att.get.nc(nc, T_name, "units"),
value = var.get.nc(nc, T_name, unpack = TRUE),
type = "c"
)
}
dT <- diff(time_steps)
g <- data.frame(expand.grid(unique(dT), units(dT)))
g <- g[order(g$Var1), ]
g$n <- as.numeric(table(dT))
names(g) <- c("value", "interval", "n")
if (nrow(g) > 1 & all(g$value %in% c(28, 29, 30, 31))) {
g <- data.frame(value = 1, interval = "months")
} else {
g <- g[which.max(g$n), ]
}
# If time is within 5 days of today then we call the range Open
maxDate <- ifelse(max(time_steps) >= Sys.time() - (5 * 86400) & max(time_steps) <= Sys.time() + 1,
"..",
as.character(max(time_steps))
)
nT <- ifelse(maxDate == "..", NA, length(time_steps))
int <- paste(g$value, g$interval)
if (length(int) == 0) {
int <- "0"
}
list(
duration = paste0(min(time_steps), "/", maxDate),
interval = int,
nT = nT
)
}
#' Read from a OpenDAP landing page
#' @description Reads an OpenDap resources and returns metadata
#' @param URL URL to OpenDap resource
#' @param id character. Uniquely named dataset identifier
#' @inheritParams dap_xyzv
#' @return data.frame
#' @family dap
#' @export
read_dap_file <- function(URL, varname = NULL, id, varmeta = TRUE) {
nc <- open.nc(URL)
on.exit(close.nc(nc))
raw <- dap_xyzv(obj = nc, varname, varmeta = varmeta)
raw$URL <- URL
raw$id <- id
raw <- merge(raw, data.frame(.resource_time(URL, T_name = raw$T_name[1]), id = id), by = "id")
raw <- merge(raw, .resource_grid(URL, X_name = raw$X_name[1], Y_name = raw$Y_name[1]))
raw
}
#' Read formated DAP URL as SpatRast
#' @param dap output from dap_crop
#' @return SpatRast
#' @family dap
#' @export
go_get_dap_data <- function(dap) {
tryCatch({
if (grepl("http", dap$URL)) {
var_to_terra(var = get_data(dap), dap)
} else {
var_to_terra(var = dap_to_local(dap), dap)
}
},
error = function(e) {
dap$URL
})
}
#' Convert catalog entry to extent
#' @param cat catalog entry (data.frame with an {Xn, X1, Yn, Y1, crs})
#' @return SpatExtent
#' @family dap
#' @export
make_ext <-
function(cat) {
ext(c(
min(cat$Xn, cat$X1),
max(cat$Xn, cat$X1),
min(cat$Yn, cat$Y1),
max(cat$Yn, cat$Y1)
))
}
#' Make Vector
#' @param cat catalog entry (data.frame with an {Xn, X1, Yn, Y1, crs})
#' @return SpatVect
#' @family dap
#' @export
make_vect = function (cat) { as.polygons(make_ext(cat), crs = cat$crs) }
#' Variable Array to SpatRast
#' @param var numeric array
#' @param dap dap description
#' @return SpatRast
#' @family dap
#' @export
var_to_terra <- function(var, dap) {
if(dap$startDate == dap$endDate){
dates = as.POSIXct(dap$startDate, tz = "UTC")
} else {
dates <- seq.POSIXt(as.POSIXct(dap$startDate, tz = "UTC"),
as.POSIXct(dap$endDate, tz = "UTC"),
by = dap$interval)
}
name <- gsub("_NA", "",paste(dap$variable,
dates,
dap$model,
dap$ensemble,
dap$scenario,
sep = "_"))
vars = dap$variable
if(length(vars) == 0){ vars = dap$varname }
names_ts = sub("_$", "",
gsub("__", "", gsub("_NA", "",
paste(
vars,
dap$model,
dap$ensemble,
dap$scenario,
sep = "_")))
)
if (dap$X1 == dap$Xn & dap$Y1 == dap$Yn) {
df = data.frame(date = dates, var)
names(df) = c("date", names_ts)
return(df)
}
resx <- dap$resX#(dap$Xn - dap$X1) / (dap$ncols - 1)
resy <- dap$resY#(dap$Yn - dap$Y1) / (dap$nrows - 1)
xmin <- dap$X1 - 0.5 * resx
xmax <- dap$Xn + 0.5 * resx
ymin <- dap$Y1 - 0.5 * resy
ymax <- dap$Yn + 0.5 * resy
if (length(dim(var)) == 2) {
dim(var) <- c(dim(var), 1)
}
r = rast(nrows = dap$nrows, ncols = dap$ncols,
crs = dap$crs,
nlyrs = dap$Tdim,
extent = c(
xmin = min(xmin, xmax),
xmax = max(xmin, xmax),
ymin = min(ymin, ymax),
ymax = max(ymin, ymax)
))
if(grepl("XY", dap$dim_order)){
terra::values(r) = var
} else {
r[] = var
}
if (dap$toptobottom) { r <- flip(r) }
units(r) <- dap$units
time(r) <- dates
names(r) <- name
r
}
#' Get DAP Array
#' @param dap dap description
#' @return SpatRast
#' @family dap
#' @export
get_data <- function(dap) {
nc <- open.nc(paste0(dap$URL, "#fillmismatch"))
on.exit(close.nc(nc))
var.get.nc(nc, dap$varname, unpack = TRUE)
}
#' Convert OpenDAP to start/count call
#' @param dap dap description
#' @param get should data be collected?
#' @return numeric array
#' @family dap
#' @export
dap_to_local <- function(dap, get = TRUE) {
if (nrow(dap) != 1) {
stop("This function processes only 1 DAP row at a time ... currently there are ", nrow(dap))
}
nc <- open.nc(sub("\\?.*", "", dap$URL))
on.exit(close.nc(nc))
k <- regmatches(dap$URL, gregexpr("\\[.*?\\]", dap$URL))[[1]]
k <- gsub("[", "", k, fixed = TRUE)
k <- gsub("]", "", k, fixed = TRUE)
nc_var_info <- var.inq.nc(nc, dap$varname)
X_var_info <- var.inq.nc(nc, dap$X_name)$dimids
Y_var_info <- var.inq.nc(nc, dap$Y_name)$dimids
T_var_info <- var.inq.nc(nc, dap$T_name)$dimids
dimid_order <- match(
nc_var_info$dimids,
c(T_var_info, Y_var_info, X_var_info)
)
start <- (as.numeric(sapply(strsplit(k, ":"), "[[", 1)) + 1)[dimid_order]
count <- (c(dap$Tdim, dap$nrows, dap$ncols))[dimid_order]
if (get) {
var.get.nc(nc, dap$varname,
start = start,
count = count,
unpack = TRUE
)
} else {
data.frame(
file = sub("\\?.*", "", dap$URL),
variable = dap$varname,
start = I(list(start)),
count = I(list(count)), unpack = TRUE
)
}
}
Add the following code to your website.
For more information on customizing the embed code, read Embedding Snippets.