Nothing
# Use sparingly to minimize aws costs.
# Verify all `targets` buckets are deleted afterwards.
tar_test("pipeline continuously uploads metadata", {
skip_if_no_aws()
bucket_name <- random_bucket_name()
s3 <- paws.storage::s3()
s3$create_bucket(Bucket = bucket_name)
on.exit(aws_s3_delete_bucket(bucket_name))
expr <- quote({
tar_option_set(
resources = tar_resources(
aws = tar_resources_aws(bucket = !!bucket_name, prefix = "_targets"),
network = tar_resources_network(max_tries = 10L)
),
repository = "aws"
)
list(
tar_target(a, 1),
tar_target(
b, {
Sys.sleep(2)
a
}
),
tar_target(
c, {
Sys.sleep(2)
b
}
),
tar_target(
d, {
Sys.sleep(200)
c
}
)
)
})
expr <- tar_tidy_eval(expr, environment(), TRUE)
eval(as.call(list(`tar_script`, expr, ask = FALSE)))
R.utils::withTimeout(
expr = tar_make(seconds_meta_upload = 1, reporter = "silent"),
timeout = 30,
onTimeout = "silent"
)
tar_destroy(destroy = "local")
temp <- tempfile()
meta <- path_meta(temp)
aws_s3_download(
file = meta,
bucket = bucket_name,
key = "_targets/meta/meta",
max_tries = 3
)
out <- tar_meta(store = temp, targets_only = TRUE)
expect_equal(sort(out$name), sort(c("a", "b", "c")))
})
Any scripts or data that you put into this service are public.
Add the following code to your website.
For more information on customizing the embed code, read Embedding Snippets.