to_h3j()
method for streaming data to H3J format duckdb_secrets()
as more modern credential managerwrite_geo()
method, currently writes geojson #37The file was built for DuckDB version 'v1.1.3', but we can only load extensions built for DuckDB version '19864453f7'.
by using nightly repo for extensions by default. cached_connection()
helper will configure a temporary storage location by default.
It also now supports all options supported by duckdb::duckdb()
for connection creation. as_dataset()
utility copies a local in-memory data.frame into the connection.**
.
Also, trying to read from an existing local file won't try and append recursive search
even when given the default recursive=TRUE option.open_dataset()
uses random table name by default, avoid naming collisions.write_dataset()
no longer adds **
into paths when writing some partitions.open_dataset()
gains the ability to read spatial vector data formats
(objects read by sf
) using format="sf"
to_sf()
is now termed geom
, to match the default
used in duckdb
's st_read()
function.open_dataset()
now tries to guess the data format instead of defaulting to
parquet when no format is explicitly provided.
a new function, spatial_join()
, allows a variety of spatial joins.
st_read_meta()
, exposes the spatial metadata of remote spatial objects.as_view()
, creates a temporary view of a query.write_dataset()
now understands lazy queries, not just lazy tables.duckdbfs now has spatial data query support! Users can leverage spatial
data operations like st_distance()
and st_area()
and request return
values as sf
objects. Supports network-based access too. See README.md
Added write_dataset()
which can write to (potentially partitioned) parquet
to local directories or remote (S3) buckets.
The S3 interface supports arrow
-compatible URI notation:
s3://userid:secret_token@bucket-name?endpoint_override=data.ecoforecast.org
Users can omit the use of *
(match any file) or **
(recursive search) and just supply a path. Recursive search is then
assumed automatically. Note: unlike arrow
, this still supports the
use of globs (*
) elsewhere in the path, e.g. s3://bucket/*/path
duckdb_s3_config
gains argument anonymous
allowing users to ignore existing
AWS keys that may be set in environmental variables or AWS configuration files.
This can also be passed as the username position in URI notation, e.g.
s3://anonymous@bucket_name
.
open_dataset
drops use of endpoint
as an argument. Instead, alternative
S3 endpoints can be set either by using the URI query notation or calling
duckdb_s3_config()
first. Additionally, any arguments to duckdb_s3_config()
,
including s3_endpoint
, can now be passed to open_dataset
through the ...
.
Note these settings will override any set by the URI notation.
Any scripts or data that you put into this service are public.
Add the following code to your website.
For more information on customizing the embed code, read Embedding Snippets.