- The default
cached_connection()
helper will configure a temporary storage location by default. It also now supports all options supported byduckdb::duckdb()
for connection creation. - New
as_dataset()
utility copies a local in-memory data.frame into the connection.
- bugfix: reading from local disk recursively no longer requires manual
**
. Also, trying to read from an existing local file won't try and append recursive search even when given the default recursive=TRUE option. - bugfix:
open_dataset()
uses random table name by default, avoid naming collisions.
- bugfix
write_dataset()
no longer adds**
into paths when writing some partitions. - Protect from unsupported table names generated from file names that start with a digit, fixes #21.
-
open_dataset()
gains the ability to read spatial vector data formats (objects read bysf
) usingformat="sf"
-
default geometry column in
to_sf()
is now termedgeom
, to match the default used induckdb
'sst_read()
function. -
open_dataset()
now tries to guess the data format instead of defaulting to parquet when no format is explicitly provided. -
a new function,
spatial_join()
, allows a variety of spatial joins. -
a new function,
st_read_meta()
, exposes the spatial metadata of remote spatial objects. -
new helper function,
as_view()
, creates a temporary view of a query.
write_dataset()
now understands lazy queries, not just lazy tables.
-
duckdbfs now has spatial data query support! Users can leverage spatial data operations like
st_distance()
andst_area()
and request return values assf
objects. Supports network-based access too. See README.md -
Added
write_dataset()
which can write to (potentially partitioned) parquet to local directories or remote (S3) buckets. -
The S3 interface supports
arrow
-compatible URI notation:- Alternate endpoints can now be passed like so
s3://userid:secret_token@bucket-name?endpoint_override=data.ecoforecast.org
- Users can omit the use of
*
(match any file) or**
(recursive search) and just supply a path. Recursive search is then assumed automatically. Note: unlikearrow
, this still supports the use of globs (*
) elsewhere in the path, e.g.s3://bucket/*/path
- Alternate endpoints can now be passed like so
-
duckdb_s3_config
gains argumentanonymous
allowing users to ignore existing AWS keys that may be set in environmental variables or AWS configuration files. This can also be passed as the username position in URI notation, e.g.s3://anonymous@bucket_name
. -
open_dataset
drops use ofendpoint
as an argument. Instead, alternative S3 endpoints can be set either by using the URI query notation or callingduckdb_s3_config()
first. Additionally, any arguments toduckdb_s3_config()
, includings3_endpoint
, can now be passed toopen_dataset
through the...
. Note these settings will override any set by the URI notation.
- Initial release to CRAN