Description Usage Arguments Examples
Create a table from scratch, guessing the table schema
1 2 3 4 5 6 7 8 9 | rs_create_table(df, dbcon, table_name = deparse(substitute(df)),
split_files, bucket = Sys.getenv("AWS_BUCKET_NAME"),
region = Sys.getenv("AWS_DEFAULT_REGION"),
access_key = Sys.getenv("AWS_ACCESS_KEY_ID"),
secret_key = Sys.getenv("AWS_SECRET_ACCESS_KEY"),
session_token = Sys.getenv("AWS_SESSION_TOKEN"),
iam_role_arn = Sys.getenv("AWS_IAM_ROLE_ARN"), wlm_slots = 1,
sortkeys, sortkey_style = "compound", distkey,
distkey_style = "even", compression = T, additional_params = "")
|
df |
a data frame |
dbcon |
an RPostgres/RJDBC connection to the redshift server |
table_name |
the name of the table to create |
split_files |
optional parameter to specify amount of files to split into. If not specified will look at amount of slices in Redshift to determine an optimal amount. |
bucket |
the name of the temporary bucket to load the data. Will look for AWS_BUCKET_NAME on environment if not specified. |
region |
the region of the bucket. Will look for AWS_DEFAULT_REGION on environment if not specified. |
access_key |
the access key with permissions for the bucket. Will look for AWS_ACCESS_KEY_ID on environment if not specified. |
secret_key |
the secret key with permissions fot the bucket. Will look for AWS_SECRET_ACCESS_KEY on environment if not specified. |
session_token |
the session key with permissions for the bucket, this will be used instead of the access/secret keys if specified. Will look for AWS_SESSION_TOKEN on environment if not specified. |
iam_role_arn |
an iam role arn with permissions fot the bucket. Will look for AWS_IAM_ROLE_ARN on environment if not specified. This is ignoring access_key and secret_key if set. |
wlm_slots |
amount of WLM slots to use for this bulk load http://docs.aws.amazon.com/redshift/latest/dg/tutorial-configuring-workload-management.html |
sortkeys |
Column or columns to sort the table by |
sortkey_style |
Sortkey style, can be compound or interleaved http://docs.aws.amazon.com/redshift/latest/dg/t_Sorting_data-compare-sort-styles.html |
distkey |
Distkey column, can only be one, if chosen the table is distributed among clusters according to a hash of this column's value. |
distkey_style |
Distkey style, can be even or all, for the key distribution use the distkey parameter. http://docs.aws.amazon.com/redshift/latest/dg/t_Distributing_data.html |
compression |
Add encoding for columns whose compression algorithm is easy to guess, for the rest you should upload it to Redshift and run ANALYZE COMPRESSION |
additional_params |
Additional params to send to the COPY statement in Redshift |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 | library(DBI)
a=data.frame(a=seq(1,10000), b=seq(10000,1))
## Not run:
con <- dbConnect(RPostgres::Postgres(), dbname="dbname",
host='my-redshift-url.amazon.com', port='5439',
user='myuser', password='mypassword',sslmode='require')
rs_create_table(df=a, dbcon=con, table_name='testTable',
bucket="my-bucket", split_files=4)
## End(Not run)
|
Add the following code to your website.
For more information on customizing the embed code, read Embedding Snippets.