Output to Amazon Redshift

Prerequisites

Ensure that you have a Redshift connection with the correct permissions to write to your target table. Additionally, this target table and columns should already exist within Redshift before writing to it using Upsolver.

You also need a storage connection that has access to the bucket you would like the job to use to store the intermediate files used while running the job.

Finally, you should also have a staging table created previously that contains the data you intend to write to Redshift.

Create a job writing to Redshift

After you have fulfilled the prerequisites, you can create an INSERT job as follows:

CREATE JOB load_data_to_redshift
    START_FROM = BEGINNING
    SKIP_FAILED_FILES = TRUE
    FAIL_ON_WRITE_ERROR = FALSE
AS INSERT INTO REDSHIFT <redshift_connection>.<schema_name>.<target_table_name> 
    MAP_COLUMNS_BY_NAME            
    SELECT orderid AS app_name
    FROM <glue_catalog_name>.<database_name>.<table_name>
    WHERE time_filter();

This example only uses a subset of all job options available when writing to Redshift. Depending on your use case, you may want to configure a different set of options.


Learn More

For the full list of job options with syntax and detailed descriptions, see the transformation job options for Amazon Redshift.


See the INSERT SQL command reference for more details and examples.

Last updated