Documentation / Product / Integrations / Amazon Web Services / Amazon Redshift

Amazon Redshift: Import Table

Import user data directly from your AWS Redshift database into Lytics, resulting in new user profiles or updates to fields on existing profiles.

Integration Details

  • Implementation Type: Server-side.
  • Implementation Technique: PostgreSQL connection.
  • Frequency: Batch with frequency configurable from hourly to weekly.
  • Resulting data: User profiles.

This integration connects to your Redshift table through a PostgreSQL connection and then completes the following steps:

  1. Create a temporary table containing a snapshot of the database to query against.
    • Adds a column for consistent ordering of data.
    • Only rows that have a timestamp after the last import, or Since Date will be included.
  2. Query a batch of rows from the temporary table.
  3. Emits rows to the data stream.
  4. Repeat steps 2 and 3 until the entire temporary table is read.

Fields

Because of the nature of Redshift data, there is no default mapping defined. By default the data will be added to the redshift_table_name data stream, where table_name is the name of the Redshift table imported.

Configuration

Follow these steps to set up and configure an import Redshift table job for Amazon Web Services (AWS) in the Lytics platform. If you are new to creating jobs in Lytics, see the Jobs Dashboard documentation for more information.

  1. Select Amazon Web Services (AWS) from the list of providers.
  2. Select the Import Redshift Table job type from the list.
  3. Select the Authorization you would like to use or create a new one.
  4. Enter a Label to identify this job you are creating in Lytics.
  5. (Optional) Enter a Description for further context on your job. aws_redshift_import_config
  6. Complete the configuration steps for your job.
  7. (Optional) In the DB URL text box, enter the Redshift URL to connect to. Leave blank to use the authorization's setting. Follow these instructions to obtain your DB URL; copy the Endpoint from the general information page (the area as the JDBC URL in the Amazon instructions). This will be of the form: redshift-cluster-name.VCP-cluster.region.redshift.amazonaws.com:port/db-name
  8. (Optional) From the SSL mode input, select the SSL mode to use to connect to the database. Leave blank to use the authorization's setting.
  9. From the Table input, select the Redshift table to import data from.
  10. From the Timestamp Column input, select the timestamp column to order the events.
  11. (Optional) From the Record Timestamp Column input, select the timestamp column to use as event timestamps, if left blank the Timestamp Column will be used.
  12. (Optional) From the Since Timestamp input, enter the earliest timestamp to import records from; only records with a Timestamp Column after this date will be imported. Use yyyy-mm-dd HH:MM:SS UTC format.
  13. (Optional) In the Stream Override text box, enter the data stream name you want to add data to. if the data stream does not exist, it will be created. If left blank the data will go to the redshift_TABLE data stream, where TABLE is the name of the Redshift table.
  14. (Optional) Select the Keep Updated checkbox to repeatedly run this import on a schedule.
  15. (Optional) From the Import Frequency input, select how often a repeated import should run.
  16. (Optional) From the Time of Day input, select the time of day to start import. Only applies to Daily, Weekly, or Monthly import frequencies.
  17. (optional) From the Timezone input, select the timezone for time of day.
  18. (Optional) In the Query Timeout numeric field, enter the maximum time a query is allowed to run in minutes.
  19. Click Start Import.