Amazon S3

You can connect an Amazon S3 target to the data flow, and load the ETL into Amazon S3 as a delimited Text file.

Configure the Amazon S3 Target

To load the ETL into an Amazon S3 target, add the Amazon S3 node from the Targets panel to the data flow.

From its Properties panel, select the required Amazon S3 target from the server list drop-down, then choose the required Bucket and provide the access path to the bucket. The path can be a static value pasted in the Path field, or you can create a dynamic PQL expression in the PQL editor.

Next, click ‘Connect All’ to connect the target to the data flow.

From the File Properties window, set the properties of the Amazon S3 file output:

  • Encoding: provide the character encoding set for the new database file.
  • Delimiters: select or enter the value delimiter.
  • Text Delimiter: enter the text delimiter.
  • Headers: if the file's first row is made up of column names, enable 'First row has column names'.

Target Tables

From the Target Tables window, you can rename the table outputs and change the writing type.

  • Click here to learn more about target tables.


Expand the Description window to add a description or notes to the node. The description is visible only from the Properties panel of the node, and does not produce any outputs. This is a useful way to document the ETL pipeline for yourself and other users.

Run the ETL

As there is no database or in-memory destination, the Data Model and Security stages are not relevant. Skip these steps and simply run the ETL from the Data Flow.