# Data Integration - ETL/ELT

Extract, Transform, and Load (ETL) and Extract, Load, and Transform (ELT) are processes used in data integration and data warehousing to extract, transform, and load data from various sources into a target destination, such as a data warehouse or a data lake.

This guide will take you through sample recipes with explanations, that represent practical use cases of ELT/ETL.

# Bulk vs Batch

Bulk/Batch actions/triggers are available throughout Workato. Bulk processing gives you the ability to process large amounts of data in a single job, especially suited for ETL/ELT. Batch processing is restricted by batch sizes and memory constraints, and are generally less suitable in the context of ETL/ELT.

# Extract, Transform, and Load (ETL)

ETL begins with the extraction phase, where data is sourced from multiple heterogeneous sources, including databases, files, APIs, and web services. This raw data is then subjected to a transformation phase (cleansing, filtering, etc.) and finally loaded into a target system, typically a data warehouse.

# Sample ETL Recipe

A copy of the sample recipe can be found here (opens new window)

ETL Recipe Sample ETL Recipe

# Recipe Breakdown

The sample ETL recipe is a setup of how you can extract data from an on-prem data source (SQL Server (opens new window)), merge it together with a product catalog stored in Workato FileStorage (opens new window) and load this transformed output into a data warehouse (BigQuery (opens new window)). Additionally, Workato file streaming capabilities allows you to transfer data without having to worry about time or memory constraints. This recipe serves as a general guide when building ETL recipes, and performs basic transformations on data extracted from an on prem data source before loading them into a data warehouse.

# Setup

This recipe does not require any additional steps behind the scenes.

1

This recipe triggers on a schedule everyday at 0030. This trigger should be customized to your use case.

2

Used Generate On-Prem File URL (opens new window) to prepare an on-prem location for the subsequent files to be loaded to.

3

Used Export Query Result (Bulk) (opens new window) to extract the required Sales data from SQL Server, into the on-prem location specified in the previous step.

  • Loads data into the on-prem file created earlier
4

Used Download an On-Prem File (opens new window) to retrieve the Sales data from the on-prem location that was specified earlier.

  • Retrieves the CSV content from the on-prem file created earlier
5

Used Query Data using SQL Transformations (opens new window) to merge the Sales data with Product data retrieved from Workato FileStorage (opens new window).

  • Merge Sales and Product Data to produce a transformed output, appropriate for the business requirement
6

Used Load Data into BigQuery (opens new window) to load the transformed data into a Data Warehouse.

  • Final loading of the transformed output into a data warehouse

# Extract, Load, and Transform (ELT)

Similar to ETL, ELT starts with the extraction phase, where data is extracted from various sources. ELT focuses on loading the extracted data into a target system such as a data lake or distributed storage. Once the data is loaded, transformations occur within the target system.

# Sample ELT Recipe

The copy of the sample recipe can be found here (opens new window)

ELT Recipe Sample ELT Recipe

# Recipe Breakdown

The sample ELT recipe is a setup of how you can extract data from a cloud data source (Salesforce) into a data warehouse (Snowflake). This recipe serves as a guide when building ELT recipes, and performs the basic functionality of extracting bulk data from cloud data sources and loading them into a database.

# Setup
  • You should have set up an external stage on Snowflake with the necessary permissions. More information can be found here (opens new window).
1

Create a new recipe, using Export new/updated records (Bulk) as the trigger.

  • Configure the trigger according to your use case
  • Bulk actions are best suited for extracting large amounts of data in a single job
2

Used S3 Upload file streaming (opens new window) action to upload the extracted CSV file contents to designated S3 bucket.

  • Action streams the data extracted to the designated S3 bucket
3

Used Snowflake Replicate Schema (opens new window) action to create/update an new/existing table.

  • This step is optional. If you have an existing table that matches the schema of the staged data, this step is not necessary
  • Replicate schema ensures that the schema of the destination table in Snowflake exists and/or has an appropriate schema matching the source data.
4

Used Bulk load data to a table from an external stage (opens new window) to load data from staged files to an existing table.

  • Loads data from staged file setup in the previous step into an existing table using the COPY command
5

Data Transformation

This recipe does not show how transformation can be done in Snowflake. Data transformations can happen after loading data into a table, or even before. More information can be found here (opens new window)


Last updated: 1/29/2024, 8:20:37 AM