linkedin insight
Omax Tech

Loading...

Amazon Aurora MySQL Zero-ETL Integration set up with Amazon RedShift

Amazon Aurora MySQL Zero-ETL Integration set up with Amazon RedShift

Cloud/DevOps
Sep 23, 2024
7-8 min

Share blog

Introduction

AWS Zero-ETL Integration is a streamlined approach that eliminates the traditional Extract, Transform, Load (ETL) process for replicating data between services. Instead of manually replicating data between different databases or services, Zero-ETL enables seamless, near real-time data sharing and integration without requiring extensive data pipelines.

Amazon work station announced Amazon Aurora zero-ETL integration with Amazon Redshift at AWS re:Invent 2022 and is now available for Aurora MySQL 3.05.0 (compatible with MySQL 8.0.32) and higher versions in multiple regions.

Benefits of Zero-ETL

In the organization’s data strategy, zero-ETL offers many benefits.

Agility

Zero-ETL streamlines data architecture and minimizes the need for extensive data engineering. It enables the addition of new data sources without requiring the reprocessing of large datasets. This flexibility boosts agility, fostering data-driven decision-making and accelerating innovation.

Cost efficiency

Zero-ETL uses cloud native and scalable integration techniques which allows businesses to optimize cost based on their actual usage. It also reduces infrastructure costs, development efforts, and maintenance overheads.

Real time insights

Zero-ETL provides near-real-time data access, ensuring freshed data for analytics, AI/ML, and reporting. It guarantees accurate and timely insights for use cases like real-time dashboards, optimized gaming experience, data quality monitoring, and customer behavior analysis.

In this post, a step-by-step guidance is provided on how to set up zero-ETL between Amazon Aurora and Amazon RedShift.

You need an Amazon Aurora MySQL cluster up and running, if you haven’t set up, please follow our blog for Creating an Amazon Aurora DB Cluster

Configure the Aurora MySQL source with a customized DB cluster parameter group

Once Amazon Aurora MySQL cluster is setup, we need to create a custom parameter group with the following parameters,

On the Amazon RDS console, create a DB cluster parameter group called

javascript
zero-etl-database
React Query features

Zero-ETL integrations require specific values for the Aurora DB cluster parameters that control binary logging (binlog). For example, enhanced binlog mode must be turned on

javascript
(aurora_enhanced_binlog=1).

2. Set the following binlog cluster parameter settings:

javascript
1. binlog_backup=0
2. binlog_replication_globaldb=0
3. binlog_format=ROW
4. aurora_enhanced_binlog=1
5. binlog_row_metadata=FULL
6. binlog_row_image=FULL

In addition, make sure that the binlog_transaction_compression parameter is not set to ON, and that the binlog_row_value_options parameter is not set to PARTIAL_JSON. By default, these parameters are not set.

3. Select Save changes.

React Query features
React Query features

Configure AWS RedShift destination with enabling case sensitive identifiers

We have set up the source DB cluster, now we need to configure Amazon RedShift as target data warehouse with the following requirements;

  • 1
    Must be an RA3 node type (ra3.16xlarge, ra3.4xlarge, or ra3.xlplus), or RedShift Serverless
  • 2
    Must be Encrypted, only in case of provisioned cluster

For our use case, we will be creating a Redshift Serverless data warehouse;

  • 1
    On the Amazon Redshift console, choose Serverless dashboard in the navigation pane.
  • 2
    Choose Create workgroup.

The following screenshot shows the default settings for Amazon Redshift Serverless.

React Query features
React Query features
React Query features
React Query features

3. For Workgroup name, enter custom-wg. Choose base RPU capacity in 8 to 512 range available in increments of 8.

React Query features
React Query features

4. For Namespace, select Create a new namespace and enter custom-ns, also provide database name, Admin username, and password of your own choice.

React Query features
React Query features

Keep all the rest as the default.

React Query features
React Query features

After setup completes, choose Continue to go to your Serverless dashboard. You can see that the serverless workgroup and namespace are available.

React Query features

Create zero-ETL integration

To create the zero-ETL integration, complete the following steps:

1. On the Amazon RDS console, In the navigation pane choose Zero-ETL integrations and then on the Zero-ETL integration page, choose Create zero-ETL integration

React Query features
React Query features

2. For Integration identifier, enter a name, for example zero-etl-integration-rds.

React Query features

3. For Source database, choose Browse RDS databases and choose the source cluster, in our case the cluster name is database-1 and choose Next.

React Query features

4. Under Target, for Amazon Redshift data warehouse, choose Browse Redshift data warehouses and choose the Redshift Serverless destination namespace (custom-ns).

React Query features

Add tags and encryption is an optional.

React Query features

5. On the Review and create page, verify the integration name, source, target and other settings. When confirmed, choose Create zero-ETL integration.

React Query features
React Query features

Choose the integration to view the details and monitor its progress. It takes a few minutes to change the status from Creating to Active depending on the size of the dataset already available in the source.

React Query features

Create a database from the integration in Amazon Redshift

To create your database, follow the steps:

1. On the Redshift Serverless dashboard, navigate to the custom-ns namespace.

2. Choose Query data and click on Query Editor v2 to open the query editor.

3. Connect to the Redshift Serverless data warehouse by choosing Create connection.

React Query features

4. Run the following query to obtain the integration_id from the svv_integration system table:

javascript
$ select integration_id from svv_integration;

5. Use the integration_id from the previous step to create a new database from the integration:

javascript
$ CREATE DATABASE database-1-rs FROM INTEGRATION '';

The integration is now complete, and an entire snapshot of the source will reflect as is in the destination. Ongoing changes will also be synced in near-real time.

Blogs

Discover the latest insights and trends in technology with the Omax Tech Blog.

View All Blogs
Startup MVP architecture illustration with rocket and analytics icons.
6-8 min
Feb 25, 2026

Why Building the Right MVP Architecture No Longer Slows You Down

Just build a simple monolith for your MVP. You can fix the architecture later...

Read More
Modern AI development cycle showing code, system, and automation flow.
4-6 min
Feb 11, 2026

AI-Assisted MVP Development (Vibe Coding)

Building a startup MVP used to be slow, expensive, and stressful especially if you weren’t technical....

Read More
Illustration showing SEO evolving into AEO and GEO, with search, analytics, and automation icons representing QA teams driving AI search visibility
4-6 min
Feb 2, 2026

From SEO to AEO & GEO: Why QA Teams Will Own Search Visibility in the AI Era

Search is no longer just a list of links. It’s becoming a decision layer, A place where users expect an immediate, synthesized answer, a recommendation, or a next action...

Read More
Amazon EventBridge logo representing AWS event-driven architecture service
4-6 min
Feb 2, 2026

Common Amazon EventBridge Pitfalls in Production (and How to Avoid Them)

Amazon EventBridge simplifies the implementation of event-driven architectures. Publish an event, configure a rule, attach a target-and the system appears to work seamlessly...

Read More
Digital network concept with interconnected computer icons over a glowing circuit board background.
8-10 min
Jan 28, 2026

Building Production-Ready RAG Microservices: A Complete Serverless Architecture Guide

Large Language Models like GPT-4 and Claude have a critical flaw for businesses: they don't know your proprietary data. They can't answer questions about your products...

Read More
Illustration showing a modern data lakehouse architecture with interconnected data servers and centralized data processing.
4-6 min
Jan 22, 2026

What is a Data Lake, Data Warehouse, and Data Lakehouse? - A Simple Beginner’s Guide

Data has become one of the most valuable assets for modern businesses. Every click, transaction, message, and app interaction generates information that companies want to store, analyze, and learn from....

Read More
AWS cloud architecture diagram showing core services and infrastructure
4-6 min
Jan 19, 2026

Implementing a Scalable AWS Landing Zone: A Practical Guide for DevOps Teams

An AWS Landing Zone is a well-architected, multi-account AWS environment designed to support scalability, security, compliance, and operational excellence from day one....

Read More
Abstract illustration of scalable cloud servers representing modern distributed system architecture.
4-6 min
Jan 19, 2026

Using EventBridge for Async Communication in a Serverless Microservice Architecture

Microservices often begin with simple, synchronous communication: Service A calls Service B’s API and waits for a response...

Read More
illustration of an Amazon DynamoDB database on a blue background, representing pros and cons of using DynamoDB.
4-6 min
Jan 16, 2026

Pros and cons of using DynamoDB

Amazon DynamoDB has become one of the most popular NoSQL databases in the cloud, offering a fully managed, serverless experience....

Read More

Get In Touch

Build Your Next Big Idea with Us

From MVPs to full-scale applications, we help you bring your vision to life on time and within budget. Our expert team delivers scalable, high-quality software tailored to your business goals.