Manage VantageCloud Lake Compute Clusters with Apache Airflow
Overview
This tutorial demonstrates the process of utilizing the Teradata Airflow Compute Cluster Operators to manage VantageCloud Lake compute clusters. The objective is to execute dbt transformations defined on jaffle_shop dbt project through VantageCloud Lake compute clusters.
Use The Windows Subsystem for Linux (WSL) on Windows
to try this quickstart example.
Prerequisites
- Ensure you have the necessary credentials and access rights to use Teradata VantageCloud Lake.
- Python 3.8, 3.9, 3.10 or 3.11 and python3-env, python3-pip installed.
- Linux
- WSL
- macOS
Run in Powershell:
Refer Installation Guide if you face any issues.
Install Apache Airflow and Astronomer Cosmos
-
Create a new python environment to manage airflow and its dependencies. Activate the environment:
노트This will install Apache Airflow as well.
-
Install the Apache Airflow Teradata provider
-
Set the AIRFLOW_HOME environment variable.
Install dbt
-
Create a new python environment to manage dbt and its dependencies. Activate the environment:
-
Install
dbt-teradata
anddbt-core
modules: