If you are unfamiliar with Airflow or the Astro CLI start with this tutorial.
Welcome! This repository contains an examples for data quality checks running on two tables created by two different DAGs.
To run the repository you will need to configure a Snowflake connection. With at least the following:
- conn_id: snowflake_conn
- conn_type: Snowflake
- account: your Snowflake account
- database: your Snowflake database
- schema: your Snowflake schema
- login: your Snowflake login
- password: your Snowflake password
- location: your Snowflake location (e.g. us-east-01)
The repository contains 3 dags in the dags folder. create_table_1_dag
will create table_1
in Snowflake, create_table_2_dag
will create table_2
. data_quality_checks_dag
is triggered once the other two dags have completed (see: Datasets and Data-Aware Scheduling in Airflow) and will run a set of data quality checks on both tables. All checks are set up to pass.