Skip to content
This repository has been archived by the owner on Apr 12, 2023. It is now read-only.

Action exposing datacoves/ci-airflow image, including dbt, dbt-coves and airflow libraries

License

Notifications You must be signed in to change notification settings

datacoves/ci-airflow-action

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

73 Commits
 
 
 
 
 
 
 
 
 
 

Repository files navigation

ci-airflow-action

A GitHub Action to @run dbt and dbt-coves commands, as well as test Apache Airflow DAGs integrity in a Docker container. You can use dbt commands such as run, test and debug. This action captures the dbt console output for use in subsequent steps.

Supported versions

Library Version
dbt 1.1.1
dbt-coves 1.1.1a1
apache-airflow 2.3.1
dag-factory 0.8.0-32
airflow-dbt 0.4.0-2

Usage

    - name: DBT Run
      uses: datacoves/[email protected]
      with:
        command: "dbt run --profiles-dir ."
      env:
        DBT_USER: ${{ secrets.DBT_USER }}
        DBT_PASSWORD: ${{ secrets.DBT_PASSWORD }}

Outputs

The result of the dbt command is either failed or passed and is saved into the result output if you want to use it in a next step:

    - name: DBT Run
      id: dbt-run
      uses: datacoves/[email protected]
      with:
        command: "dbt run --profiles-dir ."
      env:
        DBT_USER: ${{ secrets.DBT_USER }}
        DBT_PASSWORD: ${{ secrets.DBT_PASSWORD }}
    - name: Get the result
      if: ${{ always() }}
      run: echo "${{ steps.dbt-run.outputs.result }}"
      shell: bash

The result output is also saved in the DBT_RUN_STATE environment variable. The location of the dbt console log output can be accessed via the environment variable DBT_LOG_PATH. See the "Suggested workflow" section on how to use these.

General Setup

This action assumes that your dbt project is in the top-level directory of your repo, such as this sample dbt project. If your dbt project files are in a folder, you can specify it as such:

    - name: DBT Run
      uses: datacoves/[email protected]
      with:
        command: "dbt run --profiles-dir ."
        project_folder: "dbt_project"
      env:
        DBT_USER: ${{ secrets.DBT_USER }}
        DBT_PASSWORD: ${{ secrets.DBT_PASSWORD }}

Important: dbt projects use a profiles.yml file to connect to your dataset. ci-airflow-action currently requires .config/profiles.yml to be in your repo unless changed using the --profiles-dir argument or the DBT_PROFILES_DIR environment variable.

Airflow validation usage

    env:
      DATACOVES__REPO_PATH: /path/to/repo
      DATACOVES__YAML_DAGS_FOLDER: /path/to/airflow/yaml/files
      AIRBYTE__EXTRACT_LOCATION: /path/to/airbyte/extracted/files
      AIRFLOW__CORE__DAGS_FOLDER: /path/to/airflow/dags/location
      AIRFLOW__CORE__DAGBAG_IMPORT_TIMEOUT: 300
      TEST_MODE: "True" # Necessary to avoid connection issues between Airflow and Airbyte

    steps:
      - name: Test DAG structure integrity (DagBag Loading)
        uses: datacoves/[email protected]
        with:
            command: "python /usr/app/load_dagbag.py"
      
      - name: Test DBT Sources against DAGs' YAML files
        uses: datacoves/[email protected]
        with:
          dbt_project_folder: "transform"
          command: "python /usr/app/test_dags.py"
        

Thanks

This repo was inspired on https://github.com/mwhitaker/dbt-action. Thanks to Michael Whitaker.

Suggested workflow and other tips

Here is a sample workflow that sends dbt console logs by email.

License

MIT

About

Action exposing datacoves/ci-airflow image, including dbt, dbt-coves and airflow libraries

Resources

License

Stars

Watchers

Forks

Packages

No packages published

Contributors 3

  •  
  •  
  •