All Projects → konosp → dbt-on-airflow

konosp / dbt-on-airflow

Licence: other
No description or website provided.

Programming Languages

python
139335 projects - #7 most used programming language
shell
77523 projects

Projects that are alternatives of or similar to dbt-on-airflow

dbt-cloud-plugin
DBT Cloud Plugin for Airflow
Stars: ✭ 35 (+16.67%)
Mutual labels:  airflow, dbt
airflow-dbt-python
A collection of Airflow operators, hooks, and utilities to elevate dbt to a first-class citizen of Airflow.
Stars: ✭ 111 (+270%)
Mutual labels:  airflow, dbt
dataops-platform-airflow-dbt
Build DataOps platform with Apache Airflow and dbt on AWS
Stars: ✭ 33 (+10%)
Mutual labels:  airflow, dbt
dbt-airflow-docker-compose
Execution of DBT models using Apache Airflow through Docker Compose
Stars: ✭ 76 (+153.33%)
Mutual labels:  airflow, dbt
airflow-dbt
Apache Airflow integration for dbt
Stars: ✭ 233 (+676.67%)
Mutual labels:  airflow, dbt
astro
Astro allows rapid and clean development of {Extract, Load, Transform} workflows using Python and SQL, powered by Apache Airflow.
Stars: ✭ 79 (+163.33%)
Mutual labels:  airflow
udacity-data-eng-proj2
A production-grade data pipeline has been designed to automate the parsing of user search patterns to analyze user engagement. Extract data from S3, apply a series of transformations and load into S3 and Redshift.
Stars: ✭ 25 (-16.67%)
Mutual labels:  airflow
AirDataComputer
Air Data Computer
Stars: ✭ 29 (-3.33%)
Mutual labels:  airflow
airflow-prometheus-exporter
Export Airflow metrics (from mysql) in prometheus format
Stars: ✭ 25 (-16.67%)
Mutual labels:  airflow
tellery
Tellery lets you build metrics using SQL and bring them to your team. As easy as using a document. As powerful as a data modeling tool.
Stars: ✭ 219 (+630%)
Mutual labels:  dbt
pre-commit-dbt
🎣 List of `pre-commit` hooks to ensure the quality of your `dbt` projects.
Stars: ✭ 149 (+396.67%)
Mutual labels:  dbt
ecs-airflow
Cloudformation templates for deploying Airflow in ECS
Stars: ✭ 37 (+23.33%)
Mutual labels:  airflow
airflow-tutorial
Use Airflow to move data from multiple MySQL databases to BigQuery
Stars: ✭ 96 (+220%)
Mutual labels:  airflow
snowflake-starter
A _simple_ starter template for Snowflake Cloud Data Platform
Stars: ✭ 31 (+3.33%)
Mutual labels:  dbt
ria-jit
Lightweight and performant dynamic binary translation for RISC–V code on x86–64
Stars: ✭ 38 (+26.67%)
Mutual labels:  dbt
airflow-user-management-plugin
A plugin for Apache Airflow that allows you to manage the users that can login
Stars: ✭ 13 (-56.67%)
Mutual labels:  airflow
kuwala
Kuwala is the no-code data platform for BI analysts and engineers enabling you to build powerful analytics workflows. We are set out to bring state-of-the-art data engineering tools you love, such as Airbyte, dbt, or Great Expectations together in one intuitive interface built with React Flow. In addition we provide third-party data into data sc…
Stars: ✭ 474 (+1480%)
Mutual labels:  dbt
airflow multi dagrun
triggering a DAG run multiple times
Stars: ✭ 74 (+146.67%)
Mutual labels:  airflow
ap-airflow
Astronomer Core Docker Images
Stars: ✭ 87 (+190%)
Mutual labels:  airflow
openverse-catalog
Identifies and collects data on cc-licensed content across web crawl data and public apis.
Stars: ✭ 27 (-10%)
Mutual labels:  airflow

dbt on Airflow

Experimental project on managing dbt models on Apache Airflow.

This project has a dependency on Adobe Analytics data being load on BigQuery as described: https://github.com/konosp/adobe-clickstream-dbt

Requirements

There two sample files in the misc/ folder; profile-demo_sample.yml and service_account_key_sample.json. There files are needed for the dbt/BigQuery configuration and for the BigQuery service account that runs within the docker container.

The files need to be renamed so that "_sample" is removed. For example:

  • profile-demo_sample.yml -> profile-demo.yml
  • service_account_key_sample.json -> service_account_key.json

Then as part of the dockefile command below, the files are utilised:

COPY misc/ /project/misc/

BigQuery permissions

In order for the Google Cloud service account to be able to run the jobs, the following permissions are needed:

  • BigQuery Data Editor
  • BigQuery Job User
  • BigQuery User

How to run

Build your docker image and use 'dbt-airflow' as image name

docker build -t dbt-airflow .

Run the Docker container, expose port 8080 and pass as an argument the dbt project url:

docker run -it --rm -p 8080:8080 dbt-airflow https://github.com/konosp/adobe-clickstream-dbt.git
Note that the project description data, including the texts, logos, images, and/or trademarks, for each open source project belongs to its rightful owner. If you wish to add or remove any projects, please contact us at [email protected].