Skip to content

Venkatesh-admin/ELT-Pipeline-Using-DBT-Snowflake-Airflow

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

7 Commits
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

Welcome to Netflix( Movielens) ELT Pipeline

This project demonstrates an end-to-end ELT (Extract, Load, Transform) pipeline using Azure Storage for raw data ingestion, Snowflake as the cloud data warehouse, and dbt for data transformations and data modeling and airflow for orchestration.

🏗️ Architecture Diagram

Below is the high-level architecture of the pipeline:

Architecture Diagram

🗂️ Resources

Data in S3

⚙️ Snowflake

Queries to setup datawarehouse,role,user,database,schema,tables and copy the data from Azure storage Snowflake

⚙️ Key dbt Commands in development mode

Run these commands from your dbt project root folder after deveolping models

🔨 Build Models

dbt build

🏗️ Run Specific Models

dbt run --select model_name

🧪 Run Tests

dbt test

📄 Generate Documentation

dbt docs generate

🌐 Serve Documentation Locally

dbt docs serve

📊 Run Analysis Queries

dbt compile --select path:analyses/

Airflow Sample pipeline

Use SnowflakeOperator to copy data from the azure storage and DbtTaskGroup to execute dbt models

alt text

About

Netflix( Movielens) ELT Pipeline

Topics

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published