ysfesr / Building-Data-LakeHouseLinks
Creation of a data lakehouse and an ELT pipeline to enable the efficient analysis and use of data
☆46Updated last year
Alternatives and similar repositories for Building-Data-LakeHouse
Users that are interested in Building-Data-LakeHouse are comparing it to the libraries listed below
Sorting:
- ☆16Updated last year
- ☆14Updated 2 years ago
- Dockerizing an Apache Spark Standalone Cluster☆43Updated 2 years ago
- Delta-Lake, ETL, Spark, Airflow☆47Updated 2 years ago
- Building a Data Pipeline with an Open Source Stack☆55Updated 11 months ago
- Playground for Lakehouse (Iceberg, Hudi, Spark, Flink, Trino, DBT, Airflow, Kafka, Debezium CDC)☆58Updated last year
- Sample Data Lakehouse deployed in Docker containers using Apache Iceberg, Minio, Trino and a Hive Metastore. Can be used for local testin…☆72Updated last year
- This project shows how to capture changes from postgres database and stream them into kafka☆36Updated last year
- This project demonstrates how to use Apache Airflow to submit jobs to Apache spark cluster in different programming laguages using Python…☆44Updated last year
- Create a streaming data, transfer it to Kafka, modify it with PySpark, take it to ElasticSearch and MinIO☆62Updated last year
- To provide a deeper understanding of how the modern, open-source data stack consisting of Iceberg, dbt, Trino, and Hive operates within a…☆35Updated last year
- Spark data pipeline that processes movie ratings data.☆28Updated last week
- Build Data Lake using Open Source tools☆101Updated 3 weeks ago
- A workspace to experiment with Apache Spark, Livy, and Airflow in a Docker environment.☆38Updated 4 years ago
- In this project, we setup and end to end data engineering using Apache Spark, Azure Databricks, Data Build Tool (DBT) using Azure as our …☆32Updated last year
- Trino dbt demo project to mix and load BigQuery data with and in a local PostgreSQL database☆75Updated 3 years ago
- End-to-end data platform: A PoC Data Platform project utilizing modern data stack (Spark, Airflow, DBT, Trino, Lightdash, Hive metastore,…☆41Updated 8 months ago
- Simple stream processing pipeline☆102Updated last year
- Simplified ETL process in Hadoop using Apache Spark. Has complete ETL pipeline for datalake. SparkSession extensions, DataFrame validatio…☆55Updated 2 years ago
- A repository of sample code to show data quality checking best practices using Airflow.☆77Updated 2 years ago
- Materials for the next course☆24Updated 2 years ago
- Near real time ETL to populate a dashboard.☆72Updated last year
- Execution of DBT models using Apache Airflow through Docker Compose☆116Updated 2 years ago
- Series follows learning from Apache Spark (PySpark) with quick tips and workaround for daily problems in hand☆53Updated last year
- Docker with Airflow and Spark standalone cluster☆258Updated last year
- The goal of this project is to build a docker cluster that gives access to Hadoop, HDFS, Hive, PySpark, Sqoop, Airflow, Kafka, Flume, Pos…☆64Updated 2 years ago
- Simple repo to demonstrate how to submit a spark job to EMR from Airflow☆33Updated 4 years ago
- ☆18Updated last year
- End to end data engineering project☆56Updated 2 years ago
- build dw with dbt☆46Updated 8 months ago