ysfesr / Building-Data-LakeHouseLinks
Creation of a data lakehouse and an ELT pipeline to enable the efficient analysis and use of data
☆47Updated last year
Alternatives and similar repositories for Building-Data-LakeHouse
Users that are interested in Building-Data-LakeHouse are comparing it to the libraries listed below
Sorting:
- Dockerizing an Apache Spark Standalone Cluster☆43Updated 3 years ago
- Docker with Airflow and Spark standalone cluster☆261Updated 2 years ago
- Simple stream processing pipeline☆103Updated last year
- Spark data pipeline that processes movie ratings data.☆29Updated 3 weeks ago
- ☆14Updated 2 years ago
- Playground for Lakehouse (Iceberg, Hudi, Spark, Flink, Trino, DBT, Airflow, Kafka, Debezium CDC)☆60Updated last year
- One click deploy docker-compose with Kafka, Spark Streaming, Zeppelin UI and Monitoring (Grafana + Kafka Manager)☆121Updated 4 years ago
- Simplified ETL process in Hadoop using Apache Spark. Has complete ETL pipeline for datalake. SparkSession extensions, DataFrame validatio…☆56Updated 2 years ago
- Delta-Lake, ETL, Spark, Airflow☆48Updated 2 years ago
- Trino dbt demo project to mix and load BigQuery data with and in a local PostgreSQL database☆76Updated 3 years ago
- Execution of DBT models using Apache Airflow through Docker Compose☆118Updated 2 years ago
- Code for dbt tutorial☆160Updated 2 months ago
- A repository of sample code to show data quality checking best practices using Airflow.☆78Updated 2 years ago
- ☆17Updated last year
- Building a Data Pipeline with an Open Source Stack☆55Updated 2 months ago
- Open source stack lakehouse☆26Updated last year
- To provide a deeper understanding of how the modern, open-source data stack consisting of Iceberg, dbt, Trino, and Hive operates within a…☆40Updated last year
- Delta Lake examples☆226Updated 10 months ago
- ☆88Updated 2 years ago
- ETL pipeline using pyspark (Spark - Python)☆117Updated 5 years ago
- Apache Spark 3 - Structured Streaming Course Material☆122Updated 2 years ago
- In this project, we setup and end to end data engineering using Apache Spark, Azure Databricks, Data Build Tool (DBT) using Azure as our …☆33Updated last year
- Series follows learning from Apache Spark (PySpark) with quick tips and workaround for daily problems in hand☆55Updated last year
- Apche Spark Structured Streaming with Kafka using Python(PySpark)☆40Updated 6 years ago
- Code snippets for Data Engineering Design Patterns book☆150Updated 5 months ago
- A full data warehouse infrastructure with ETL pipelines running inside docker on Apache Airflow for data orchestration, AWS Redshift for …☆137Updated 5 years ago
- Data pipeline performing ETL to AWS Redshift using Spark, orchestrated with Apache Airflow☆152Updated 5 years ago
- Simple repo to demonstrate how to submit a spark job to EMR from Airflow☆34Updated 4 years ago
- Create a streaming data, transfer it to Kafka, modify it with PySpark, take it to ElasticSearch and MinIO☆63Updated 2 years ago
- Sample Data Lakehouse deployed in Docker containers using Apache Iceberg, Minio, Trino and a Hive Metastore. Can be used for local testin…☆73Updated last year