ysfesr / Building-Data-LakeHouseLinks
Creation of a data lakehouse and an ELT pipeline to enable the efficient analysis and use of data
☆47Updated last year
Alternatives and similar repositories for Building-Data-LakeHouse
Users that are interested in Building-Data-LakeHouse are comparing it to the libraries listed below
Sorting:
- Dockerizing an Apache Spark Standalone Cluster☆43Updated 3 years ago
- Delta-Lake, ETL, Spark, Airflow☆48Updated 2 years ago
- Playground for Lakehouse (Iceberg, Hudi, Spark, Flink, Trino, DBT, Airflow, Kafka, Debezium CDC)☆60Updated last year
- Docker with Airflow and Spark standalone cluster☆261Updated 2 years ago
- ☆14Updated 2 years ago
- A repository of sample code to show data quality checking best practices using Airflow.☆78Updated 2 years ago
- Simple stream processing pipeline☆108Updated last year
- Code for dbt tutorial☆162Updated last week
- Spark data pipeline that processes movie ratings data.☆30Updated 2 weeks ago
- One click deploy docker-compose with Kafka, Spark Streaming, Zeppelin UI and Monitoring (Grafana + Kafka Manager)☆121Updated 4 years ago
- Simplified ETL process in Hadoop using Apache Spark. Has complete ETL pipeline for datalake. SparkSession extensions, DataFrame validatio…☆55Updated 2 years ago
- Open source stack lakehouse☆26Updated last year
- A full data warehouse infrastructure with ETL pipelines running inside docker on Apache Airflow for data orchestration, AWS Redshift for …☆138Updated 5 years ago
- Real-time Data Warehouse with Apache Flink & Apache Kafka & Apache Hudi☆116Updated last year
- Execution of DBT models using Apache Airflow through Docker Compose☆118Updated 2 years ago
- In this project, we setup and end to end data engineering using Apache Spark, Azure Databricks, Data Build Tool (DBT) using Azure as our …☆35Updated last year
- Building a Data Pipeline with an Open Source Stack☆54Updated 2 months ago
- Git Repo for EDW Best Practice Assets on the Lakehouse☆15Updated last year
- Trino dbt demo project to mix and load BigQuery data with and in a local PostgreSQL database☆76Updated 4 years ago
- ☆89Updated 7 months ago
- ☆88Updated 3 years ago
- Create a streaming data, transfer it to Kafka, modify it with PySpark, take it to ElasticSearch and MinIO☆63Updated 2 years ago
- A sample implementation of stream writes to an Iceberg table on GCS using Flink and reading it using Trino☆21Updated 3 years ago
- A docker using the airflow with Hadoop ecosystem (hive, spark, and sqoop)☆12Updated 4 years ago
- Example for article Running Spark 3 with standalone Hive Metastore 3.0☆100Updated 2 years ago
- ☆19Updated last year
- Data pipeline performing ETL to AWS Redshift using Spark, orchestrated with Apache Airflow☆155Updated 5 years ago
- Infraestructura para Big Data : Hadoop + NiFi +Spark + Hive usando Docker☆20Updated 2 years ago
- Apache Spark 3 - Structured Streaming Course Material☆123Updated 2 years ago
- Sample Data Lakehouse deployed in Docker containers using Apache Iceberg, Minio, Trino and a Hive Metastore. Can be used for local testin…☆74Updated 2 years ago