leehuwuj / olh
Open source stack lakehouse
☆25Updated last year
Alternatives and similar repositories for olh:
Users that are interested in olh are comparing it to the libraries listed below
- A Python Library to support running data quality rules while the spark job is running⚡☆181Updated last week
- Repo for everything open table formats (Iceberg, Hudi, Delta Lake) and the overall Lakehouse architecture☆60Updated 3 months ago
- Building Data Lakehouse by open source technology. Support end to end data pipeline, from source data on AWS S3 to Lakehouse, visualize a…☆25Updated last year
- Playground for Lakehouse (Iceberg, Hudi, Spark, Flink, Trino, DBT, Airflow, Kafka, Debezium CDC)☆56Updated last year
- A Python package to submit and manage Apache Spark applications on Kubernetes.☆41Updated last week
- Delta Lake helper methods. No Spark dependency.☆23Updated 7 months ago
- A Table format agnostic data sharing framework☆38Updated last year
- A Micosoft Power BI Custom Connector allowing you to import Trino data into Power BI.☆68Updated 3 months ago
- Quick Guides from Dremio on Several topics☆70Updated 3 months ago
- Code snippets for Data Engineering Design Patterns book☆78Updated last month
- Pythonic Programming Framework to orchestrate jobs in Databricks Workflow☆215Updated this week
- Trino dbt demo project to mix and load BigQuery data with and in a local PostgreSQL database☆74Updated 3 years ago
- A write-audit-publish implementation on a data lake without the JVM☆46Updated 8 months ago
- trino monitoring with JMX metrics through Prometheus and Grafana☆13Updated 8 months ago
- ☆262Updated 5 months ago
- Delta Lake examples☆221Updated 6 months ago
- Delta lake and filesystem helper methods☆51Updated last year
- Adapter for dbt that executes dbt pipelines on Apache Flink☆94Updated last year
- Apache Hive Metastore as a Standalone server in Docker☆72Updated 7 months ago
- Delta Lake helper methods in PySpark☆322Updated 7 months ago
- Yet Another (Spark) ETL Framework☆20Updated last year
- Sample Data Lakehouse deployed in Docker containers using Apache Iceberg, Minio, Trino and a Hive Metastore. Can be used for local testin…☆64Updated last year
- Performance Observability for Apache Spark☆248Updated 2 weeks ago
- Sample code to collect Apache Iceberg metrics for table monitoring☆26Updated 8 months ago
- The Lakehouse Engine is a configuration driven Spark framework, written in Python, serving as a scalable and distributed engine for sever…☆243Updated 2 months ago
- ☆79Updated last year
- New generation opensource data stack☆66Updated 2 years ago
- PyJaws: A Pythonic Way to Define Databricks Jobs and Workflows☆43Updated 9 months ago
- A Docker Compose template that builds a interactive development environment for PySpark with Jupyter Lab, MinIO as object storage, Hive M…☆43Updated 4 months ago
- Low Cost, Simple and Scalable Way of Data Replication to Apache Iceberg/Cloud/Data Lake☆243Updated this week