AbsaOSS / atum
A dynamic data completeness and accuracy library at enterprise scale for Apache Spark
☆30Updated 2 weeks ago
Related projects ⓘ
Alternatives and complementary repositories for atum
- Dynamic Conformance Engine☆31Updated this week
- Extensible streaming ingestion pipeline on top of Apache Spark☆44Updated 7 months ago
- Scala API for Apache Spark SQL high-order functions☆14Updated last year
- type-class based data cleansing library for Apache Spark SQL☆79Updated 5 years ago
- A library that brings useful functions from various modern database management systems to Apache Spark☆56Updated last year
- Bulletproof Apache Spark jobs with fast root cause analysis of failures.☆72Updated 3 years ago
- Waimak is an open-source framework that makes it easier to create complex data flows in Apache Spark.☆75Updated 6 months ago
- Nested array transformation helper extensions for Apache Spark☆36Updated last year
- Schema Registry integration for Apache Spark☆39Updated 2 years ago
- Code and examples of how to write and deploy Apache Spark Plugins. Spark plugins allow runnig custom code on the executors as they are in…☆85Updated 7 months ago
- JSON schema parser for Apache Spark☆81Updated 2 years ago
- Circus Train is a dataset replication tool that copies Hive tables between clusters and clouds.☆86Updated 8 months ago
- Custom state store providers for Apache Spark☆93Updated 2 years ago
- Sample processing code using Spark 2.1+ and Scala☆51Updated 4 years ago
- Lighthouse is a library for data lakes built on top of Apache Spark. It provides high-level APIs in Scala to streamline data pipelines an…☆60Updated 2 months ago
- Filling in the Spark function gaps across APIs☆50Updated 3 years ago
- Flowchart for debugging Spark applications☆101Updated last month
- ☆63Updated 5 years ago
- A Spark-based data comparison tool at scale which facilitates software development engineers to compare a plethora of pair combinations o…☆48Updated 10 months ago
- Spark package to "plug" holes in data using SQL based rules ⚡️ 🔌☆28Updated 4 years ago
- Spark-Radiant is Apache Spark Performance and Cost Optimizer☆25Updated 2 years ago
- Spark Structured Streaming State Tools☆34Updated 4 years ago
- Template for Spark Projects☆101Updated 6 months ago
- Smart Automation Tool for building modern Data Lakes and Data Pipelines☆111Updated this week
- A framework for creating composable and pluggable data processing pipelines using Apache Spark, and running them on a cluster.☆47Updated 8 years ago
- ☆25Updated 3 years ago
- Scalable CDC Pattern Implemented using PySpark☆18Updated 5 years ago
- Magic to help Spark pipelines upgrade☆34Updated last month