FINRAOS / MegaSparkDiff
A Spark-based data comparison tool at scale which facilitates software development engineers to compare a plethora of pair combinations of possible data sources. Multiple execution modes in multiple environments enable the user to generate a diff report as a Java/Scala-friendly DataFrame or as a file for future use. Comes with out of the box Spa…
☆50Updated last year
Alternatives and similar repositories for MegaSparkDiff:
Users that are interested in MegaSparkDiff are comparing it to the libraries listed below
- Bulletproof Apache Spark jobs with fast root cause analysis of failures.☆72Updated 4 years ago
- Apache-Spark based Data Flow(ETL) Framework which supports multiple read, write destinations of different types and also support multiple…☆26Updated 3 years ago
- Waimak is an open-source framework that makes it easier to create complex data flows in Apache Spark.☆75Updated 10 months ago
- Circus Train is a dataset replication tool that copies Hive tables between clusters and clouds.☆88Updated last year
- type-class based data cleansing library for Apache Spark SQL☆78Updated 5 years ago
- A Spark datasource for the HadoopOffice library☆38Updated 2 years ago
- Schema Registry integration for Apache Spark☆40Updated 2 years ago
- A dynamic data completeness and accuracy library at enterprise scale for Apache Spark☆30Updated 4 months ago
- Spark cloud integration: tests, cloud committers and more☆19Updated last month
- Lighthouse is a library for data lakes built on top of Apache Spark. It provides high-level APIs in Scala to streamline data pipelines an…☆61Updated 6 months ago
- Quark is a data virtualization engine over analytic databases.☆98Updated 7 years ago
- Shunting Yard is a real-time data replication tool that copies data between Hive Metastores.☆20Updated 3 years ago
- JSON schema parser for Apache Spark☆81Updated 2 years ago
- Provide functionality to build statistical models to repair dirty tabular data in Spark☆12Updated last year
- PySpark for ETL jobs including lineage to Apache Atlas in one script via code inspection☆18Updated 8 years ago
- A library that brings useful functions from various modern database management systems to Apache Spark☆58Updated last year
- Yet Another (Spark) ETL Framework☆20Updated last year
- Spark-Radiant is Apache Spark Performance and Cost Optimizer☆25Updated 2 months ago
- Scalable CDC Pattern Implemented using PySpark☆18Updated 5 years ago
- Apache Spark ETL Utilities☆40Updated 4 months ago
- Examples for High Performance Spark☆15Updated 4 months ago
- Spark functions to run popular phonetic and string matching algorithms☆60Updated 3 years ago
- A library for strong, schema based conversion between 'natural' JSON documents and Avro☆18Updated last year
- Spark Structured Streaming State Tools☆34Updated 4 years ago
- Splittable Gzip codec for Hadoop☆70Updated 3 weeks ago
- An implementation of the DatasourceV2 interface of Apache Spark™ for writing Spark Datasets to Apache Druid™.☆41Updated 5 months ago
- Enabling Spark Optimization through Cross-stack Monitoring and Visualization☆47Updated 7 years ago
- A framework for creating composable and pluggable data processing pipelines using Apache Spark, and running them on a cluster.☆47Updated 8 years ago
- Sample processing code using Spark 2.1+ and Scala☆51Updated 4 years ago
- Spark stream from kafka(json) to s3(parquet)☆15Updated 6 years ago