hortonworks-spark / cloud-integration
Spark cloud integration: tests, cloud committers and more
☆19Updated 3 months ago
Alternatives and similar repositories for cloud-integration
Users that are interested in cloud-integration are comparing it to the libraries listed below
Sorting:
- Spark Structured Streaming State Tools☆34Updated 4 years ago
- Schema Registry integration for Apache Spark☆40Updated 2 years ago
- Sample processing code using Spark 2.1+ and Scala☆52Updated 4 years ago
- Rocksdb state storage implementation for Structured Streaming.☆17Updated 4 years ago
- Bulletproof Apache Spark jobs with fast root cause analysis of failures.☆72Updated 4 years ago
- Circus Train is a dataset replication tool that copies Hive tables between clusters and clouds.☆88Updated last year
- Spark structured streaming with Kafka data source and writing to Cassandra☆62Updated 5 years ago
- Yet Another Spark SQL JDBC/ODBC server based on the PostgreSQL V3 protocol☆34Updated 2 years ago
- Spark stream from kafka(json) to s3(parquet)☆15Updated 6 years ago
- A library to expose more of Apache Spark's metrics system☆146Updated 5 years ago
- Custom state store providers for Apache Spark☆92Updated 3 months ago
- functionstest☆33Updated 8 years ago
- type-class based data cleansing library for Apache Spark SQL☆78Updated 5 years ago
- Streaming Analytics platform, built with Apache Flink and Kafka☆35Updated last year
- Low level integration of Spark and Kafka☆130Updated 7 years ago
- The iterative broadcast join example code.☆69Updated 7 years ago
- Apache Spark and Apache Kafka integration example☆124Updated 7 years ago
- A small project to show how to add lineage to Atlas when using Spark as ETL tool☆12Updated 8 years ago
- File compaction tool that runs on top of the Spark framework.☆59Updated 6 years ago
- Big Data Toolkit for the JVM☆146Updated 4 years ago
- A library you can include in your Spark job to validate the counters and perform operations on success. Goal is scala/java/python support…☆109Updated 7 years ago
- A dynamic data completeness and accuracy library at enterprise scale for Apache Spark☆30Updated 6 months ago
- Waimak is an open-source framework that makes it easier to create complex data flows in Apache Spark.☆75Updated last year
- Hadoop output committers for S3☆109Updated 4 years ago
- Flink Examples☆39Updated 9 years ago
- A Spark-based data comparison tool at scale which facilitates software development engineers to compare a plethora of pair combinations o…☆51Updated last year
- ☆54Updated 10 years ago
- Cascading on Apache Flink®☆54Updated last year
- Scripts for parsing / making sense of yarn logs☆52Updated 8 years ago
- Profiler for large-scale distributed java applications (Spark, Scalding, MapReduce, Hive,...) on YARN.☆126Updated 6 years ago