phdata / sdc-api-toolLinks
A set of utilities to help with management of Streamsets pipelines.
☆13Updated 7 years ago
Alternatives and similar repositories for sdc-api-tool
Users that are interested in sdc-api-tool are comparing it to the libraries listed below
Sorting:
- File compaction tool that runs on top of the Spark framework.☆59Updated 6 years ago
- kafka-connect-s3 : Ingest data from Kafka to Object Stores(s3)☆95Updated 6 years ago
- Read - Write JSON SerDe for Apache Hive.☆21Updated 6 years ago
- Hadoop MapReduce tool to convert Avro data files to Parquet format.☆34Updated 12 years ago
- Spark cloud integration: tests, cloud committers and more☆19Updated 4 months ago
- SQL for Kafka Connectors☆98Updated last year
- Circus Train is a dataset replication tool that copies Hive tables between clusters and clouds.☆88Updated last year
- Spark structured streaming with Kafka data source and writing to Cassandra☆62Updated 5 years ago
- ☆26Updated 5 years ago
- JSON schema parser for Apache Spark☆81Updated 2 years ago
- A small project to show how to add lineage to Atlas when using Spark as ETL tool☆12Updated 8 years ago
- Kafka to Avro Writer based on Apache Beam. It's a generic solution that reads data from multiple kafka topics and stores it on in cloud s…☆25Updated 4 years ago
- Apache Spark ETL Utilities☆40Updated 7 months ago
- JDBC driver for Apache Kafka☆87Updated 3 years ago
- Data pipeline automation tool☆26Updated last year
- ☆14Updated 8 years ago
- Shunting Yard is a real-time data replication tool that copies data between Hive Metastores.☆20Updated 3 years ago
- Ambari Service definition for deploying R & RHadoop libraries☆18Updated 9 years ago
- A Spark-based data comparison tool at scale which facilitates software development engineers to compare a plethora of pair combinations o…☆51Updated last year
- This will help you to generate AVRO schema from JSON schema.☆34Updated 2 years ago
- Schema Registry integration for Apache Spark☆40Updated 2 years ago
- Waimak is an open-source framework that makes it easier to create complex data flows in Apache Spark.☆75Updated last year
- A bridge to Apache Atlas for provenance metadata created in course of using Apache NiFi☆15Updated 2 years ago
- Spark stream from kafka(json) to s3(parquet)☆15Updated 6 years ago
- A High Performance Cluster Consumer for Kafka that creates Avro (boom) files in Hadoop in time based directory paths☆42Updated 9 years ago
- Apache-Spark based Data Flow(ETL) Framework which supports multiple read, write destinations of different types and also support multiple…☆26Updated 3 years ago
- This repository is to help with the Partner Demonstration of the Apache Atlas project.☆30Updated 9 years ago
- A library for querying Druid data sources with Apache Spark☆23Updated 4 years ago
- Demonstrates NiFi template deployment and configuration via a REST API☆70Updated 8 years ago
- ☆26Updated 8 years ago