qubole / streamx
kafka-connect-s3 : Ingest data from Kafka to Object Stores(s3)
☆95Updated 5 years ago
Alternatives and similar repositories for streamx:
Users that are interested in streamx are comparing it to the libraries listed below
- Hadoop output committers for S3☆108Updated 4 years ago
- The Schema Repo is a RESTful web service for storing and serving mappings between schema identifiers and schema definitions.☆156Updated 2 years ago
- A Kafka-Connect Sink for S3 with no Hadoop dependencies.☆57Updated 2 years ago
- A library to expose more of Apache Spark's metrics system☆146Updated 5 years ago
- A Spark WordCountJob example as a standalone SBT project with Specs2 tests, runnable on Amazon EMR☆118Updated 8 years ago
- A Spark Streaming job reading events from Amazon Kinesis and writing event counts to DynamoDB☆94Updated 4 years ago
- Schedoscope is a scheduling framework for painfree agile development, testing, (re)loading, and monitoring of your datahub, lake, or what…☆95Updated 5 years ago
- Big Data Toolkit for the JVM☆146Updated 4 years ago
- Simplify getting Zeppelin up and running☆56Updated 8 years ago
- Read SparkSQL parquet file as RDD[Protobuf]☆93Updated 6 years ago
- SQL for Kafka Connectors☆98Updated last year
- Kafka Connect Cassandra Connector. This project includes source/sink connectors for Cassandra to/from Kafka.☆78Updated 8 years ago
- Low level integration of Spark and Kafka☆130Updated 7 years ago
- ☆76Updated 9 years ago
- Support Highcharts in Apache Zeppelin☆81Updated 7 years ago
- A utility for generating Oozie workflows from a YAML definition☆48Updated 6 years ago
- functionstest☆33Updated 8 years ago
- File compaction tool that runs on top of the Spark framework.☆59Updated 5 years ago
- Hadoop MapReduce tool to convert Avro data files to Parquet format.☆34Updated 11 years ago
- Bulletproof Apache Spark jobs with fast root cause analysis of failures.☆72Updated 4 years ago
- Google Dataflow Runner for Apache Flink™ (deprecated; please use the up-to-date Beam Runner)☆88Updated 8 years ago
- Circus Train is a dataset replication tool that copies Hive tables between clusters and clouds.☆88Updated last year
- Ephemeral Hadoop clusters using Google Compute Platform☆135Updated 2 years ago
- A super simple utility for testing Apache Hive scripts locally for non-Java developers.☆72Updated 8 years ago
- Spark structured streaming with Kafka data source and writing to Cassandra☆62Updated 5 years ago
- Spark cloud integration: tests, cloud committers and more☆19Updated last month
- Kafka Connect Tooling☆118Updated 3 years ago
- A library you can include in your Spark job to validate the counters and perform operations on success. Goal is scala/java/python support…☆109Updated 7 years ago
- Hadoop mapreduce job to bulk load data into Cassandra☆75Updated 2 years ago
- Google BigQuery support for Spark, SQL, and DataFrames☆155Updated 5 years ago