jeoffreylim / maelstromLinks
Maelstrom is an open source Kafka integration with Spark that is designed to be developer friendly, high performance (millisecond stream processing), scalable (consumes messges at Spark worker nodes), and is extremely reliable.
☆22Updated 8 years ago
Alternatives and similar repositories for maelstrom
Users that are interested in maelstrom are comparing it to the libraries listed below
Sorting:
- Spooker is a dynamic framework for processing high volume data streams via processing pipelines☆30Updated 9 years ago
- A library for strong, schema based conversion between 'natural' JSON documents and Avro☆18Updated last year
- Flink performance tests☆28Updated 5 years ago
- Cascading on Apache Flink®☆54Updated last year
- spark structured streaming via HTTP communication☆18Updated 3 years ago
- A High Performance Cluster Consumer for Kafka that creates Avro (boom) files in Hadoop in time based directory paths☆42Updated 9 years ago
- An example of using Flink for Fault-Tolerant Stream Processing☆12Updated 6 years ago
- Common utilities for Apache Kafka☆36Updated 2 years ago
- UberScriptQuery, a SQL-like DSL to make writing Spark jobs super easy☆63Updated last year
- Flink Examples☆39Updated 9 years ago
- Yet Another Spark SQL JDBC/ODBC server based on the PostgreSQL V3 protocol☆34Updated 3 years ago
- SamzaSQL: Streaming SQL implementation on top of Apache Samza and Apache Kafka☆29Updated 9 years ago
- Camus Compressor merges files created by Camus and saves them in a compressed format.☆13Updated 2 years ago
- HDFS compatible Distributed Filesystem backed Cassandra☆25Updated 10 years ago
- Schema Registry integration for Apache Spark☆40Updated 2 years ago
- Collection of generic Apache Flink operators☆17Updated 8 years ago
- ☆21Updated 2 years ago
- A set of tools to ease working with Zookeeper and Kafka.☆23Updated 9 years ago
- Sql interface to druid.☆77Updated 9 years ago
- Apache Amaterasu☆56Updated 5 years ago
- Starter project for building MemSQL Streamliner Pipelines☆32Updated 8 years ago
- Spark package to "plug" holes in data using SQL based rules ⚡️ 🔌☆29Updated 5 years ago
- Akka persistance plugin implementation with Apache Ignite☆21Updated 6 years ago
- A distributed generic query layer for Apache Kafka Interactive Queries☆26Updated 7 years ago
- Library and a Framework for building fast, scalable, fault-tolerant Data APIs based on Akka, Avro, ZooKeeper and Kafka☆25Updated 4 years ago
- SQL for Kafka Connectors☆99Updated last year
- A library for financial and time series calculations on Apache Spark☆28Updated 9 years ago
- Hadoop MapReduce tool to convert Avro data files to Parquet format.☆34Updated 12 years ago
- The Scala Rule Engine☆42Updated 4 years ago
- Streaming Analytics platform, built with Apache Flink and Kafka☆35Updated last year