anemos-io / protobeam
☆22Updated 6 years ago
Alternatives and similar repositories for protobeam:
Users that are interested in protobeam are comparing it to the libraries listed below
- Lenses.io JDBC driver for Apache Kafka☆21Updated 4 years ago
- Rokku project. This project acts as a proxy on top of any S3 storage solution providing services like authentication, authorization, shor…☆66Updated 2 months ago
- Using the Parquet file format (with Avro) to process data with Apache Flink☆14Updated 9 years ago
- Bullet is a streaming query engine that can be plugged into any singular data stream using a Stream Processing framework like Apache Stor…☆41Updated 2 years ago
- Demonstration of a Hive Input Format for Iceberg☆26Updated 4 years ago
- A small project to allow publishing data to Apache Kafka, Apache Pulsar or any other target system☆14Updated 4 years ago
- ☆26Updated 5 years ago
- Spark stream from kafka(json) to s3(parquet)☆15Updated 6 years ago
- Apache Beam Site☆29Updated this week
- Use SQL to transform your avro schema/records☆28Updated 7 years ago
- Bulletproof Apache Spark jobs with fast root cause analysis of failures.☆72Updated 4 years ago
- Fast Apache Avro serialization/deserialization library☆43Updated 4 years ago
- ☆81Updated last year
- An application that records stats about consumer group offset commits and reports them as prometheus metrics☆14Updated 6 years ago
- Kafka to Avro Writer based on Apache Beam. It's a generic solution that reads data from multiple kafka topics and stores it on in cloud s…☆25Updated 4 years ago
- A Kafka Streams process to convert __consumer_offsets to a JSON-readable topic☆13Updated 5 years ago
- Kafka Streams + Memcached (e.g. AWS ElasticCache) for low-latency in-memory lookups☆13Updated 5 years ago
- Extensible streaming ingestion pipeline on top of Apache Spark☆44Updated last year
- A dynamic data completeness and accuracy library at enterprise scale for Apache Spark☆30Updated 6 months ago
- ☆13Updated 2 years ago
- Spark UDFs to deserialize Avro messages with schemas stored in Schema Registry.☆19Updated 7 years ago
- A protobuf schema registry on steroids. It will keep track of the contracts throughout your organization, making sure no contract is brok…☆43Updated 5 years ago
- A streaming key-value store implementation using native Flink Streaming operators☆23Updated 9 years ago
- Scalable CDC Pattern Implemented using PySpark☆18Updated 5 years ago
- Paper: A Zero-rename committer for object stores☆20Updated 3 years ago
- Spark job for compacting avro files together☆12Updated 7 years ago
- Common components used across the datamountaineer kafka connect connectors☆21Updated 4 years ago
- Lab project to showcase Flink's performance differences between using a SQL query and implementing the same logic via the DataStream API☆14Updated 5 years ago
- A testing framework for Trino☆26Updated last month
- ☆14Updated 2 months ago