MartijnVisser / flink-only-sqlLinks
Traditionally, engineers were needed to implement business logic via data pipelines before business users can start using it. Using this demo, we would explain how data analysts and non-engineers can use only Flink SQL to explore and transform data into insights and actions, without writing any Java or Python code.
☆12Updated last week
Alternatives and similar repositories for flink-only-sql
Users that are interested in flink-only-sql are comparing it to the libraries listed below
Sorting:
- Examples for using Apache Flink® with DataStream API, Table API, Flink SQL and connectors such as MySQL, JDBC, CDC, Kafka.☆64Updated last year
- Kubernetes Operator for the Ververica Platform☆35Updated 2 years ago
- ☆58Updated last year
- Example of a tested Apache Flink application.☆42Updated 6 years ago
- Apache flink☆18Updated 2 years ago
- This project contains a couple of tools to analyze data around the Apache Flink community.☆18Updated last year
- Lab project to showcase Flink's performance differences between using a SQL query and implementing the same logic via the DataStream API☆14Updated 5 years ago
- ☆40Updated 2 years ago
- An example of using Flink for Fault-Tolerant Stream Processing☆12Updated 6 years ago
- Streaming data changes to a Data Lake with Debezium and Delta Lake pipeline☆75Updated 2 years ago
- An application that records stats about consumer group offset commits and reports them as prometheus metrics☆14Updated 6 years ago
- Code Samples for my Ververica Webinar "99 Ways to Enrich Streaming Data with Apache Flink"☆41Updated 3 years ago
- Demos for Nessie. Nessie provides Git-like capabilities for your Data Lake.☆29Updated 2 weeks ago
- Multi-hop declarative data pipelines☆117Updated this week
- Ecosystem website for Apache Flink☆12Updated last year
- Maelstrom is an open source Kafka integration with Spark that is designed to be developer friendly, high performance (millisecond stream …☆22Updated 8 years ago
- An implementation of the DatasourceV2 interface of Apache Spark™ for writing Spark Datasets to Apache Druid™.☆43Updated last month
- Fork of Apache Kafka implenting KIP-1150 -- Diskless Topics☆43Updated last week
- Schema Registry integration for Apache Spark☆40Updated 2 years ago
- A streaming key-value store implementation using native Flink Streaming operators☆23Updated 9 years ago
- Scalable CDC Pattern Implemented using PySpark☆18Updated 6 years ago
- Spark UDFs to deserialize Avro messages with schemas stored in Schema Registry.☆20Updated 7 years ago
- Demonstration of a Hive Input Format for Iceberg☆26Updated 4 years ago
- Yet Another Spark SQL JDBC/ODBC server based on the PostgreSQL V3 protocol☆34Updated 2 years ago
- A tool to benchmark L (loading) workloads within ETL workloads☆26Updated 2 months ago
- Spark Connector to read and write with Pulsar☆115Updated last month
- Java event logs collector for hadoop and frameworks☆40Updated 4 months ago
- JDBC driver for Apache Kafka☆86Updated 3 years ago
- Db2 JDBC connector for Trino☆19Updated 2 years ago
- 🌟 Examples of use cases that utilize Decodable, as well as demos for related open-source projects such as Apache Flink, Debezium, and Po…☆80Updated last month