Aiven-Labs / python-fake-data-producer-for-apache-kafka
The Python fake data producer for Apache Kafka® is a complete demo app allowing you to quickly produce JSON fake streaming datasets and push it to an Apache Kafka topic.
☆82Updated 8 months ago
Alternatives and similar repositories for python-fake-data-producer-for-apache-kafka:
Users that are interested in python-fake-data-producer-for-apache-kafka are comparing it to the libraries listed below
- A repository of sample code to show data quality checking best practices using Airflow.☆74Updated last year
- Sample project that use Dagster, dbt, DuckDB and Dash to visualize car and motorcycle Spanish market☆54Updated 2 years ago
- A Series of Notebooks on how to start with Kafka and Python☆154Updated last year
- Make simple storing test results and visualisation of these in a BI dashboard☆40Updated 3 weeks ago
- Streaming Synthetic Sales Data Generator: Streaming sales data generator for Apache Kafka, written in Python☆44Updated 2 years ago
- ☆14Updated 11 months ago
- Apache Flink (Pyflink) and Related Projects☆29Updated 7 months ago
- Docker envinroment to stream data from Kafka to Iceberg tables☆24Updated 10 months ago
- Pipeline definitions for managing data flows to power analytics at MIT Open Learning☆41Updated this week
- Evaluation Matrix for Change Data Capture☆24Updated 5 months ago
- Full stack data engineering tools and infrastructure set-up☆47Updated 3 years ago
- Materials of the Official Helm Chart Webinar☆27Updated 3 years ago
- Code snippets for Data Engineering Design Patterns book☆49Updated last week
- New generation opensource data stack☆65Updated 2 years ago
- Delta Lake Documentation☆48Updated 7 months ago
- A Python package that creates fine-grained dbt tasks on Apache Airflow☆62Updated 3 months ago
- A CLI tool to streamline getting started with Apache Airflow™ and managing multiple Airflow projects☆204Updated this week
- Learn how to add data validation and documentation to a data pipeline built with dbt and Airflow.☆166Updated last year
- ☆19Updated 3 years ago
- Execution of DBT models using Apache Airflow through Docker Compose☆113Updated 2 years ago
- Simple stream processing pipeline☆94Updated 7 months ago
- dbt + Trino demo project, using TPC-H sample data☆19Updated 9 months ago
- Pylint plugin for static code analysis on Airflow code☆91Updated 4 years ago
- ☆62Updated this week
- Fast iterative local development and testing of Apache Airflow workflows☆195Updated last month
- A simple and easy to use Data Quality (DQ) tool built with Python.☆49Updated last year
- Example for article Running Spark 3 with standalone Hive Metastore 3.0☆97Updated last year
- Trino dbt demo project to mix and load BigQuery data with and in a local PostgreSQL database☆71Updated 3 years ago
- Repository of helm charts for deploying DataHub on a Kubernetes cluster☆168Updated this week