airscholar / ApacheFlink-SalesAnalyticsLinks
This repository contains an end-to-end data engineering project using Apache Flink, focused on performing sales analytics. The project demonstrates how to ingest, process, and analyze sales data, showcasing the capabilities of Apache Flink for big data processing.
☆11Updated 2 years ago
Alternatives and similar repositories for ApacheFlink-SalesAnalytics
Users that are interested in ApacheFlink-SalesAnalytics are comparing it to the libraries listed below
Sorting:
- This project shows how to capture changes from postgres database and stream them into kafka☆39Updated last year
- This repository contains an Apache Flink application for real-time sales analytics built using Docker Compose to orchestrate the necessar…☆46Updated 2 years ago
- This project showcases how to integrate the world of DevOps, focusing on Continuous Integration (CI) and Continuous Deployment (CD) with …☆14Updated 2 years ago
- Apache Airflow advanced functionalities examples☆21Updated last year
- This repository contains the code for a realtime election voting system. The system is built using Python, Kafka, Spark Streaming, Postgr…☆45Updated 2 years ago
- Simple stream processing pipeline☆110Updated last year
- A custom end-to-end analytics platform for customer churn☆11Updated 7 months ago
- In this project, we setup and end to end data engineering using Apache Spark, Azure Databricks, Data Build Tool (DBT) using Azure as our …☆37Updated 2 years ago
- This project serves as a comprehensive guide to building an end-to-end data engineering pipeline using TCP/IP Socket, Apache Spark, OpenA…☆44Updated 2 years ago
- Analytics engineering with dbt - projects and developer environment☆22Updated last year
- ☆15Updated 2 years ago
- End-to-end data platform: A PoC Data Platform project utilizing modern data stack (Spark, Airflow, DBT, Trino, Lightdash, Hive metastore,…☆47Updated last year
- End to end data engineering project☆57Updated 3 years ago
- A data engineering project with Airflow, dbt, Terrafrom, GCP and much more!☆25Updated 3 years ago
- End-to-end data pipeline that ingests, processes, and stores data. It uses Apache Airflow to schedule scripts that fetch data from an API…☆20Updated last year
- This project demonstrates how to use Apache Airflow to submit jobs to Apache spark cluster in different programming laguages using Python…☆46Updated last year
- Realtime Data Engineering Project☆30Updated 11 months ago
- Git Repo for EDW Best Practice Assets on the Lakehouse☆16Updated 2 years ago
- ☆18Updated 2 years ago
- ☆70Updated last week
- Sample project to demonstrate data engineering best practices☆203Updated last year
- A sample implementation of stream writes to an Iceberg table on GCS using Flink and reading it using Trino☆22Updated 3 years ago
- Code to demonstrate data engineering metadata & logging best practices☆21Updated last year
- An end-to-end data engineering pipeline that orchestrates data ingestion, processing, and storage using Apache Airflow, Python, Apache Ka…☆303Updated 10 months ago
- Data Engineering with Scala, published by Packt☆27Updated last year
- End-to-end data platform leveraging the Modern data stack☆52Updated last year
- Code for blog at https://www.startdataengineering.com/post/python-for-de/☆92Updated last year
- A demonstration of an ELT (Extract, Load, Transform) pipeline☆31Updated last year
- Repo for everything open table formats (Iceberg, Hudi, Delta Lake) and the overall Lakehouse architecture☆126Updated 2 months ago
- Data Engineering examples for Airflow, Prefect; dbt for BigQuery, Redshift, ClickHouse, Postgres, DuckDB; PySpark for Batch processing; K…☆68Updated last month