PritomDas / Real-Time-Streaming-Data-Pipeline-and-Dashboard
Building Real Time Data Pipeline using Apache Kafka, Apache Spark, Hadoop, PostgreSQL, Django and Flexomonster on Docker to track status of Servers in the Data Center across the Globe.
☆20Updated 4 years ago
Alternatives and similar repositories for Real-Time-Streaming-Data-Pipeline-and-Dashboard:
Users that are interested in Real-Time-Streaming-Data-Pipeline-and-Dashboard are comparing it to the libraries listed below
- This repository contains an Apache Flink application for real-time sales analytics built using Docker Compose to orchestrate the necessar…☆41Updated last year
- This is my Apache Airflow Local development setup on Windows 10 WSL2/Mac using docker-compose. It will also include some sample DAGs and …☆27Updated last year
- Spark, Airflow, Kafka☆26Updated last year
- ☆29Updated last year
- This repository contains the code for a realtime election voting system. The system is built using Python, Kafka, Spark Streaming, Postgr…☆34Updated last year
- ☆13Updated 2 years ago
- ☆14Updated 5 years ago
- A data pipeline with Kafka, Spark Streaming, dbt, Docker, Airflow, and GCP!☆11Updated last year
- ☆37Updated last year
- In this project, we setup and end to end data engineering using Apache Spark, Azure Databricks, Data Build Tool (DBT) using Azure as our …☆26Updated last year
- This project demonstrates how to use Apache Airflow to submit jobs to Apache spark cluster in different programming laguages using Python…☆38Updated 11 months ago
- Project for real-time anomaly detection using Kafka and python☆59Updated 2 years ago
- ☆41Updated 7 months ago
- A Postgres data warehouse for processing synthetic data using IAC principles☆16Updated last year
- Writes the CSV file to Postgres, read table and modify it. Write more tables to Postgres with Airflow.☆36Updated last year
- A real-time streaming ETL pipeline for streaming and performing sentiment analysis on Twitter data using Apache Kafka, Apache Spark and D…☆30Updated 4 years ago
- The goal of this project is to build a docker cluster that gives access to Hadoop, HDFS, Hive, PySpark, Sqoop, Airflow, Kafka, Flume, Pos…☆58Updated last year
- This project serves as a comprehensive guide to building an end-to-end data engineering pipeline using TCP/IP Socket, Apache Spark, OpenA…☆32Updated last year
- A data pipeline moving data from a Relational database system (RDBMS) to a Hadoop file system (HDFS).☆15Updated 3 years ago
- The goal of this project is to offer an AWS EMR template using Spot Fleet and On-Demand Instances that you can use quickly. Just focus on…☆26Updated 2 years ago
- PySpark Tutorial for Beginners on Google Colab: Hands-On Guide☆16Updated 4 years ago
- Predict the number of deaths due to covid19 in the next two weeks☆11Updated 2 years ago
- The Data Pipeline and Analytics Stack is a comprehensive solution designed for processing, storing, and visualizing data. Explore a compl…☆11Updated last year
- ☆11Updated 4 years ago
- ☆11Updated 2 years ago
- Example of event-driven architecture with FastAPI Gateway, Kafka, Redis pub/sub and Faust-streaming☆14Updated 3 years ago
- Create a streaming data, transfer it to Kafka, modify it with PySpark, take it to ElasticSearch and MinIO☆59Updated last year
- This project focuses on building a robust data pipeline using Apache Airflow to automate the ingestion of weather data from the OpenWeath…☆21Updated last year
- reating a modern data pipeline using a combination of Terraform, AWS Lambda and S3, Snowflake, DBT, Mage AI, and Dash.☆14Updated last year
- This project shows how to capture changes from postgres database and stream them into kafka☆35Updated 9 months ago