dbusteed / kafka-spark-streaming-example
☆37Updated 5 years ago
Alternatives and similar repositories for kafka-spark-streaming-example:
Users that are interested in kafka-spark-streaming-example are comparing it to the libraries listed below
- Apche Spark Structured Streaming with Kafka using Python(PySpark)☆40Updated 5 years ago
- Design/Implement stream/batch architecture on NYC taxi data | #DE☆25Updated 4 years ago
- Classwork projects and home works done through Udacity data engineering nano degree☆74Updated last year
- PySpark Cheatsheet☆36Updated 2 years ago
- PySpark-ETL☆23Updated 5 years ago
- Developed an ETL pipeline for a Data Lake that extracts data from S3, processes the data using Spark, and loads the data back into S3 as …☆16Updated 5 years ago
- Simplified ETL process in Hadoop using Apache Spark. Has complete ETL pipeline for datalake. SparkSession extensions, DataFrame validatio…☆54Updated 2 years ago
- Apache Spark 3 - Structured Streaming Course Material☆122Updated last year
- ☆40Updated 10 months ago
- ☆150Updated 7 years ago
- A real-time streaming ETL pipeline for streaming and performing sentiment analysis on Twitter data using Apache Kafka, Apache Spark and D…☆30Updated 4 years ago
- ☆87Updated 2 years ago
- A workspace to experiment with Apache Spark, Livy, and Airflow in a Docker environment.☆38Updated 4 years ago
- PySpark Tutorial for Beginners on Google Colab: Hands-On Guide☆16Updated 4 years ago
- Simple ETL pipeline using Python☆26Updated last year
- Project for real-time anomaly detection using Kafka and python☆58Updated 2 years ago
- ☆33Updated 5 years ago
- PySpark functions and utilities with examples. Assists ETL process of data modeling☆102Updated 4 years ago
- Dockerizing an Apache Spark Standalone Cluster☆43Updated 2 years ago
- Code examples on Apache Spark using python☆107Updated 2 years ago
- My solutions for the Udacity Data Engineering Nanodegree☆34Updated 5 years ago
- Data pipeline performing ETL to AWS Redshift using Spark, orchestrated with Apache Airflow☆144Updated 4 years ago
- This project helps me to understand the core concepts of Apache Airflow. I have created custom operators to perform tasks such as staging…☆85Updated 5 years ago
- Developed a data pipeline to automate data warehouse ETL by building custom airflow operators that handle the extraction, transformation,…☆90Updated 3 years ago
- This set of code and instructions has the porpouse to instanciate a compiled environment with set of docker images like airflow webserver…☆3Updated last year
- Solution to all projects of Udacity's Data Engineering Nanodegree: Data Modeling with Postgres & Cassandra, Data Warehouse with Redshift,…☆56Updated 2 years ago
- ETL pipeline using pyspark (Spark - Python)☆114Updated 5 years ago
- Data Quest - Data Engineer Learning and Projects☆24Updated 5 years ago
- Create a streaming data, transfer it to Kafka, modify it with PySpark, take it to ElasticSearch and MinIO☆60Updated last year
- A production-grade data pipeline has been designed to automate the parsing of user search patterns to analyze user engagement. Extract d…☆24Updated 3 years ago