colbyford / sparkitecture
A collection of “cookbook-style” scripts for simplifying data engineering and machine learning in Apache Spark.
☆13Updated 3 years ago
Alternatives and similar repositories for sparkitecture:
Users that are interested in sparkitecture are comparing it to the libraries listed below
- Apache-Spark based Data Flow(ETL) Framework which supports multiple read, write destinations of different types and also support multiple…☆26Updated 3 years ago
- Multi-stage, config driven, SQL based ETL framework using PySpark☆25Updated 5 years ago
- A Spark-based data comparison tool at scale which facilitates software development engineers to compare a plethora of pair combinations o…☆50Updated last year
- Spark functions to run popular phonetic and string matching algorithms☆60Updated 3 years ago
- Spark and Delta Lake Workshop☆22Updated 2 years ago
- Examples for High Performance Spark☆15Updated 5 months ago
- Nested Data (JSON/AVRO/XML) Parsing and Flattening in Spark☆16Updated last year
- Read Delta tables without any Spark☆47Updated last year
- Data validation library for PySpark 3.0.0☆33Updated 2 years ago
- A library that brings useful functions from various modern database management systems to Apache Spark☆58Updated last year
- Notebooks for the ML Link Prediction Course☆14Updated 4 years ago
- Fuzzy matching function in spark (https://spark-packages.org/package/itspawanbhardwaj/spark-fuzzy-matching)☆24Updated 5 years ago
- A project with examples of using few commonly used data manipulation/processing/transformation APIs in Apache Spark 2.0.0☆25Updated 3 years ago
- Lighthouse is a library for data lakes built on top of Apache Spark. It provides high-level APIs in Scala to streamline data pipelines an…☆61Updated 7 months ago
- Filling in the Spark function gaps across APIs☆50Updated 4 years ago
- PySpark phonetic and string matching algorithms☆39Updated last year
- Magic to help Spark pipelines upgrade☆34Updated 6 months ago
- ☆10Updated 2 years ago
- Waimak is an open-source framework that makes it easier to create complex data flows in Apache Spark.☆75Updated last year
- Mastering Spark for Data Science, published by Packt☆47Updated 2 years ago
- Asynchronous actions for PySpark☆47Updated 3 years ago
- Scalable CDC Pattern Implemented using PySpark☆18Updated 5 years ago
- Examples of Spark 3.0☆47Updated 4 years ago
- A systematic Benchmarking on the performance of Spark-SQL for processing Vast RDF datasets☆14Updated 2 years ago
- Data quality control tool built on spark and deequ☆24Updated last month
- Code snippets used in demos recorded for the blog.☆33Updated this week
- notebooks for nlp-on-spark☆13Updated 8 years ago
- Machine Learning Pipeline Stages for Spark (exposed in Scala/Java + Python)☆74Updated last year
- PySpark Algorithms Book: https://www.amazon.com/dp/B07X4B2218/ref=sr_1_2☆84Updated 5 years ago
- A Spark datasource for the HadoopOffice library☆38Updated 2 years ago