bernhard-42 / pyspark-atlasLinks
PySpark for ETL jobs including lineage to Apache Atlas in one script via code inspection
☆18Updated 8 years ago
Alternatives and similar repositories for pyspark-atlas
Users that are interested in pyspark-atlas are comparing it to the libraries listed below
Sorting:
- Circus Train is a dataset replication tool that copies Hive tables between clusters and clouds.☆88Updated last year
- A Spark-based data comparison tool at scale which facilitates software development engineers to compare a plethora of pair combinations o…☆51Updated last year
- The sane way of building a data layer in Airflow☆24Updated 5 years ago
- A library that brings useful functions from various modern database management systems to Apache Spark☆59Updated last year
- Shunting Yard is a real-time data replication tool that copies data between Hive Metastores.☆20Updated 3 years ago
- Schema Registry integration for Apache Spark☆40Updated 2 years ago
- Bulletproof Apache Spark jobs with fast root cause analysis of failures.☆72Updated 4 years ago
- Scalable CDC Pattern Implemented using PySpark☆18Updated 5 years ago
- Apache-Spark based Data Flow(ETL) Framework which supports multiple read, write destinations of different types and also support multiple…☆26Updated 3 years ago
- A small project to show how to add lineage to Atlas when using Spark as ETL tool☆12Updated 8 years ago
- A dynamic data completeness and accuracy library at enterprise scale for Apache Spark☆30Updated 6 months ago
- Examples for High Performance Spark☆15Updated 7 months ago
- A testing framework for Trino☆26Updated 2 months ago
- Dremio Flight connector. Access Dremio using Arrow flight☆40Updated 4 years ago
- ☆14Updated 3 months ago
- Avro Schema Shredder is a REST API that enables storage of Avro Schemas in Apache Atlas. This API enables an organization to use Apache A…☆13Updated 8 years ago
- Profiles the data, validates the schema and runs data quality checks and produces a report☆20Updated 6 years ago
- An opinionated auto-deployer for the Hortonworks Platform☆34Updated 4 years ago
- Ambari Service definition for deploying R & RHadoop libraries☆18Updated 9 years ago
- minio as local storage and DynamoDB as catalog☆15Updated last year
- Spark Structured Streaming State Tools☆34Updated 4 years ago
- ☆10Updated 3 years ago
- Using the Parquet file format (with Avro) to process data with Apache Flink☆14Updated 9 years ago
- hive_compared_bq compares/validates 2 (SQL like) tables, and graphically shows the rows/columns that are different.☆28Updated 7 years ago
- Yet Another (Spark) ETL Framework☆21Updated last year
- Hadoop Data Pipeline using Falcon☆15Updated 9 years ago
- Waimak is an open-source framework that makes it easier to create complex data flows in Apache Spark.☆75Updated last year
- Paper: A Zero-rename committer for object stores☆20Updated 4 years ago
- Apiary provides modules which can be combined to create a federated cloud data lake☆36Updated last year
- ☆63Updated 5 years ago