PacktPublishing / Superset-Quick-Start-Guide
Superset Quick Start Guide, published by Packt
☆56Updated last year
Alternatives and similar repositories for Superset-Quick-Start-Guide:
Users that are interested in Superset-Quick-Start-Guide are comparing it to the libraries listed below
- Airflow workflow management platform chef cookbook.☆71Updated 5 years ago
- A repository of sample code to show data quality checking best practices using Airflow.☆74Updated 2 years ago
- Data validation library for PySpark 3.0.0☆33Updated 2 years ago
- Learn how to add data validation and documentation to a data pipeline built with dbt and Airflow.☆167Updated last year
- ☆110Updated 2 months ago
- An Airflow docker image preconfigured to work well with Spark and Hadoop/EMR☆174Updated last year
- Big Data Demystified meetup and blog examples☆31Updated 7 months ago
- Data Brewery is an ETL (Extract-Transform-Load) program that connect to many data sources (cloud services, databases, ...) and manage dat…☆16Updated 4 years ago
- PyConDE & PyData Berlin 2019 Airflow Workshop: Airflow for machine learning pipelines.☆47Updated last year
- Trino dbt demo project to mix and load BigQuery data with and in a local PostgreSQL database☆72Updated 3 years ago
- ELT Code for your Data Warehouse☆26Updated last year
- ☆73Updated last week
- Execution of DBT models using Apache Airflow through Docker Compose☆116Updated 2 years ago
- Data processing and modelling framework for automating tasks (incl. Python & SQL transformations).☆123Updated 10 months ago
- Containerized end-to-end analytics of Spotify data using Python, dbt, Postgres, and Metabase☆125Updated 2 years ago
- dagster scikit-learn pipeline example.☆45Updated 2 years ago
- Airflow training for the crunch conf☆105Updated 6 years ago
- 🚨 Simple, self-contained fraud detection system built with Apache Kafka and Python☆86Updated 5 years ago
- locopy: Loading/Unloading to Redshift and Snowflake using Python.☆107Updated last week
- A project for exploring how Great Expectations can be used to ensure data quality and validate batches within a data pipeline defined in …☆21Updated 2 years ago
- Simplified ETL process in Hadoop using Apache Spark. Has complete ETL pipeline for datalake. SparkSession extensions, DataFrame validatio…☆53Updated last year
- PySpark phonetic and string matching algorithms☆39Updated last year
- 🐍💨 Airflow tutorial for PyCon 2019☆85Updated 2 years ago
- How to manage Slowly Changing Dimensions with Apache Hive☆55Updated 5 years ago
- Quickly ingest messy CSV and XLS files. Export to clean pandas, SQL, parquet☆195Updated last year
- Data lake, data warehouse on GCP☆56Updated 3 years ago
- Mapping of DWH database tables to business entities, attributes & metrics in Python, with automatic creation of flattened tables☆73Updated last year
- A project with examples of using few commonly used data manipulation/processing/transformation APIs in Apache Spark 2.0.0☆25Updated 3 years ago
- A plugin to Apache Airflow to allow you to run Spark Submit Commands as an Operator☆73Updated 5 years ago
- Skeleton project for Apache Airflow training participants to work on.☆16Updated 4 years ago