crealytics / spark-google-analytics
☆32Updated this week
Related projects: ⓘ
- SQL for many helpful Redshift UDFs, and the scripts for generating and testing those UDFs☆125Updated 5 years ago
- Google BigQuery support for Spark, SQL, and DataFrames☆155Updated 4 years ago
- A decisioning and response platform☆69Updated 2 years ago
- Mirrors a Kinesis stream to Amazon S3 using the KCL☆42Updated 2 weeks ago
- Apache Spark AWS Lambda Executor (SAMBA)☆44Updated 6 years ago
- DataPipeline for humans.☆252Updated 2 years ago
- JSONs -> JSON Schema☆151Updated 4 years ago
- Redshift Ops Console☆93Updated 8 years ago
- The Schema Repo is a RESTful web service for storing and serving mappings between schema identifiers and schema definitions.☆155Updated 2 years ago
- kafka-connect-s3 : Ingest data from Kafka to Object Stores(s3)☆97Updated 5 years ago
- A Spark Streaming job reading events from Amazon Kinesis and writing event counts to DynamoDB☆94Updated 3 years ago
- Simplify getting Zeppelin up and running☆56Updated 8 years ago
- Google Analytics plugin for sending events to Snowplow☆17Updated 3 years ago
- Tools for working with parquet, impala, and hive☆134Updated 3 years ago
- Iglu is a machine-readable, open-source schema repository for JSON Schema from the team at Snowplow☆206Updated last month
- Amazon Kinesis Aggregators provides a simple way to create real time aggregations of data on Amazon Kinesis.☆151Updated 3 years ago
- DonorsChoose.org Data Science Team Opensource Code☆77Updated last year
- Scala SDK for working with Snowplow enriched events in Spark, AWS Lambda, Flink et al.☆20Updated 8 months ago
- Google BigQuery support for Spark, Structured Streaming, SQL, and DataFrames with easy Databricks integration.☆70Updated last year
- Empower Curiosity / Redshift analytics platform☆77Updated 3 years ago
- Ephemeral Hadoop clusters using Google Compute Platform☆134Updated 2 years ago
- ☆56Updated this week
- A Spark WordCountJob example as a standalone SBT project with Specs2 tests, runnable on Amazon EMR☆118Updated 8 years ago
- Export Redshift data and convert to Parquet for use with Redshift Spectrum or other data warehouses.☆116Updated last year
- ☆56Updated this week
- ☆30Updated this week
- Functional, Typesafe, Declarative Data Pipelines☆139Updated 6 years ago
- A super simple utility for testing Apache Hive scripts locally for non-Java developers.☆72Updated 7 years ago
- An example of using Avro and Parquet in Spark SQL☆60Updated 8 years ago
- Amazon Elastic MapReduce code samples☆63Updated 9 years ago