hammerlab / magic-rdds
Miscellaneous functionality for manipulating Apache Spark RDDs.
☆22Updated 6 years ago
Alternatives and similar repositories for magic-rdds:
Users that are interested in magic-rdds are comparing it to the libraries listed below
- Scriptable scheduler for periodical Hadoop workflows☆22Updated 7 years ago
- A collection of Apache Parquet add-on modules☆30Updated this week
- something to help you spark☆65Updated 6 years ago
- Use Cascading Taps and Scalding DSL with Spark☆49Updated 8 years ago
- Apache Spark jobs such as Principal Coordinate Analysis.☆74Updated 8 years ago
- Secondary sort and streaming reduce for Apache Spark☆78Updated last year
- Starter project for building MemSQL Streamliner Pipelines☆32Updated 7 years ago
- low-level helpers for Apache Spark libraries and tests☆16Updated 6 years ago
- Bucketing and partitioning system for Parquet☆30Updated 6 years ago
- Example project to show how to use Spark to read and write Avro/Parquet files☆50Updated 11 years ago
- Joins for skewed datasets in Spark☆57Updated 7 years ago
- Utilities for writing tests that use Apache Spark.☆24Updated 6 years ago
- Dependency and data pipeline management framework for Spark and Scala☆15Updated 7 years ago
- functionstest☆33Updated 8 years ago
- Library for organizing batch processing pipelines in Apache Spark☆41Updated 8 years ago
- Interactive Audience Analytics with Spark and HyperLogLog☆55Updated 9 years ago
- Cascading on Apache Flink®☆54Updated last year
- Scala client for the Lightning data visualization server (WIP)☆47Updated 5 years ago
- Experiments with the GDELT dataset and Cassandra schemas.☆25Updated 9 years ago
- Data-Driven Spark allows quick data exploration based on Apache Spark.☆28Updated 8 years ago
- General utility code used across BDG products. Apache 2 licensed.☆18Updated last month
- A quotation-based Scala DSL for scalable data analysis.☆63Updated 2 years ago
- Efficient, distributed downloads of large files from S3 to HDFS using Spark.☆17Updated 7 years ago
- Example Spark project using Parquet as a columnar store with Thrift objects.☆48Updated 10 years ago
- Spooker is a dynamic framework for processing high volume data streams via processing pipelines☆29Updated 9 years ago
- Embedded Kafka for testing and quick prototyping.☆14Updated 8 years ago
- ScalaCheck for Spark☆63Updated 7 years ago
- A framework for creating composable and pluggable data processing pipelines using Apache Spark, and running them on a cluster.☆47Updated 8 years ago
- Fast JVM collection☆59Updated 10 years ago
- This is an introduction of Apache Spark DataFrames.☆41Updated 10 years ago