memsql / singlestore-spark-connectorLinks
A connector for SingleStore and Spark
☆162Updated this week
Alternatives and similar repositories for singlestore-spark-connector
Users that are interested in singlestore-spark-connector are comparing it to the libraries listed below
Sorting:
- Live-updating Spark UI built with Meteor☆189Updated 4 years ago
- ☆92Updated 8 years ago
- kafka-connect-s3 : Ingest data from Kafka to Object Stores(s3)☆94Updated 6 years ago
- Support Highcharts in Apache Zeppelin☆81Updated 7 years ago
- Spark SQL index for Parquet tables☆134Updated 4 years ago
- Low level integration of Spark and Kafka☆130Updated 7 years ago
- Bulletproof Apache Spark jobs with fast root cause analysis of failures.☆72Updated 4 years ago
- Enabling Spark Optimization through Cross-stack Monitoring and Visualization☆47Updated 7 years ago
- Google Dataflow Runner for Apache Flink™ (deprecated; please use the up-to-date Beam Runner)☆88Updated 8 years ago
- Starter project for building MemSQL Streamliner Pipelines☆32Updated 8 years ago
- Sparkline BI Accelerator provides fast ad-hoc query capability over Logical Cubes. This has been folded into our SNAP Platform(http://bit…☆282Updated 6 years ago
- Interactive Audience Analytics with Spark and HyperLogLog☆55Updated 9 years ago
- Quark is a data virtualization engine over analytic databases.☆98Updated 7 years ago
- A library to expose more of Apache Spark's metrics system☆146Updated 5 years ago
- Schedoscope is a scheduling framework for painfree agile development, testing, (re)loading, and monitoring of your datahub, lake, or what…☆96Updated 5 years ago
- Read SparkSQL parquet file as RDD[Protobuf]☆93Updated 6 years ago
- An efficient updatable key-value store for Apache Spark☆251Updated 8 years ago
- Scripts for generating Grafana dashboards for monitoring Spark jobs☆242Updated 10 years ago
- ☆111Updated 8 years ago
- Druid indexing plugin for using Spark in batch jobs☆101Updated 3 years ago
- functionstest☆33Updated 8 years ago
- Hadoop MapReduce tool to convert Avro data files to Parquet format.☆34Updated 12 years ago
- Google BigQuery support for Spark, SQL, and DataFrames☆155Updated 5 years ago
- A High Performance Cluster Consumer for Kafka that creates Avro (boom) files in Hadoop in time based directory paths☆42Updated 9 years ago
- Hadoop output committers for S3