duhanmin / structured-streaming-Kafka2HBase
Spark structured-streaming 消费kafka数据写入hbase
☆33Updated 5 years ago
Alternatives and similar repositories for structured-streaming-Kafka2HBase:
Users that are interested in structured-streaming-Kafka2HBase are comparing it to the libraries listed below
- SparkStreaming中利用MySQL保存Kafka偏移量保证0数据丢失☆45Updated 7 years ago
- Kafka delivery semantics in the case of failure depend on how and when offsets are stored. Spark output operations are at-least-once. So …☆37Updated 7 years ago
- sql实现Structured Streaming☆39Updated 6 years ago
- Spark Streaming监控平台,支持任务部署与告警、自启动☆128Updated 6 years ago
- flink技术学习笔记分享☆84Updated 5 years ago
- 使用spark对hive、hbase、ES的读写, 实现一次配置可对不同数据库进行导入导出,并对ES、hbase进行封装☆32Updated 7 years ago
- spark将hdfs数据高性能灌入kafka,然后spark streaming/structured streaming高速消费,关注性能,欢迎提供性能/代码优化建议☆33Updated 5 years ago
- Encapsulated spark 与其他组件的结合api,方便使用,例如 es,hbase,kudu,kafka,mq等☆35Updated 5 years ago
- flink实时处理kafka传来的数据通过连接池技术写入hbase☆94Updated 2 years ago
- Streaming 相关项目☆15Updated 7 years ago
- Learning Flink : Flink CEP,Flink Core,Flink SQL☆71Updated 2 years ago
- spark-scala-maven☆58Updated 6 years ago
- 手动管理spark streaming集成kafka的数据偏移量到zookeeper中☆22Updated 6 years ago
- tools for bigData☆37Updated 6 years ago
- Flink 案例代码☆43Updated 2 years ago
- My Blog☆76Updated 6 years ago
- 一个手动管理spark streaming集成kafka时的偏移量到zookeeper中的小项目☆135Updated 5 months ago
- 一个为spark批量导入数据到hbase的库☆43Updated 8 years ago
- Flink Sql 教程☆34Updated last month
- ☆53Updated 6 years ago
- Spark 脚手架工程,标准化 spark 开发、部署、测试流程。☆93Updated 3 months ago
- 基于flink 1.8 源码二次开发,详见MD☆82Updated 4 years ago
- A playground for Spark jobs.☆44Updated 6 years ago
- SparkStreaming项目,显示flume->Kafka->Spark->hbase(实时数据处理方案),Scala实现☆36Updated 6 years ago
- Use Scala API to read/write data from different databases,HBase,MySQL,etc.☆24Updated 6 years ago
- DirectKafka examples for Spark Streaming : 1. with checkpointing 2. Custom offset management☆60Updated 8 years ago
- ☆33Updated 5 years ago
- ☆29Updated 6 years ago
- This project compose of two parts: 1) write, spark job to write to hbase using bulk load to; 2)read, rest api reading from hbase base on …☆20Updated 7 years ago