nonokang / bd-utils
大数据工具封装
☆19Updated 2 years ago
Alternatives and similar repositories for bd-utils:
Users that are interested in bd-utils are comparing it to the libraries listed below
- Flink Sql 教程☆34Updated 3 months ago
- 基于canal/kafka conenct的mysql/oracle数据实时同步、flink rest api、flink sql以及udf☆50Updated 2 years ago
- 基于DataX的通用数据同步微服务,一个Restful接口搞定所有通用数据同步☆53Updated 2 years ago
- 大数据组件学习;包括dataflow,spring cloud stream;elasticsearch;flink;spark;kafka;phoenix;Hive;Hbase;☆23Updated 2 years ago
- Encapsulated spark 与其他组件的结合api,方便使用,例如 es,hbase,kudu,kafka,mq等☆35Updated 5 years ago
- Learning Flink : Flink CEP,Flink Core,Flink SQL☆71Updated 3 years ago
- ☆38Updated last year
- 此项目主要应用于数据中台或数据平台的数据总线,支持直接实时监听MySQL、MongoDB、PostgreSQL、Oracle、SQL Server、Db2和Cassandra等数据库的数据变更。☆62Updated last year
- A distributed data factory, providing data access, etl, scheduling. Easily manage tasks such as hive, spark, clickhouse, flink, shell, py…☆32Updated 2 years ago
- Data self exporting and monitoring platform based on Hive data warehouse. https://hc.smartloli.org☆36Updated 7 years ago
- sql实现Structured Streaming☆39Updated 6 years ago
- poseidonX 是一个基于jstorm和flink的一体化实时计算服务平台☆55Updated 6 years ago
- java性能采集工具☆51Updated 6 years ago
- flink-sql 在 flink 上运行 sql 和 构建数据流的平台 基于 apache flink 1.10.0☆110Updated 2 years ago
- 易观开源大数据互联网百亿级记录互传Backquarter项目☆19Updated 2 years ago
- Flink 案例代码☆43Updated 2 years ago
- 基于flink 1.8 源码二次开发,详见MD☆82Updated 4 years ago
- Hive-JDBC-Proxy是一个高性能的HiveServer2和Spark ThriftServer的代理服务,具备负载均衡、基于规则转发Hive JDBC Client的请求给到HiveServer2和Spark ThriftServer的能力。☆32Updated 2 years ago
- SparkStreaming中利用MySQL保存Kafka偏移量保证0数据丢失☆45Updated 7 years ago
- 基于插件架构的数据源服务,统一接口,可操作不同类型数据源☆46Updated 2 years ago
- Spark Streaming监控平台,支持任务部署与告警、自启动☆128Updated 6 years ago
- log、event 、time 、window 、table、sql、connect、join、async IO、维表、CEP☆68Updated 2 years ago
- 因现有的datax、sqoop满足不了需求,使用spark封装了一个数据同步工具。☆9Updated 5 years ago
- 通过HBase Observer同步数据到ElasticSearch☆55Updated 9 years ago
- Real-time ETL developed by Flink, data from MySQL to Greenplum. Use canal to parse the MySQL binlog, put it into kafka, use Flink to cons…☆78Updated 11 months ago
- spark将hdfs数据高性能灌入kafka,然后spark streaming/structured streaming高速消费,关注性能,欢迎提供性能/代码优化建议☆33Updated 5 years ago
- Guardian of Waterdrop and Spark☆30Updated 2 years ago
- hive sql 字段血缘☆22Updated 2 years ago
- 一个为spark批量导入数据到hbase的库☆43Updated 8 years ago
- 执行Flink SQL 文件的客户端☆26Updated 3 years ago