qsyao / cudaBERT
A Fast Muti-processing BERT-Inference System
☆101Updated 2 years ago
Alternatives and similar repositories for cudaBERT:
Users that are interested in cudaBERT are comparing it to the libraries listed below
- OneFlow models for benchmarking.☆105Updated 7 months ago
- InsNet Runs Instance-dependent Neural Networks with Padding-free Dynamic Batching.☆66Updated 3 years ago
- pytorch源码阅读 0.2.0 版本☆90Updated 5 years ago
- Simple Dynamic Batching Inference☆145Updated 3 years ago
- Place for meetup slides☆140Updated 4 years ago
- oneflow documentation☆68Updated 8 months ago
- A small deep-learning framework with C++/Python/CUDA☆53Updated 6 years ago
- DeepLearning Framework Performance Profiling Toolkit☆285Updated 2 years ago
- Running BERT without Padding☆472Updated 3 years ago
- Fast implementation of BERT inference directly on NVIDIA (CUDA, CUBLAS) and Intel MKL☆543Updated 4 years ago
- Transformer related optimization, including BERT, GPT☆59Updated last year
- ☆214Updated last year
- ☆125Updated 3 years ago
- A high-performance distributed deep learning system targeting large-scale and automated distributed training. If you have any interests, …☆108Updated last year
- tensorflow源码阅读笔记☆190Updated 6 years ago
- an automatic differentiation framework with dynamic graph/支持动态图的自动求导框架☆101Updated 5 years ago
- A way to use cuda to accelerate top k algorithm☆29Updated 7 years ago
- ☆51Updated last year
- Easy Parallel Library (EPL) is a general and efficient deep learning framework for distributed model training.☆267Updated last year
- A high-performance framework for training wide-and-deep recommender systems on heterogeneous cluster☆157Updated 11 months ago
- ☆139Updated 11 months ago
- Tutorial code on how to build your own Deep Learning System in 2k Lines☆125Updated 7 years ago
- Dynamic Tensor Rematerialization prototype (modified PyTorch) and simulator. Paper: https://arxiv.org/abs/2006.09616☆133Updated last year
- TePDist (TEnsor Program DISTributed) is an HLO-level automatic distributed system for DL models.☆92Updated last year
- Performance of the C++ interface of flash attention and flash attention v2 in large language model (LLM) inference scenarios.☆35Updated 3 weeks ago
- ☆79Updated 3 months ago
- Inference framework for MoE layers based on TensorRT with Python binding☆41Updated 3 years ago
- ☆127Updated 2 months ago
- PyTorch distributed training acceleration framework☆44Updated last month
- TensorFlow code and pre-trained models for BERT☆24Updated 5 years ago