Oneflow-Inc / libai
LiBai(李白): A Toolbox for Large-Scale Distributed Parallel Training
☆390Updated this week
Related projects ⓘ
Alternatives and complementary repositories for libai
- Models and examples built with OneFlow☆96Updated last month
- ☆209Updated last year
- Best practice for training LLaMA models in Megatron-LM☆628Updated 10 months ago
- Efficient Training (including pre-training and fine-tuning) for Big Models☆564Updated 3 months ago
- optimized BERT transformer inference on NVIDIA GPU. https://arxiv.org/abs/2210.03052☆457Updated 8 months ago
- Easy Parallel Library (EPL) is a general and efficient deep learning framework for distributed model training.☆263Updated last year
- FlagScale is a large model toolkit based on open-sourced projects.☆169Updated this week
- USP: Unified (a.k.a. Hybrid, 2D) Sequence Parallel Attention for Long Context Transformers Model Training and Inference☆357Updated this week
- Efficient Inference for Big Models☆571Updated last year
- OneFlow models for benchmarking.☆104Updated 3 months ago
- Running BERT without Padding☆460Updated 2 years ago
- ☆589Updated 5 months ago
- ☆123Updated 2 weeks ago
- A flexible and efficient training framework for large-scale alignment tasks☆206Updated this week
- ☆74Updated 11 months ago
- Tutel MoE: An Optimized Mixture-of-Experts Implementation☆735Updated this week
- ☆289Updated this week
- DeepLearning Framework Performance Profiling Toolkit☆277Updated 2 years ago
- RTP-LLM: Alibaba's high-performance LLM inference engine for diverse applications.☆545Updated last month
- Collaborative Training of Large Language Models in an Efficient Way☆411Updated 2 months ago
- ☆144Updated this week
- A collection of memory efficient attention operators implemented in the Triton language.☆219Updated 5 months ago
- Model Compression for Big Models☆151Updated last year
- oneflow documentation☆68Updated 4 months ago
- The road to hack SysML and become an system expert☆437Updated last month
- veRL: Volcano Engine Reinforcement Learning for LLM☆318Updated this week
- The official repo of Pai-Megatron-Patch for LLM & VLM large scale training developed by Alibaba Cloud.☆721Updated this week
- InternEvo is an open-sourced lightweight training framework aims to support model pre-training without the need for extensive dependencie…☆310Updated this week
- PyTorch bindings for CUTLASS grouped GEMM.☆68Updated 4 months ago
- [NeurIPS'23] H2O: Heavy-Hitter Oracle for Efficient Generative Inference of Large Language Models.☆391Updated 3 months ago