FlagOpen / FlagScale
FlagScale is a large model toolkit based on open-sourced projects.
☆167Updated this week
Related projects ⓘ
Alternatives and complementary repositories for FlagScale
- USP: Unified (a.k.a. Hybrid, 2D) Sequence Parallel Attention for Long Context Transformers Model Training and Inference☆352Updated last week
- A collection of memory efficient attention operators implemented in the Triton language.☆215Updated 5 months ago
- Best practice for training LLaMA models in Megatron-LM☆627Updated 10 months ago
- Disaggregated serving system for Large Language Models (LLMs).☆348Updated 2 months ago
- ☆198Updated this week
- ☆74Updated 10 months ago
- RTP-LLM: Alibaba's high-performance LLM inference engine for diverse applications.☆541Updated 3 weeks ago
- FlagGems is an operator library for large language models implemented in Triton Language.☆328Updated this week
- InternEvo is an open-sourced lightweight training framework aims to support model pre-training without the need for extensive dependencie…☆307Updated this week
- ☆282Updated last week
- ☆123Updated this week
- The official repo of Pai-Megatron-Patch for LLM & VLM large scale training developed by Alibaba Cloud.☆711Updated last week
- [USENIX ATC '24] Accelerating the Training of Large Language Models using Efficient Activation Rematerialization and Optimal Hybrid Paral…☆46Updated 3 months ago
- Analyze the inference of Large Language Models (LLMs). Analyze aspects like computation, storage, transmission, and hardware roofline mod…☆310Updated last month
- Zero Bubble Pipeline Parallelism☆279Updated this week
- ☆142Updated this week
- veRL: Volcano Engine Reinforcement Learning for LLM☆279Updated this week
- [NeurIPS'23] H2O: Heavy-Hitter Oracle for Efficient Generative Inference of Large Language Models.☆387Updated 3 months ago
- PyTorch bindings for CUTLASS grouped GEMM.☆67Updated 3 months ago
- Model Compression for Big Models☆151Updated last year
- ☆208Updated last year
- Ascend PyTorch adapter (torch_npu). Mirror of https://gitee.com/ascend/pytorch☆253Updated this week
- Easy Parallel Library (EPL) is a general and efficient deep learning framework for distributed model training.☆264Updated last year
- Dynamic Memory Management for Serving LLMs without PagedAttention☆222Updated this week
- [EMNLP 2024 Industry Track] This is the official PyTorch implementation of "LLMC: Benchmarking Large Language Model Quantization with a V…☆315Updated this week
- GLake: optimizing GPU memory management and IO transmission.☆375Updated 3 months ago
- ☆140Updated 6 months ago
- Inferflow is an efficient and highly configurable inference engine for large language models (LLMs).☆236Updated 7 months ago
- Ring attention implementation with flash attention☆578Updated this week
- optimized BERT transformer inference on NVIDIA GPU. https://arxiv.org/abs/2210.03052☆457Updated 7 months ago