YaFSDP: Yet another Fully Sharded Data Parallel
☆985Jan 30, 2026Updated last month
Alternatives and similar repositories for YaFSDP
Users that are interested in YaFSDP are comparing it to the libraries listed below
Sorting:
- Library for industrial alignment.☆405Sep 24, 2025Updated 5 months ago
- Effective LLM Alignment Toolkit☆152Jun 25, 2025Updated 8 months ago
- My Implementation of Q-Sparse: All Large Language Models can be Fully Sparsely-Activated☆33Aug 14, 2024Updated last year
- Official implementation of the paper "Linear Transformers with Learnable Kernel Functions are Better In-Context Models"☆169Jan 16, 2025Updated last year
- A PyTorch native platform for training generative AI models☆5,098Updated this week
- Minimalistic large language model 3D-parallelism training☆2,579Feb 19, 2026Updated last week
- ☆93Jul 5, 2024Updated last year
- Official Pytorch repository for Extreme Compression of Large Language Models via Additive Quantization https://arxiv.org/pdf/2401.06118.p…☆1,315Aug 8, 2025Updated 6 months ago
- PyTorch native quantization and sparsity for training and inference☆2,707Updated this week
- PyTorch native post-training library☆5,691Updated this week
- Efficient Triton Kernels for LLM Training☆6,162Updated this week
- Tile primitives for speedy kernels☆3,183Updated this week
- Open weights language model from Google DeepMind, based on Griffin.☆663Feb 6, 2026Updated 3 weeks ago
- Layer-Condensed KV cache w/ 10 times larger batch size, fewer params and less computation. Dramatic speed up with better task performance…☆156Apr 7, 2025Updated 10 months ago
- MERA (Multimodal Evaluation for Russian-language Architectures) is a new open benchmark for the Russian language for evaluating fundament…☆62Oct 7, 2024Updated last year
- Q-GaLore: Quantized GaLore with INT4 Projection and Layer-Adaptive Low-Rank Gradients.☆203Jul 17, 2024Updated last year
- ☆415Oct 23, 2024Updated last year
- Tools for merging pretrained large language models.☆6,814Jan 26, 2026Updated last month
- [ICLR 2025] Samba: Simple Hybrid State Space Models for Efficient Unlimited Context Language Modeling☆952Nov 16, 2025Updated 3 months ago
- PyTorch compiler that accelerates training and inference. Get built-in optimizations for performance, memory, parallelism, and easily wri…☆1,444Updated this week
- SAGE: Spelling correction, corruption and evaluation for multiple languages☆165Dec 8, 2025Updated 2 months ago
- Minimalistic 4D-parallelism distributed training framework for education purpose☆2,090Aug 26, 2025Updated 6 months ago
- A library built upon PyTorch for building embeddings on discrete event sequences using self-supervision☆254Dec 23, 2025Updated 2 months ago
- ☆71Aug 27, 2024Updated last year
- Freeing data processing from scripting madness by providing a set of platform-agnostic customizable pipeline processing blocks.☆2,903Feb 24, 2026Updated last week
- ☆22Oct 4, 2023Updated 2 years ago
- 🚀 Efficiently (pre)training foundation models with native PyTorch features, including FSDP for training and SDPA implementation of Flash…☆280Nov 24, 2025Updated 3 months ago
- YTsaurus is a scalable and fault-tolerant open-source big data platform.☆2,131Updated this week
- Efficient Deep Learning Systems course materials (HSE, YSDA)☆962Updated this week
- Reference implementation of "Softmax Attention with Constant Cost per Token" (Heinsen, 2024)☆24Jun 6, 2024Updated last year
- Helpful tools and examples for working with flex-attention☆1,136Feb 8, 2026Updated 3 weeks ago
- Odysseus: Playground of LLM Sequence Parallelism☆79Jun 17, 2024Updated last year
- S-LoRA: Serving Thousands of Concurrent LoRA Adapters☆1,899Jan 21, 2024Updated 2 years ago
- This library supports evaluating disparities in generated image quality, diversity, and consistency between geographic regions.☆20Jun 3, 2024Updated last year
- A throughput-oriented high-performance serving framework for LLMs☆946Oct 29, 2025Updated 4 months ago
- MII makes low-latency and high-throughput inference possible, powered by DeepSpeed.☆2,095Jun 30, 2025Updated 8 months ago
- FlashInfer: Kernel Library for LLM Serving☆5,057Updated this week
- Official repository of paper "RNNs Are Not Transformers (Yet): The Key Bottleneck on In-context Retrieval"☆27Apr 17, 2024Updated last year
- Transformer related optimization, including BERT, GPT☆6,394Mar 27, 2024Updated last year