Model parallel transformers in JAX and Haiku
☆6,363Jan 21, 2023Updated 3 years ago
Alternatives and similar repositories for mesh-transformer-jax
Users that are interested in mesh-transformer-jax are comparing it to the libraries listed below
Sorting:
- An implementation of model parallel autoregressive transformers on GPUs, based on the Megatron and DeepSpeed libraries☆7,396Feb 3, 2026Updated last month
- An implementation of model parallel GPT-2 and GPT-3-style models using the mesh-tensorflow library.☆8,287Feb 25, 2022Updated 4 years ago
- Repo for external large-scale work☆6,543Apr 27, 2024Updated last year
- ☆2,950Jan 15, 2026Updated last month
- Swarm training framework using Haiku + JAX + Ray for layer parallel transformer language models on unreliable, heterogeneous nodes☆241May 12, 2023Updated 2 years ago
- OpenAssistant is a chat-based assistant that understands tasks, can interact with third-party systems, and retrieve information dynamical…☆37,444Aug 17, 2024Updated last year
- JAX-based neural network library☆3,193Mar 2, 2026Updated last week
- Facebook AI Research Sequence-to-Sequence Toolkit written in Python.☆32,176Sep 30, 2025Updated 5 months ago
- A minimal PyTorch re-implementation of the OpenAI GPT (Generative Pretrained Transformer) training☆23,845Aug 15, 2024Updated last year
- Composable transformations of Python+NumPy programs: differentiate, vectorize, JIT to GPU/TPU, and more☆34,987Updated this week
- 🌸 Run LLMs at home, BitTorrent-style. Fine-tuning and inference up to 10x faster than offloading☆9,982Sep 7, 2024Updated last year
- Code for the paper "Language Models are Unsupervised Multitask Learners"☆24,668Aug 14, 2024Updated last year
- Running large language models on a single GPU for throughput-oriented scenarios.☆9,381Oct 28, 2024Updated last year
- A repo for distributed training of language models with Reinforcement Learning via Human Feedback (RLHF)☆4,738Jan 8, 2024Updated 2 years ago
- RWKV (pronounced RwaKuv) is an RNN with great LLM performance, which can also be directly trained like a GPT transformer (parallelizable)…☆14,393Feb 21, 2026Updated 2 weeks ago
- 🤗 Transformers: the model-definition framework for state-of-the-art machine learning models in text, vision, audio, and multimodal model…☆157,462Updated this week
- DeepSpeed is a deep learning optimization library that makes distributed training and inference easy, efficient, and effective.☆41,759Updated this week
- CodeGen is a family of open-source model for program synthesis. Trained on TPU-v4. Competitive with OpenAI Codex.☆5,170Oct 27, 2025Updated 4 months ago
- Large-scale Self-supervised Pre-training Across Tasks, Languages, and Modalities☆22,030Jan 23, 2026Updated last month
- Flax is a neural network library for JAX that is designed for flexibility.