Modalities / modalitiesLinks
Modalities, a PyTorch-native framework for distributed and reproducible foundation model training.
☆91Updated this week
Alternatives and similar repositories for modalities
Users that are interested in modalities are comparing it to the libraries listed below
Sorting:
- nanoGPT-like codebase for LLM training☆110Updated 2 weeks ago
- SDLG is an efficient method to accurately estimate aleatoric semantic uncertainty in LLMs☆27Updated last year
- some common Huggingface transformers in maximal update parametrization (µP)☆86Updated 3 years ago
- Flexible library for merging large language models (LLMs) via evolutionary optimization (ACL 2025 Demo).☆91Updated 3 months ago
- Efficient LLM inference on Slurm clusters using vLLM.☆82Updated last week
- ☆61Updated last year
- Official Repository of Pretraining Without Attention (BiGS), BiGS is the first model to achieve BERT-level transfer learning on the GLUE …☆115Updated last year
- ☆166Updated 2 years ago
- Notebooks accompanying Anthropic's "Toy Models of Superposition" paper☆130Updated 3 years ago
- ☆82Updated last year
- Code for NeurIPS 2024 Spotlight: "Scaling Laws and Compute-Optimal Training Beyond Fixed Training Durations"☆85Updated last year
- Supercharge huggingface transformers with model parallelism.☆77Updated 4 months ago
- A MAD laboratory to improve AI architecture designs 🧪☆133Updated 11 months ago
- One Initialization to Rule them All: Fine-tuning via Explained Variance Adaptation☆45Updated last month
- Understand and test language model architectures on synthetic tasks.☆240Updated last month
- The simplest, fastest repository for training/finetuning medium-sized GPTs.☆173Updated 4 months ago
- Running Jax in PyTorch Lightning☆114Updated 11 months ago
- PyTorch library for Active Fine-Tuning☆95Updated last month
- Implementation of the BatchTopK activation function for training sparse autoencoders (SAEs)☆55Updated 4 months ago
- A Jax-based library for building transformers, includes implementations of GPT, Gemma, LlaMa, Mixtral, Whisper, SWin, ViT and more.☆297Updated last year
- Yet another random morning idea to be quickly tried and architecture shared if it works; to allow the transformer to pause for any amount…☆53Updated 2 years ago
- ☆83Updated 8 months ago
- Official implementation of "BERTs are Generative In-Context Learners"☆32Updated 8 months ago
- [NeurIPS 2023] Learning Transformer Programs☆162Updated last year
- Unofficial but Efficient Implementation of "Mamba: Linear-Time Sequence Modeling with Selective State Spaces" in JAX☆89Updated last year
- A fast implementation of T5/UL2 in PyTorch using Flash Attention☆110Updated 3 weeks ago
- Official implementation of "GPT or BERT: why not both?"☆62Updated 3 months ago
- Interpretating the latent space representations of attention head outputs for LLMs☆34Updated last year
- Minimum Description Length probing for neural network representations☆20Updated 9 months ago
- Scrape papers from OpenReview using OpenReview API☆54Updated 8 months ago