Modalities / modalities
Modalities, a PyTorch-native framework for distributed and reproducible foundation model training.
☆77Updated last week
Alternatives and similar repositories for modalities
Users that are interested in modalities are comparing it to the libraries listed below
Sorting:
- Efficient LLM inference on Slurm clusters using vLLM.☆62Updated this week
- nanoGPT-like codebase for LLM training☆94Updated last month
- Code for NeurIPS 2024 Spotlight: "Scaling Laws and Compute-Optimal Training Beyond Fixed Training Durations"☆72Updated 6 months ago
- Repository for the code of the "PPL-MCTS: Constrained Textual Generation Through Discriminator-Guided Decoding" paper, NAACL'22☆65Updated 2 years ago
- Understand and test language model architectures on synthetic tasks.☆195Updated 2 months ago
- SDLG is an efficient method to accurately estimate aleatoric semantic uncertainty in LLMs☆25Updated 11 months ago
- Language models scale reliably with over-training and on downstream tasks☆97Updated last year
- ☆73Updated 2 months ago
- ☆37Updated last year
- Code Release for "Broken Neural Scaling Laws" (BNSL) paper☆58Updated last year
- Simple-to-use scoring function for arbitrarily tokenized texts.☆39Updated 2 months ago
- ☆72Updated last year
- some common Huggingface transformers in maximal update parametrization (µP)☆80Updated 3 years ago
- A framework for few-shot evaluation of autoregressive language models.☆103Updated 2 years ago
- unofficial re-implementation of "Grokking: Generalization Beyond Overfitting on Small Algorithmic Datasets"☆78Updated 2 years ago
- LLM finetuning in resource-constrained environments.☆47Updated 10 months ago
- Yet another random morning idea to be quickly tried and architecture shared if it works; to allow the transformer to pause for any amount…☆53Updated last year
- Notebooks accompanying Anthropic's "Toy Models of Superposition" paper☆121Updated 2 years ago
- ☆31Updated last month
- Supercharge huggingface transformers with model parallelism.☆76Updated 7 months ago
- Efficient Transformers with Dynamic Token Pooling☆61Updated last year
- ☆38Updated last week
- A MAD laboratory to improve AI architecture designs 🧪☆115Updated 4 months ago
- A library for efficient patching and automatic circuit discovery.☆64Updated 3 weeks ago
- Discovering Data-driven Hypotheses in the Wild☆80Updated 5 months ago
- ☆54Updated last year
- NeuroSurgeon is a package that enables researchers to uncover and manipulate subnetworks within models in Huggingface Transformers☆41Updated 3 months ago
- Code and Configs for Asynchronous RLHF: Faster and More Efficient RL for Language Models☆51Updated 3 weeks ago
- ☆64Updated 2 years ago
- Scaling is a distributed training library and installable dependency designed to scale up neural networks, with a dedicated module for tr…☆60Updated 6 months ago