stanford-crfm / mistral
Mistral: A strong, northwesterly wind: Framework for transparent and accessible large-scale language model training, built with Hugging Face 🤗 Transformers.
☆562Updated last year
Related projects ⓘ
Alternatives and complementary repositories for mistral
- Task-based datasets, preprocessing, and evaluation for sequence models.☆558Updated this week
- Reproduce results and replicate training fo T0 (Multitask Prompted Training Enables Zero-Shot Task Generalization)☆457Updated 2 years ago
- An open collection of implementation tips, tricks and resources for training large language models☆459Updated last year
- Scaling Data-Constrained Language Models☆321Updated last month
- Code for T-Few from "Few-Shot Parameter-Efficient Fine-Tuning is Better and Cheaper than In-Context Learning"☆430Updated last year
- Central place for the engineering/scaling WG: documentation, SLURM scripts and logs, compute environment and data.☆980Updated 3 months ago
- NeurIPS Large Language Model Efficiency Challenge: 1 LLM + 1GPU + 1Day☆251Updated last year
- Seminar on Large Language Models (COMP790-101 at UNC Chapel Hill, Fall 2022)☆308Updated last year
- Interpretability for sequence generation models 🐛 🔍☆374Updated this week
- Repository containing code for "How to Train BERT with an Academic Budget" paper☆309Updated last year
- A framework for few-shot evaluation of autoregressive language models.☆101Updated last year
- Pipeline for pulling and processing online language model pretraining data from the web☆174Updated last year
- An open collection of methodologies to help with successful training of large language models.☆459Updated 8 months ago
- Used for adaptive human in the loop evaluation of language and embedding models.☆303Updated last year
- Build, evaluate, understand, and fix LLM-based apps☆484Updated 9 months ago
- Manage scalable open LLM inference endpoints in Slurm clusters☆237Updated 3 months ago
- Code repository for supporting the paper "Atlas Few-shot Learning with Retrieval Augmented Language Models",(https//arxiv.org/abs/2208.03…☆514Updated 11 months ago
- Expanding natural instructions☆956Updated 10 months ago
- Organize your experiments into discrete steps that can be cached and reused throughout the lifetime of your research project.☆532Updated 5 months ago
- distributed trainer for LLMs☆542Updated 5 months ago
- A repository for research on medium sized language models.☆479Updated this week
- Implementation of RETRO, Deepmind's Retrieval based Attention net, in Pytorch☆851Updated last year
- Tools for understanding how transformer predictions are built layer-by-layer☆429Updated 5 months ago
- This project studies the performance and robustness of language models and task-adaptation methods.☆141Updated 5 months ago
- A prize for finding tasks that cause large language models to show inverse scaling☆597Updated last year
- The official code of LM-Debugger, an interactive tool for inspection and intervention in transformer-based language models.☆172Updated 2 years ago
- ☆258Updated 3 weeks ago
- Fast Inference Solutions for BLOOM☆560Updated last month
- Scalable toolkit for efficient model alignment☆611Updated this week