bigscience-workshop / bigscience
Central place for the engineering/scaling WG: documentation, SLURM scripts and logs, compute environment and data.
☆992Updated 8 months ago
Alternatives and similar repositories for bigscience:
Users that are interested in bigscience are comparing it to the libraries listed below
- Ongoing research training transformer language models at scale, including: BERT & GPT-2☆1,386Updated last year
- Fast Inference Solutions for BLOOM☆561Updated 6 months ago
- distributed trainer for LLMs☆572Updated 11 months ago
- Task-based datasets, preprocessing, and evaluation for sequence models.☆573Updated 2 weeks ago
- Ongoing research training transformer language models at scale, including: BERT & GPT-2☆2,049Updated last month
- Mistral: A strong, northwesterly wind: Framework for transparent and accessible large-scale language model training, built with Hugging F…☆571Updated last year
- Crosslingual Generalization through Multitask Finetuning☆531Updated 7 months ago
- Cramming the training of a (BERT-type) language model into limited compute.☆1,331Updated 10 months ago
- Implementation of the specific Transformer architecture from PaLM - Scaling Language Modeling with Pathways