TurkuNLP / Megatron-DeepSpeedLinks
Ongoing research training transformer language models at scale, including: BERT & GPT-2
☆20Updated 2 years ago
Alternatives and similar repositories for Megatron-DeepSpeed
Users that are interested in Megatron-DeepSpeed are comparing it to the libraries listed below
Sorting:
- ☆72Updated 2 years ago
- [ICLR 2023] Guess the Instruction! Flipped Learning Makes Language Models Stronger Zero-Shot Learners☆116Updated 7 months ago
- ☆38Updated last year
- Code for Zero-Shot Tokenizer Transfer☆142Updated last year
- Repo for training MLMs, CLMs, or T5-type models on the OLM pretraining data, but it should work with any hugging face text dataset.☆96Updated 3 years ago
- ☆65Updated 2 years ago
- The official code of LM-Debugger, an interactive tool for inspection and intervention in transformer-based language models.☆182Updated 3 years ago
- Pipeline for pulling and processing online language model pretraining data from the web☆177Updated 2 years ago
- ☆102Updated 3 years ago
- Experiments with generating opensource language model assistants☆97Updated 2 years ago
- Techniques used to run BLOOM at inference in parallel☆37Updated 3 years ago
- Repo for ICML23 "Why do Nearest Neighbor Language Models Work?"☆59Updated 3 years ago
- SILO Language Models code repository☆83Updated last year
- ☆44Updated last year
- ☆67Updated 3 years ago
- Minimum Bayes Risk Decoding for Hugging Face Transformers☆60Updated last year
- Experiments on including metadata such as URLs, timestamps, website descriptions and HTML tags during pretraining.☆31Updated 2 years ago
- A framework for few-shot evaluation of autoregressive language models.☆106Updated 2 years ago
- Tools for managing datasets for governance and training.☆87Updated 2 weeks ago
- some common Huggingface transformers in maximal update parametrization (µP)☆87Updated 3 years ago
- ☆77Updated last year
- ☆52Updated 2 years ago
- INCOME: An Easy Repository for Training and Evaluation of Index Compression Methods in Dense Retrieval. Includes BPR and JPQ.☆24Updated 2 years ago
- [EMNLP'23] Official Code for "FOCUS: Effective Embedding Initialization for Monolingual Specialization of Multilingual Models"☆36Updated 8 months ago
- [EMNLP'23] Execution-Based Evaluation for Open Domain Code Generation☆49Updated 2 years ago
- SWIM-IR is a Synthetic Wikipedia-based Multilingual Information Retrieval training set with 28 million query-passage pairs spanning 33 la…☆49Updated 2 years ago
- The original implementation of Min et al. "Nonparametric Masked Language Modeling" (paper https//arxiv.org/abs/2212.01349)☆158Updated 3 years ago
- Code for SaGe subword tokenizer (EACL 2023)☆27Updated last year
- ☆54Updated last year
- Code repository for the c-BTM paper☆108Updated 2 years ago