leogao2 / lm_dataformat
☆76Updated 11 months ago
Related projects ⓘ
Alternatives and complementary repositories for lm_dataformat
- Repo for training MLMs, CLMs, or T5-type models on the OLM pretraining data, but it should work with any hugging face text dataset.☆92Updated last year
- ☆97Updated 2 years ago
- ☆67Updated 2 years ago
- ☆95Updated last year
- Transformers at any scale☆41Updated 10 months ago
- ARCHIVED. Please use https://docs.adapterhub.ml/huggingface_hub.html || 🔌 A central repository collecting pre-trained adapter modules☆68Updated 5 months ago
- ☆73Updated last year
- ☆86Updated 2 years ago
- Experiments with generating opensource language model assistants☆97Updated last year
- Tools for managing datasets for governance and training.☆78Updated 3 weeks ago
- A framework for few-shot evaluation of autoregressive language models.☆101Updated last year
- Open Instruction Generalist is an assistant trained on massive synthetic instructions to perform many millions of tasks☆206Updated 10 months ago
- [TMLR'23] Contrastive Search Is What You Need For Neural Text Generation☆118Updated last year
- Techniques used to run BLOOM at inference in parallel☆37Updated 2 years ago
- Python tools for processing the stackexchange data dumps into a text dataset for Language Models☆76Updated 11 months ago
- Pipeline for pulling and processing online language model pretraining data from the web☆174Updated last year
- Datasets collection and preprocessings framework for NLP extreme multitask learning☆149Updated 4 months ago
- ☆46Updated this week
- Implementation of Marge, Pre-training via Paraphrasing, in Pytorch☆75Updated 3 years ago
- Evaluation suite for large-scale language models.☆124Updated 3 years ago
- ☆45Updated 2 years ago
- The official code of EMNLP 2022, "SCROLLS: Standardized CompaRison Over Long Language Sequences".☆68Updated 10 months ago
- A diff tool for language models☆42Updated 10 months ago
- Official code and model checkpoints for our EMNLP 2022 paper "RankGen - Improving Text Generation with Large Ranking Models" (https://arx…☆136Updated last year
- XtremeDistil framework for distilling/compressing massive multilingual neural network models to tiny and efficient models for AI at scale☆153Updated 11 months ago
- Experiments on including metadata such as URLs, timestamps, website descriptions and HTML tags during pretraining.☆30Updated last year
- [AAAI 2024] Investigating the Effectiveness of Task-Agnostic Prefix Prompt for Instruction Following☆79Updated 2 months ago
- Tutorial to pretrain & fine-tune a 🤗 Flax T5 model on a TPUv3-8 with GCP☆58Updated 2 years ago
- A minimal PyTorch Lightning OpenAI GPT w DeepSpeed Training!☆111Updated last year
- ☆55Updated last year