huggingface / datablations
Scaling Data-Constrained Language Models
☆330Updated 3 months ago
Alternatives and similar repositories for datablations:
Users that are interested in datablations are comparing it to the libraries listed below
- DSIR large-scale data selection framework for language model training☆242Updated 9 months ago
- Simple next-token-prediction for RLHF☆222Updated last year
- Multipack distributed sampler for fast padding-free training of LLMs☆184Updated 5 months ago
- NeurIPS Large Language Model Efficiency Challenge: 1 LLM + 1GPU + 1Day☆253Updated last year
- Pytorch implementation of DoReMi, a method for optimizing the data mixture weights in language modeling datasets☆313Updated last year
- ☆251Updated last year
- Code for the paper "Rethinking Benchmark and Contamination for Language Models with Rephrased Samples"☆297Updated last year
- Positional Skip-wise Training for Efficient Context Window Extension of LLMs to Extremely Length (ICLR 2024)☆204Updated 7 months ago
- Repo for Rho-1: Token-level Data Selection & Selective Pretraining of LLMs.☆382Updated 9 months ago
- Manage scalable open LLM inference endpoints in Slurm clusters☆247Updated 6 months ago
- Self-Alignment with Principle-Following Reward Models☆150Updated 10 months ago
- Pre-training code for Amber 7B LLM☆160Updated 8 months ago
- A repository for research on medium sized language models.☆484Updated this week
- Implementation of paper Data Engineering for Scaling Language Models to 128K Context☆447Updated 9 months ago
- [EMNLP 2023] The CoT Collection: Improving Zero-shot and Few-shot Learning of Language Models via Chain-of-Thought Fine-Tuning☆221Updated last year
- Implementation of CALM from the paper "LLM Augmented LLMs: Expanding Capabilities through Composition", out of Google Deepmind☆172Updated 4 months ago
- RewardBench: the first evaluation tool for reward models.☆491Updated last week
- batched loras☆336Updated last year
- A project to improve skills of large language models☆230Updated this week
- This repository contains code to quantitatively evaluate instruction-tuned models such as Alpaca and Flan-T5 on held-out tasks.☆538Updated 10 months ago
- This project studies the performance and robustness of language models and task-adaptation methods.☆142Updated 7 months ago
- Learning to Compress Prompts with Gist Tokens - https://arxiv.org/abs/2304.08467☆274Updated last year
- Official github repo for the paper "Compression Represents Intelligence Linearly" [COLM 2024]☆130Updated 3 months ago
- Official PyTorch implementation of QA-LoRA☆122Updated 10 months ago
- Reproducible, flexible LLM evaluations☆118Updated last month
- [EMNLP 2023] Adapting Language Models to Compress Long Contexts☆292Updated 4 months ago
- Benchmarking LLMs with Challenging Tasks from Real Users☆206Updated 2 months ago
- Official repository of NEFTune: Noisy Embeddings Improves Instruction Finetuning☆388Updated 8 months ago
- Explorations into some recent techniques surrounding speculative decoding☆229Updated 3 weeks ago
- Experiments on speculative sampling with Llama models☆122Updated last year