llm-efficiency-challenge / neurips_llm_efficiency_challenge
NeurIPS Large Language Model Efficiency Challenge: 1 LLM + 1GPU + 1Day
☆252Updated last year
Related projects ⓘ
Alternatives and complementary repositories for neurips_llm_efficiency_challenge
- Scaling Data-Constrained Language Models☆321Updated last month
- DSIR large-scale data selection framework for language model training☆230Updated 7 months ago
- Manage scalable open LLM inference endpoints in Slurm clusters☆236Updated 4 months ago
- Code for the paper "Rethinking Benchmark and Contamination for Language Models with Rephrased Samples"☆293Updated 11 months ago
- Multipack distributed sampler for fast padding-free training of LLMs☆178Updated 3 months ago
- Understand and test language model architectures on synthetic tasks.☆162Updated 6 months ago
- ☆451Updated 3 weeks ago
- Code for "LayerSkip: Enabling Early Exit Inference and Self-Speculative Decoding", ACL 2024☆229Updated 3 weeks ago
- BABILong is a benchmark for LLM evaluation using the needle-in-a-haystack approach.☆152Updated last week
- batched loras☆336Updated last year
- Official PyTorch implementation of QA-LoRA☆117Updated 8 months ago
- ☆175Updated this week
- Repo for Rho-1: Token-level Data Selection & Selective Pretraining of LLMs.☆307Updated 7 months ago
- Website for hosting the Open Foundation Models Cheat Sheet.☆257Updated 4 months ago
- ☆91Updated last year
- The official evaluation suite and dynamic data release for MixEval.☆224Updated last week
- Implementation of CALM from the paper "LLM Augmented LLMs: Expanding Capabilities through Composition", out of Google Deepmind☆169Updated 2 months ago
- A set of scripts and notebooks on LLM finetunning and dataset creation☆93Updated last month
- 🚀 Efficiently (pre)training foundation models with native PyTorch features, including FSDP for training and SDPA implementation of Flash…☆193Updated this week
- ModuleFormer is a MoE-based architecture that includes two different types of experts: stick-breaking attention heads and feedforward exp…☆216Updated 7 months ago
- Tools for understanding how transformer predictions are built layer-by-layer☆430Updated 5 months ago
- Official repository for ORPO☆421Updated 5 months ago
- ☆199Updated 5 months ago
- Scalable toolkit for efficient model alignment☆620Updated this week
- The official implementation of the paper "What Matters in Transformers? Not All Attention is Needed".☆141Updated this week
- A repository for research on medium sized language models.☆479Updated this week
- Recurrent Memory Transformer☆149Updated last year
- ☆247Updated last year
- Starter pack for NeurIPS LLM Efficiency Challenge 2023.☆118Updated last year
- Explorations into some recent techniques surrounding speculative decoding☆211Updated last year