llm-efficiency-challenge / neurips_llm_efficiency_challengeLinks
NeurIPS Large Language Model Efficiency Challenge: 1 LLM + 1GPU + 1Day
☆254Updated last year
Alternatives and similar repositories for neurips_llm_efficiency_challenge
Users that are interested in neurips_llm_efficiency_challenge are comparing it to the libraries listed below
Sorting:
- Scaling Data-Constrained Language Models☆334Updated 8 months ago
- Code for the paper "Rethinking Benchmark and Contamination for Language Models with Rephrased Samples"☆302Updated last year
- Manage scalable open LLM inference endpoints in Slurm clusters☆258Updated 10 months ago
- DSIR large-scale data selection framework for language model training☆249Updated last year
- ModuleFormer is a MoE-based architecture that includes two different types of experts: stick-breaking attention heads and feedforward exp…☆221Updated last year
- Website for hosting the Open Foundation Models Cheat Sheet.☆267Updated 3 weeks ago
- batched loras☆343Updated last year
- Multipack distributed sampler for fast padding-free training of LLMs☆188Updated 9 months ago
- The official evaluation suite and dynamic data release for MixEval.☆242Updated 6 months ago
- Official PyTorch implementation of QA-LoRA☆135Updated last year
- RuLES: a benchmark for evaluating rule-following in language models☆224Updated 3 months ago
- Starter pack for NeurIPS LLM Efficiency Challenge 2023.☆122Updated last year
- ☆258Updated last year
- git extension for {collaborative, communal, continual} model development☆212Updated 6 months ago
- ☆92Updated last year
- experiments with inference on llama☆104Updated 11 months ago
- The Official Repository for "Bring Your Own Data! Self-Supervised Evaluation for Large Language Models"☆108Updated last year
- Pre-training code for Amber 7B LLM☆166Updated last year
- Fast bare-bones BPE for modern tokenizer training☆157Updated 2 months ago
- A repository for research on medium sized language models.☆497Updated last month
- A puzzle to learn about prompting☆127Updated 2 years ago
- Implementation of CALM from the paper "LLM Augmented LLMs: Expanding Capabilities through Composition", out of Google Deepmind☆178Updated 8 months ago
- This project studies the performance and robustness of language models and task-adaptation methods.☆150Updated last year
- Understand and test language model architectures on synthetic tasks.☆197Updated 2 months ago
- A set of scripts and notebooks on LLM finetunning and dataset creation☆111Updated 8 months ago
- Simple next-token-prediction for RLHF☆226Updated last year
- some common Huggingface transformers in maximal update parametrization (µP)☆80Updated 3 years ago
- Implementation of 💍 Ring Attention, from Liu et al. at Berkeley AI, in Pytorch☆514Updated 2 weeks ago
- LOFT: A 1 Million+ Token Long-Context Benchmark☆198Updated last month
- ☆125Updated last year