Saibo-creator / Awesome-LLM-Constrained-DecodingLinks
A curated list of papers related to constrained decoding of LLM, along with their relevant code and resources.
โ297Updated last month
Alternatives and similar repositories for Awesome-LLM-Constrained-Decoding
Users that are interested in Awesome-LLM-Constrained-Decoding are comparing it to the libraries listed below
Sorting:
- A simple unified framework for evaluating LLMsโ255Updated 7 months ago
- ๐ค A specialized library for integrating context-free grammars (CFG) in EBNF with the Hugging Face Transformersโ130Updated 8 months ago
- [EMNLP 2023] Adapting Language Models to Compress Long Contextsโ319Updated last year
- โ241Updated last year
- [ICML 2023] Data and code release for the paper "DS-1000: A Natural and Reliable Benchmark for Data Science Code Generation".โ260Updated last year
- open-source code for paper: Retrieval Head Mechanistically Explains Long-Context Factualityโ222Updated last year
- BABILong is a benchmark for LLM evaluation using the needle-in-a-haystack approach.โ226Updated 3 months ago
- CrossCodeEval: A Diverse and Multilingual Benchmark for Cross-File Code Completion (NeurIPS 2023)โ164Updated 3 months ago
- Reproducing R1 for Code with Reliable Rewardsโ277Updated 7 months ago
- REST: Retrieval-Based Speculative Decoding, NAACL 2024โ212Updated 2 months ago
- A simple toolkit for benchmarking LLMs on mathematical reasoning tasks. ๐งฎโจโ271Updated last year
- CRUXEval: Code Reasoning, Understanding, and Execution Evaluationโ161Updated last year
- Explorations into some recent techniques surrounding speculative decodingโ295Updated 11 months ago
- โจ RepoBench: Benchmarking Repository-Level Code Auto-Completion Systems - ICLR 2024โ182Updated last year
- Codes for the paper "โBench: Extending Long Context Evaluation Beyond 100K Tokens": https://arxiv.org/abs/2402.13718โ358Updated last year
- LOFT: A 1 Million+ Token Long-Context Benchmarkโ218Updated 5 months ago
- Efficient and general syntactical decoding for Large Language Modelsโ305Updated last week
- A curated collection of LLM reasoning and planning resources, including key papers, limitations, benchmarks, and additional learning mateโฆโ305Updated 9 months ago
- Reproducible, flexible LLM evaluationsโ293Updated 2 weeks ago
- Code and data for "Lost in the Middle: How Language Models Use Long Contexts"โ366Updated last year
- [COLM 2025] Official repository for R2E-Gym: Procedural Environment Generation and Hybrid Verifiers for Scaling Open-Weights SWE Agentsโ204Updated 4 months ago
- A Comprehensive Benchmark for Software Development.โ122Updated last year
- A comprehensive code domain benchmark review of LLM researches.โ169Updated 2 months ago
- ๐พ OAT: A research-friendly framework for LLM online alignment, including reinforcement learning, preference learning, etc.โ576Updated last month
- Awesome LLM Self-Consistency: a curated list of Self-consistency in Large Language Modelsโ113Updated 4 months ago
- Official repository for "Scaling Retrieval-Based Langauge Models with a Trillion-Token Datastore".โ220Updated last month
- [ICML '24] R2E: Turn any GitHub Repository into a Programming Agent Environmentโ136Updated 7 months ago
- Automatic evals for LLMsโ564Updated 5 months ago
- A lightweight reproduction of DeepSeek-R1-Zero with indepth analysis of self-reflection behavior.โ248Updated 7 months ago
- The HELMET Benchmarkโ187Updated 3 months ago