structuredllm / syncodeLinks
Efficient and general syntactical decoding for Large Language Models
β272Updated this week
Alternatives and similar repositories for syncode
Users that are interested in syncode are comparing it to the libraries listed below
Sorting:
- π€ A specialized library for integrating context-free grammars (CFG) in EBNF with the Hugging Face Transformersβ118Updated last month
- A certifier for bias in LLMsβ23Updated last month
- A curated list of papers related to constrained decoding of LLM, along with their relevant code and resources.β211Updated last month
- r2e: turn any github repository into a programming agent environmentβ121Updated last month
- RepoQA: Evaluating Long-Context Code Understandingβ108Updated 7 months ago
- β¨ RepoBench: Benchmarking Repository-Level Code Auto-Completion Systems - ICLR 2024β164Updated 9 months ago
- Enhancing AI Software Engineering with Repository-level Code Graphβ179Updated 2 months ago
- Iterate on LLM-based structured generation forward and backwardβ15Updated 2 months ago
- CRUXEval: Code Reasoning, Understanding, and Execution Evaluationβ141Updated 7 months ago
- EvoEval: Evolving Coding Benchmarks via LLMβ72Updated last year
- Artifact repository for the paper "Lost in Translation: A Study of Bugs Introduced by Large Language Models while Translating Code", In Pβ¦β49Updated last month
- Code and Data artifact for NeurIPS 2023 paper - "Monitor-Guided Decoding of Code LMs with Static Analysis of Repository Context". `multisβ¦β260Updated 9 months ago
- CrossCodeEval: A Diverse and Multilingual Benchmark for Cross-File Code Completion (NeurIPS 2023)β143Updated 10 months ago
- Open sourced predictions, execution logs, trajectories, and results from model inference + evaluation runs on the SWE-bench task.β178Updated last week
- β15Updated last year
- β110Updated 10 months ago
- [FORGE 2025] Graph-based method for end-to-end code completion with context awareness on repositoryβ63Updated 9 months ago
- Can It Edit? Evaluating the Ability of Large Language Models to Follow Code Editing Instructionsβ42Updated 10 months ago
- [ICML 2023] Data and code release for the paper "DS-1000: A Natural and Reliable Benchmark for Data Science Code Generation".β244Updated 7 months ago
- The repository for paper "DebugBench: "Evaluating Debugging Capability of Large Language Models".β77Updated 10 months ago
- A benchmark for LLMs on complicated tasks in the terminalβ141Updated this week
- [NeurIPS 2024] Evaluation harness for SWT-Bench, a benchmark for evaluating LLM repository-level test-generationβ49Updated last month
- β123Updated 2 years ago
- CodeRAG-Bench: Can Retrieval Augment Code Generation?β136Updated 6 months ago
- CodeSage: Code Representation Learning At Scale (ICLR 2024)β107Updated 7 months ago
- [NeurIPS'24] SelfCodeAlign: Self-Alignment for Code Generationβ307Updated 3 months ago
- Official code for the paper "CodeChain: Towards Modular Code Generation Through Chain of Self-revisions with Representative Sub-modules"β45Updated 4 months ago
- β62Updated last year
- β94Updated 10 months ago
- Data and evaluation scripts for "CodePlan: Repository-level Coding using LLMs and Planning", FSE 2024β69Updated 9 months ago