uiuc-focal-lab / syncode
Efficient and general syntactical decoding for Large Language Models
β242Updated this week
Alternatives and similar repositories for syncode:
Users that are interested in syncode are comparing it to the libraries listed below
- π€ A specialized library for integrating context-free grammars (CFG) in EBNF with the Hugging Face Transformersβ110Updated this week
- r2e: turn any github repository into a programming agent environmentβ101Updated last week
- A curated list of papers related to constrained decoding of LLM, along with their relevant code and resources.β172Updated last week
- CRUXEval: Code Reasoning, Understanding, and Execution Evaluationβ131Updated 5 months ago
- RepoQA: Evaluating Long-Context Code Understandingβ105Updated 4 months ago
- EvoEval: Evolving Coding Benchmarks via LLMβ67Updated 11 months ago
- A multi-programming language benchmark for LLMsβ235Updated last month
- β¨ RepoBench: Benchmarking Repository-Level Code Auto-Completion Systems - ICLR 2024β146Updated 6 months ago
- β106Updated 7 months ago
- CrossCodeEval: A Diverse and Multilingual Benchmark for Cross-File Code Completion (NeurIPS 2023)β133Updated 7 months ago
- Code and Data artifact for NeurIPS 2023 paper - "Monitor-Guided Decoding of Code LMs with Static Analysis of Repository Context". `multisβ¦β243Updated 7 months ago
- Benchmark ClassEval for class-level code generation.β135Updated 4 months ago
- Official implementation of paper How to Understand Whole Repository? New SOTA on SWE-bench Lite (21.3%)β70Updated 4 months ago
- β30Updated 4 months ago
- Enhancing AI Software Engineering with Repository-level Code Graphβ143Updated 2 months ago
- [NeurIPS 2024] Evaluation harness for SWT-Bench, a benchmark for evaluating LLM repository-level test-generationβ36Updated this week
- Pip compatible CodeBLEU metric implementation available for linux/macos/winβ80Updated last week
- DafnyBench: A Benchmark for Formal Software Verificationβ25Updated 3 months ago
- β121Updated last year
- Can It Edit? Evaluating the Ability of Large Language Models to Follow Code Editing Instructionsβ41Updated 7 months ago
- β13Updated last year
- Data and evaluation scripts for "CodePlan: Repository-level Coding using LLMs and Planning", FSE 2024β64Updated 6 months ago
- [NeurIPS'24] SelfCodeAlign: Self-Alignment for Code Generationβ298Updated 2 weeks ago
- XFT: Unlocking the Power of Code Instruction Tuning by Simply Merging Upcycled Mixture-of-Expertsβ30Updated 8 months ago
- xCodeEval: A Large Scale Multilingual Multitask Benchmark for Code Understanding, Generation, Translation and Retrievalβ78Updated 5 months ago
- LDB: A Large Language Model Debugger via Verifying Runtime Execution Step by Stepβ504Updated 6 months ago
- A collection of practical code generation tasks and tests in open source projects. Complementary to HumanEval by OpenAI.β134Updated 2 months ago
- β59Updated 10 months ago
- β85Updated 7 months ago