structuredllm / syncodeLinks
Efficient and general syntactical decoding for Large Language Models
β287Updated this week
Alternatives and similar repositories for syncode
Users that are interested in syncode are comparing it to the libraries listed below
Sorting:
- π€ A specialized library for integrating context-free grammars (CFG) in EBNF with the Hugging Face Transformersβ124Updated 4 months ago
- r2e: turn any github repository into a programming agent environmentβ130Updated 4 months ago
- β70Updated last year
- EvoEval: Evolving Coding Benchmarks via LLMβ76Updated last year
- β¨ RepoBench: Benchmarking Repository-Level Code Auto-Completion Systems - ICLR 2024β170Updated last year
- A curated list of papers related to constrained decoding of LLM, along with their relevant code and resources.β251Updated last week
- A multi-programming language benchmark for LLMsβ270Updated 2 weeks ago
- Code and Data artifact for NeurIPS 2023 paper - "Monitor-Guided Decoding of Code LMs with Static Analysis of Repository Context". `multisβ¦β271Updated last year
- Enhancing AI Software Engineering with Repository-level Code Graphβ205Updated 4 months ago
- CRUXEval: Code Reasoning, Understanding, and Execution Evaluationβ152Updated 10 months ago
- CrossCodeEval: A Diverse and Multilingual Benchmark for Cross-File Code Completion (NeurIPS 2023)β155Updated last week
- Open sourced predictions, execution logs, trajectories, and results from model inference + evaluation runs on the SWE-bench task.β206Updated this week
- RepoQA: Evaluating Long-Context Code Understandingβ115Updated 9 months ago
- [NeurIPS'24] SelfCodeAlign: Self-Alignment for Code Generationβ309Updated 6 months ago
- Can It Edit? Evaluating the Ability of Large Language Models to Follow Code Editing Instructionsβ47Updated this week
- CodeBERTScore: an automatic metric for code generation, based on BERTScoreβ199Updated last year
- Pip compatible CodeBLEU metric implementation available for linux/macos/winβ108Updated 4 months ago
- β108Updated 2 months ago
- [EACL 2024] ICE-Score: Instructing Large Language Models to Evaluate Codeβ78Updated last year
- β110Updated last year
- GSO: Challenging Software Optimization Tasks for Evaluating SWE-Agentsβ37Updated last month
- [ICML 2023] Data and code release for the paper "DS-1000: A Natural and Reliable Benchmark for Data Science Code Generation".β254Updated 9 months ago
- Benchmark ClassEval for class-level code generation.β145Updated 10 months ago
- Multi-SWE-bench: A Multilingual Benchmark for Issue Resolvingβ238Updated last week
- [FORGE 2025] Graph-based method for end-to-end code completion with context awareness on repositoryβ65Updated 11 months ago
- Official implementation of paper How to Understand Whole Repository? New SOTA on SWE-bench Lite (21.3%)β92Updated 5 months ago
- CodeSage: Code Representation Learning At Scale (ICLR 2024)β112Updated 10 months ago
- The repository for paper "DebugBench: "Evaluating Debugging Capability of Large Language Models".β81Updated last year
- β124Updated 2 years ago
- Reproduction Package for the paper "Type-Constrained Code Generation with Language Models" [PLDI 2025]β69Updated 2 months ago