uiuc-focal-lab / syncode
Efficient and general syntactical decoding for Large Language Models
☆177Updated this week
Related projects: ⓘ
- ☆77Updated this week
- RepoQA: Evaluating Long-Context Code Understanding☆96Updated this week
- Can It Edit? Evaluating the Ability of Large Language Models to Follow Code Editing Instructions☆38Updated last month
- EvoEval: Evolving Coding Benchmarks via LLM☆57Updated 5 months ago
- CrossCodeEval: A Diverse and Multilingual Benchmark for Cross-File Code Completion (NeurIPS 2023)☆114Updated last month
- ✨ RepoBench: Benchmarking Repository-Level Code Auto-Completion Systems - ICLR 2024☆129Updated last month
- A multi-programming language benchmark for LLMs☆189Updated this week
- CRUXEval: Code Reasoning, Understanding, and Execution Evaluation☆99Updated last month
- Enhancing AI Software Engineering with Repository-level Code Graph☆60Updated 3 weeks ago
- ☆111Updated last year
- Code and Data artifact for NeurIPS 2023 paper - "Monitor-Guided Decoding of Code LMs with Static Analysis of Repository Context". `multis…☆189Updated last month
- ☆53Updated 4 months ago
- ☆74Updated 2 months ago
- 🤗 A specialized library for integrating context-free grammars (CFG) in EBNF with the Hugging Face Transformers☆82Updated last week
- CodeMind is a generic framework for evaluating inductive code reasoning of LLMs. It is equipped with a static analysis component that ena…☆31Updated last month
- XFT: Unlocking the Power of Code Instruction Tuning by Simply Merging Upcycled Mixture-of-Experts☆24Updated 2 months ago
- ☆18Updated 2 weeks ago
- Benchmark ClassEval for class-level code generation.☆120Updated 2 weeks ago
- DafnyBench: A Benchmark for Formal Software Verification☆13Updated last month
- ☆96Updated 2 months ago
- Training language models to make programs faster☆80Updated 5 months ago
- CodeBERTScore: an automatic metric for code generation, based on BERTScore☆156Updated 6 months ago
- Open sourced predictions, execution logs, trajectories, and results from model inference + evaluation runs on the SWE-bench task.☆79Updated 2 weeks ago
- Official repository for the paper "LiveCodeBench: Holistic and Contamination Free Evaluation of Large Language Models for Code"☆173Updated 3 weeks ago
- StarCoder2-Instruct: Fully Transparent and Permissive Self-Alignment for Code Generation☆221Updated 2 months ago
- Repository for the paper "Large Language Model-Based Agents for Software Engineering: A Survey".☆205Updated last week
- Data and evaluation scripts for "CodePlan: Repository-level Coding using LLMs and Planning", FSE 2024☆44Updated 2 weeks ago
- Repo-Level Code generation papers☆66Updated 3 months ago
- [EACL 2024] ICE-Score: Instructing Large Language Models to Evaluate Code☆67Updated 3 months ago
- Code and data for XLCoST: A Benchmark Dataset for Cross-lingual Code Intelligence☆63Updated last year