ChiYeungLaw / Awsome-Code-Intelligence
In this repository, you'll find a curated selection of recent research papers, articles, and implementations from leading experts in the field of Code Intelligence.
☆16Updated last year
Alternatives and similar repositories for Awsome-Code-Intelligence:
Users that are interested in Awsome-Code-Intelligence are comparing it to the libraries listed below
- The repository for paper "DebugBench: "Evaluating Debugging Capability of Large Language Models".☆62Updated 7 months ago
- StepCoder: Improve Code Generation with Reinforcement Learning from Compiler Feedback☆62Updated 5 months ago
- Code for the TMLR 2023 paper "PPOCoder: Execution-based Code Generation using Deep Reinforcement Learning"☆108Updated last year
- Code for the paper <SelfCheck: Using LLMs to Zero-Shot Check Their Own Step-by-Step Reasoning>☆48Updated last year
- XFT: Unlocking the Power of Code Instruction Tuning by Simply Merging Upcycled Mixture-of-Experts☆29Updated 7 months ago
- Large Language Models Meet NL2Code: A Survey☆36Updated 2 months ago
- ☆13Updated 2 months ago
- Source codes for paper ”ReACC: A Retrieval-Augmented Code Completion Framework“☆61Updated 2 years ago
- Code for the AAAI 2023 paper "CodeAttack: Code-based Adversarial Attacks for Pre-Trained Programming Language Models☆26Updated last year
- Repoformer: Selective Retrieval for Repository-Level Code Completion (ICML 2024)☆50Updated 7 months ago
- Training language models to make programs faster☆85Updated 10 months ago
- [EMNLP'22] Code for 'Exploring Representation-level Augmentation for Code Search'☆26Updated last year
- Releasing code for "ReCode: Robustness Evaluation of Code Generation Models"☆52Updated 10 months ago
- R-Judge: Benchmarking Safety Risk Awareness for LLM Agents (EMNLP Findings 2024)☆65Updated last week
- [ICLR 2024] Evaluating Large Language Models at Evaluating Instruction Following☆119Updated 7 months ago
- Code for our EMNLP-2023 paper: "Active Instruction Tuning: Improving Cross-Task Generalization by Training on Prompt Sensitive Tasks"☆24Updated last year
- APIBench is a benchmark for evaluating the performance of API recommendation approaches released in the paper "Revisiting, Benchmarking a…☆53Updated last year
- ☆28Updated 3 months ago
- ☆33Updated last year
- Official implementation of Privacy Implications of Retrieval-Based Language Models (EMNLP 2023). https://arxiv.org/abs/2305.14888☆35Updated 8 months ago
- Awesome LLM Self-Consistency: a curated list of Self-consistency in Large Language Models☆86Updated 6 months ago
- ☆46Updated 2 years ago
- [ICML 2023] Data and code release for the paper "DS-1000: A Natural and Reliable Benchmark for Data Science Code Generation".☆233Updated 3 months ago
- An Evolving Code Generation Benchmark Aligned with Real-world Code Repositories☆48Updated 6 months ago
- Code and dataset for EMNLP 2022 Findings paper "Benchmarking Language Models for Code Syntax Understanding"☆14Updated 2 years ago
- A Comprehensive Benchmark for Software Development.☆91Updated 8 months ago
- Repository for the paper "Cognitive Mirage: A Review of Hallucinations in Large Language Models"☆47Updated last year
- ☆31Updated 2 months ago
- [ACL'24] A Knowledge-grounded Interactive Evaluation Framework for Large Language Models☆36Updated 6 months ago