ExpertiseModel / MuTAP
MutAP: A prompt_based learning technique to automatically generate test cases with Large Language Model
☆31Updated last month
Related projects ⓘ
Alternatives and complementary repositories for MuTAP
- This repo is for our submission for ICSE 2025.☆18Updated 5 months ago
- Replication package of a paper "Large Language Models are Few-shot Testers: Exploring LLM-based General Bug Reproduction"☆22Updated last year
- ☆85Updated 3 weeks ago
- [TOSEM 2023] A Survey of Learning-based Automated Program Repair☆68Updated 6 months ago
- Enhancing AI Software Engineering with Repository-level Code Graph☆96Updated 2 months ago
- Official implementation of paper How to Understand Whole Repository? New SOTA on SWE-bench Lite (21.3%)☆67Updated last week
- A Systematic Literature Review on Large Language Models for Automated Program Repair☆129Updated this week
- Benchmark ClassEval for class-level code generation.☆126Updated 3 weeks ago
- Large Language Models for Software Engineering☆191Updated this week
- Data and evaluation scripts for "CodePlan: Repository-level Coding using LLMs and Planning", FSE 2024☆52Updated 2 months ago
- [2023 ASE] GAMMA: Revisiting Template-based Automated Program Repair via Mask Prediction☆18Updated last year
- ☆29Updated last year
- ☆12Updated last month
- Dataflow-guided retrieval augmentation for repository-level code completion, ACL 2024 (main)☆21Updated 5 months ago
- Code and data for XLCoST: A Benchmark Dataset for Cross-lingual Code Intelligence☆66Updated last year
- ☆117Updated last year
- [2023 TDSC] Pre-trained Model-based Automated Software Vulnerability Repair: How Far are We?☆21Updated last year
- An Evolving Code Generation Benchmark Aligned with Real-world Code Repositories☆46Updated 3 months ago
- A framework to generate unit tests using LLMs☆23Updated this week
- TeCo: an ML+Execution model for test completion☆29Updated 5 months ago
- A collection of practical code generation tasks and tests in open source projects. Complementary to HumanEval by OpenAI.☆121Updated 11 months ago
- Scalable and Accurate Test Case Prioritization in Continuous Integration Contexts☆16Updated 2 years ago
- Collections of research, benchmarks and tools towards more robust and reliable language models for code; LM4Code; LM4SE; reliable LLM; L…☆22Updated 11 months ago
- ☆296Updated 10 months ago
- A Reproducible Benchmark of Recent Java Bugs☆27Updated 2 weeks ago
- MODIT: On Multi-Modal Learning of Editing Source Code.☆20Updated 3 years ago
- ☆15Updated 11 months ago
- Efficient APR with LLMs http://arxiv.org/pdf/2402.06598☆13Updated 5 months ago
- EvoEval: Evolving Coding Benchmarks via LLM☆60Updated 7 months ago
- methods2test is a supervised dataset consisting of Test Cases and their corresponding Focal Methods from a set of Java software repositor…☆134Updated 11 months ago