juyongjiang / CodeLLMSurveyLinks
[TOSEM'25] The official GitHub page for the survey paper "A Survey on Large Language Models for Code Generation".
β162Updated 3 months ago
Alternatives and similar repositories for CodeLLMSurvey
Users that are interested in CodeLLMSurvey are comparing it to the libraries listed below
Sorting:
- Repo-Level Code generation papersβ215Updated 3 months ago
- Must-read papers on Repository-level Code Generation & Issue Resolution π₯β195Updated this week
- An Evolving Code Generation Benchmark Aligned with Real-world Code Repositoriesβ63Updated last year
- Reproducing R1 for Code with Reliable Rewardsβ262Updated 5 months ago
- Neural Code Intelligence Survey 2024; Reading lists and resourcesβ275Updated 3 months ago
- The repository for paper "DebugBench: "Evaluating Debugging Capability of Large Language Models".β83Updated last year
- β23Updated 2 months ago
- A Comprehensive Benchmark for Software Development.β115Updated last year
- A comprehensive code domain benchmark review of LLM researches.β131Updated last month
- β190Updated 3 months ago
- A collection of practical code generation tasks and tests in open source projects. Complementary to HumanEval by OpenAI.β152Updated 10 months ago
- [ICML 2023] Data and code release for the paper "DS-1000: A Natural and Reliable Benchmark for Data Science Code Generation".β256Updated last year
- Generative AI Act II: Test Time Scaling Drives Cognition Engineeringβ207Updated 6 months ago
- An Awesome List of Agentic Model trained with Reinforcement Learningβ527Updated 2 weeks ago
- Survey on LLM Agents (Published on CoLing 2025)β399Updated 3 weeks ago
- Repository for the paper "Large Language Model-Based Agents for Software Engineering: A Survey". Keep updating.β506Updated 7 months ago
- Benchmark ClassEval for class-level code generation.β145Updated last year
- π A Survey of Efficient Reasoning for Large Reasoning Models: Language, Multimodality, Agent, and Beyondβ309Updated last week
- A Framework for LLM-based Multi-Agent Reinforced Training and Inferenceβ316Updated this week
- A Manually-Annotated Code Generation Benchmark Aligned with Real-World Code Repositoriesβ33Updated last year
- Awesome-Long2short-on-LRMs is a collection of state-of-the-art, novel, exciting long2short methods on large reasoning models. It containsβ¦β252Updated 2 months ago
- [EMNLP 2024] CodeJudge: Evaluating Code Generation with Large Language Modelsβ50Updated last month
- Official repository for our paper "FullStack Bench: Evaluating LLMs as Full Stack Coders"β106Updated 5 months ago
- LLM-Based Human-Agent Collaboration and Interaction Systems: A Survey | Awesome Human-Agent Collaboration | Human-AI Collaborationβ153Updated 2 weeks ago
- β303Updated 5 months ago
- CrossCodeEval: A Diverse and Multilingual Benchmark for Cross-File Code Completion (NeurIPS 2023)β159Updated 2 months ago
- Latest Advances on Long Chain-of-Thought Reasoningβ537Updated 3 months ago
- This is the repository for the Tool Learning survey.β449Updated 2 months ago
- β367Updated 2 weeks ago
- [ICLR 2025] Benchmarking Agentic Workflow Generationβ129Updated 8 months ago