huybery / Awesome-Code-LLMLinks
👨💻 An awesome and curated list of best code-LLM for research.
☆1,277Updated last year
Alternatives and similar repositories for Awesome-Code-LLM
Users that are interested in Awesome-Code-LLM are comparing it to the libraries listed below
Sorting:
- Rigourous evaluation of LLM-synthesized code - NeurIPS 2023 & COLM 2024☆1,683Updated 4 months ago
- A framework for the evaluation of autoregressive code generation language models.☆1,020Updated 6 months ago
- [TMLR] A curated list of language modeling researches for code (and other software engineering activities), plus related datasets.☆3,215Updated last week
- The papers are organized according to our survey: Evaluating Large Language Models: A Comprehensive Survey.☆792Updated last year
- A collection of awesome-prompt-datasets, awesome-instruction-dataset, to train ChatLLM such as chatgpt 收录各种各样的指令数据集, 用于训练 ChatLLM 模型。☆721Updated last year
- A collection of benchmarks and datasets for evaluating LLM.☆550Updated last year
- 🐙 OctoPack: Instruction Tuning Code Large Language Models☆479Updated last year
- ☆490Updated last year
- Run evaluation on LLMs using human-eval benchmark☆427Updated 2 years ago
- Awesome things about LLM-powered agents. Papers / Repos / Blogs / ...☆2,194Updated 9 months ago
- [ICLR'25] BigCodeBench: Benchmarking Code Generation Towards AGI☆477Updated last month
- LDB: A Large Language Model Debugger via Verifying Runtime Execution Step by Step (ACL'24)☆576Updated last year
- A library for advanced large language model reasoning☆2,328Updated 8 months ago
- Official repository for the paper "LiveCodeBench: Holistic and Contamination Free Evaluation of Large Language Models for Code"☆786Updated 6 months ago
- ☆671Updated last year
- Benchmarking large language models' complex reasoning ability with chain-of-thought prompting☆2,769Updated last year
- [ACL 2023] Reasoning with Language Model Prompting: A Survey☆994Updated 8 months ago
- Awesome-LLM-Eval: a curated list of tools, datasets/benchmark, demos, leaderboard, papers, docs and models, mainly for Evaluation on LLMs…☆607Updated 2 months ago
- LLMs can generate feedback on their work, use it to improve the output, and repeat this process iteratively.☆778Updated last year
- A curated list of awesome LLM agents frameworks.☆1,301Updated last week
- List of language agents based on paper "Cognitive Architectures for Language Agents"☆1,162Updated last year
- [ICML'24 Spotlight] LLM Maybe LongLM: Self-Extend LLM Context Window Without Tuning☆667Updated last year
- An automatic evaluator for instruction-following language models. Human-validated, high-quality, cheap, and fast.☆1,940Updated 6 months ago
- Code for the paper "Evaluating Large Language Models Trained on Code"☆3,114Updated last year
- [ACL2023] We introduce LLM-Blender, an innovative ensembling framework to attain consistently superior performance by leveraging the dive…☆976Updated last year
- A collection of open-source dataset to train instruction-following LLMs (ChatGPT,LLaMA,Alpaca)☆1,143Updated 2 years ago
- Agentless🐱: an agentless approach to automatically solve software development problems☆2,006Updated last year
- 📰 Must-read papers and blogs on LLM based Long Context Modeling 🔥☆1,909Updated 2 weeks ago
- High Accuracy and efficiency multi-task fine-tuning framework for Code LLMs. This work has been accepted by KDD 2024.☆708Updated last year
- ☆1,338Updated last year