CoIR-team / coirLinks
(ACL 2025 Main) A Comprehensive Benchmark for Code Information Retrieval.
☆142Updated 5 months ago
Alternatives and similar repositories for coir
Users that are interested in coir are comparing it to the libraries listed below
Sorting:
- A library for generating difficulty-scalable, multi-tool, and verifiable agentic tasks with execution trajectories.☆170Updated 5 months ago
- (NeurIPS 2024) AvaTaR: Optimizing LLM Agents for Tool Usage via Contrastive Reasoning☆231Updated 6 months ago
- Grimoire is All You Need for Enhancing Large Language Models☆116Updated last year
- This includes the original implementation of CtrlA: Adaptive Retrieval-Augmented Generation via Inherent Control.☆64Updated last year
- Explore concepts like Self-Correct, Self-Refine, Self-Improve, Self-Contradict, Self-Play, and Self-Knowledge, alongside o1-like reasonin…☆171Updated last year
- Recipes to train the self-rewarding reasoning LLMs.☆230Updated 9 months ago
- MPLSandbox is an out-of-the-box multi-programming language sandbox designed to provide unified and comprehensive feedback from compiler a…☆178Updated 7 months ago
- [NeurIPS 2025 Poster] Search and Refine During Think: Facilitating Knowledge Refinement for Improved Retrieval-Augmented Reasoning☆113Updated 3 weeks ago
- A toolkit for fine-tuning, inferencing, and evaluating GreenBitAI's LLMs.☆184Updated 4 months ago
- [EMNLP 2023] CodeTransOcean: A Comprehensive Multilingual Benchmark for Code Translation☆58Updated 2 years ago
- Code and dataset of CodeSteer☆87Updated 8 months ago
- [ACL 2024] CodeScope: An Execution-based Multilingual Multitask Multidimensional Benchmark for Evaluating LLMs on Code Understanding and …☆100Updated last year
- (ICML'25 Outstanding) CollabLLM: From Passive Responders to Active Collaborators☆263Updated 2 months ago
- Code and Checkpoints for "Generate rather than Retrieve: Large Language Models are Strong Context Generators" in ICLR 2023.☆291Updated 2 years ago
- A curated list of awesome leaderboard-oriented resources for large AI models☆298Updated this week
- DocAgent is a system designed to generate high-quality, context-aware code documentation for Python codebases using a multi-agent approac…☆407Updated 7 months ago
- [EMNLP 2024] DA-Code: Agent Data Science Code Generation Benchmark for Large Language Models☆83Updated 4 months ago
- Official code of paper "Beyond 'Aha!': Toward Systematic Meta-Abilities Alignment in Large Reasoning Models"☆83Updated 6 months ago
- ☆54Updated last year
- ML-Bench: Evaluating Large Language Models and Agents for Machine Learning Tasks on Repository-Level Code (https://arxiv.org/abs/2311.098…☆305Updated 4 months ago
- We leverage 14 datasets as OOD test data and conduct evaluations on 8 NLU tasks over 21 popularly used models. Our findings confirm that …☆93Updated 2 years ago
- [ACL 2024] User-friendly evaluation framework: Eval Suite & Benchmarks: UHGEval, HaluEval, HalluQA, etc.☆178Updated 6 months ago
- ☆251Updated 7 months ago
- The official repo for paper, LLMs-as-Judges: A Comprehensive Survey on LLM-based Evaluation Methods.☆500Updated 4 months ago
- This is the official code repository of MoTCoder: Elevating Large Language Models with Modular of Thought for Challenging Programming Tas…☆85Updated 8 months ago
- LLM Benchmark for Code☆32Updated last year
- Source code and utilities for the Genesys distributed language model architecture discovery system.☆150Updated 2 months ago
- (NeurIPS D&B 2024) STaRK: Benchmarking LLM Retrieval on Textual and Relational Knowledge Bases☆323Updated last month
- Official implementation of RARE: Retrieval-Augmented Reasoning Modeling☆185Updated 6 months ago
- An Extensible Framework for Retrieval-Augmented LLM Applications: Learning Relevance Beyond Simple Similarity.☆41Updated last year