zai-org / ComplexFuncBenchLinks
Complex Function Calling Benchmark.
☆135Updated 8 months ago
Alternatives and similar repositories for ComplexFuncBench
Users that are interested in ComplexFuncBench are comparing it to the libraries listed below
Sorting:
- Official repository for paper "ReasonIR Training Retrievers for Reasoning Tasks".☆202Updated 3 months ago
- LongEmbed: Extending Embedding Models for Long Context Retrieval (EMNLP 2024)☆143Updated 10 months ago
- Benchmarking LLMs with Challenging Tasks from Real Users☆241Updated 11 months ago
- ☆127Updated last year
- LOFT: A 1 Million+ Token Long-Context Benchmark☆212Updated 3 months ago
- EvolKit is an innovative framework designed to automatically enhance the complexity of instructions used for fine-tuning Large Language M…☆240Updated 11 months ago
- 🚢 Data Toolkit for Sailor Language Models☆94Updated 7 months ago
- The official evaluation suite and dynamic data release for MixEval.☆246Updated 10 months ago
- A framework to study AI models in Reasoning, Alignment, and use of Memory (RAM).☆287Updated last week
- Code repo for "Agent Instructs Large Language Models to be General Zero-Shot Reasoners"☆116Updated last year
- Official code for "MAmmoTH2: Scaling Instructions from the Web" [NeurIPS 2024]☆148Updated 11 months ago
- Official repository for "Scaling Retrieval-Based Langauge Models with a Trillion-Token Datastore".☆216Updated 2 months ago
- Reproducible, flexible LLM evaluations☆251Updated 2 months ago
- ☆211Updated last year
- Tina: Tiny Reasoning Models via LoRA☆284Updated last week
- A dataset for training and evaluating LLMs on decision making about "when (not) to call" functions☆38Updated 5 months ago
- ☆99Updated 10 months ago
- General Reasoner: Advancing LLM Reasoning Across All Domains [NeurIPS25]☆172Updated 3 months ago
- Verifiers for LLM Reinforcement Learning☆74Updated 5 months ago
- BrowseComp-Plus: A More Fair and Transparent Evaluation Benchmark of Deep-Research Agent☆83Updated last month
- [NeurIPS 2024] Spider2-V: How Far Are Multimodal Agents From Automating Data Science and Engineering Workflows?☆131Updated last year
- ☆143Updated 6 months ago
- Reformatted Alignment☆113Updated last year
- ☆91Updated 4 months ago
- Code for "Critique Fine-Tuning: Learning to Critique is More Effective than Learning to Imitate" [COLM 2025]☆172Updated 2 months ago
- BABILong is a benchmark for LLM evaluation using the needle-in-a-haystack approach.☆212Updated last month
- ☆118Updated 4 months ago
- [ACL 2025] Agentic Reward Modeling: Integrating Human Preferences with Verifiable Correctness Signals for Reliable Reward Systems☆106Updated 3 months ago
- ☆155Updated last year
- Implementation of the LongRoPE: Extending LLM Context Window Beyond 2 Million Tokens Paper☆149Updated last year