youdotcom-oss / ydc-deep-research-evalsLinks
you.com's framework for evaluating deep research systems.
☆57Updated 6 months ago
Alternatives and similar repositories for ydc-deep-research-evals
Users that are interested in ydc-deep-research-evals are comparing it to the libraries listed below
Sorting:
- Codebase accompanying the Summary of a Haystack paper.☆79Updated last year
- Official Repo for CRMArena and CRMArena-Pro☆125Updated 3 weeks ago
- ☆43Updated last year
- A method for steering llms to better follow instructions☆58Updated 3 months ago
- Matrix (Multi-Agent daTa geneRation Infra and eXperimentation framework) is a versatile engine for multi-agent conversational data genera…☆106Updated this week
- ☆58Updated last year
- ☆87Updated this week
- The first dense retrieval model that can be prompted like an LM☆89Updated 6 months ago
- Implementation of the paper: "AssistantBench: Can Web Agents Solve Realistic and Time-Consuming Tasks?"☆66Updated 11 months ago
- Leveraging Base Language Models for Few-Shot Synthetic Data Generation☆38Updated last month
- Source code for the collaborative reasoner research project at Meta FAIR.☆106Updated 7 months ago
- [EMNLP 2024] A Retrieval Benchmark for Scientific Literature Search☆101Updated 11 months ago
- Verifiers for LLM Reinforcement Learning☆79Updated 7 months ago
- Analysis code for Neurips 2025 paper "SciArena: An Open Evaluation Platform for Foundation Models in Scientific Literature Tasks"☆55Updated 3 months ago
- Code for our paper PAPILLON: PrivAcy Preservation from Internet-based and Local Language MOdel ENsembles☆60Updated 6 months ago
- A framework for standardizing evaluations of large foundation models, beyond single-score reporting and rankings.☆171Updated last week
- ReBase: Training Task Experts through Retrieval Based Distillation☆29Updated 9 months ago
- Functional Benchmarks and the Reasoning Gap☆90Updated last year
- Public code repo for paper "SaySelf: Teaching LLMs to Express Confidence with Self-Reflective Rationales"☆109Updated last year
- Code and data for "StructLM: Towards Building Generalist Models for Structured Knowledge Grounding" (COLM 2024)☆75Updated last year
- Learning to route instances for Human vs AI Feedback (ACL Main '25)☆26Updated 4 months ago
- Scalable Meta-Evaluation of LLMs as Evaluators☆43Updated last year
- ☆129Updated last year
- Mixing Language Models with Self-Verification and Meta-Verification☆110Updated 11 months ago
- Framework and toolkits for building and evaluating collaborative agents that can work together with humans.☆110Updated last month
- A framework for pitting LLMs against each other in an evolving library of games ⚔☆34Updated 7 months ago
- ☆23Updated 8 months ago
- Retrieval Augmented Generation Generalized Evaluation Dataset☆58Updated 4 months ago
- Code and Data for "Language Modeling with Editable External Knowledge"☆36Updated last year
- ☆62Updated 5 months ago