DataArcTech / LLM-as-a-JudgeLinks
☆157Updated last month
Alternatives and similar repositories for LLM-as-a-Judge
Users that are interested in LLM-as-a-Judge are comparing it to the libraries listed below
Sorting:
- [ICLR 2025] InstructRAG: Instructing Retrieval-Augmented Generation via Self-Synthesized Rationales☆132Updated 9 months ago
- [ACL'24] Selective Reflection-Tuning: Student-Selected Data Recycling for LLM Instruction-Tuning☆365Updated last year
- Github repository for "RAGTruth: A Hallucination Corpus for Developing Trustworthy Retrieval-Augmented Language Models"☆212Updated last year
- augmented LLM with self reflection☆135Updated 2 years ago
- [NAACL 2024 Outstanding Paper] Source code for the NAACL 2024 paper entitled "R-Tuning: Instructing Large Language Models to Say 'I Don't…☆125Updated last year
- ☆105Updated last year
- Code implementation of synthetic continued pretraining☆142Updated 10 months ago
- Official repository for paper "ReasonIR Training Retrievers for Reasoning Tasks".☆205Updated 5 months ago
- [ICLR 2025] BRIGHT: A Realistic and Challenging Benchmark for Reasoning-Intensive Retrieval☆179Updated 2 months ago
- [Neurips2024] Source code for xRAG: Extreme Context Compression for Retrieval-augmented Generation with One Token☆164Updated last year
- Project for the paper entitled `Instruction Tuning for Large Language Models: A Survey`☆203Updated 3 months ago
- Benchmarking LLMs with Challenging Tasks from Real Users☆246Updated last year
- ☆54Updated 3 weeks ago
- Code for "Critique Fine-Tuning: Learning to Critique is More Effective than Learning to Imitate" [COLM 2025]☆179Updated 4 months ago
- Awesome LLM Self-Consistency: a curated list of Self-consistency in Large Language Models☆113Updated 4 months ago
- A Survey on Data Selection for Language Models☆252Updated 7 months ago
- [EMNLP 2024 (Oral)] Leave No Document Behind: Benchmarking Long-Context LLMs with Extended Multi-Doc QA☆143Updated last year
- [NeurIPS 2024] The official implementation of paper: Chain of Preference Optimization: Improving Chain-of-Thought Reasoning in LLMs.☆132Updated 8 months ago
- Public code repo for paper "SaySelf: Teaching LLMs to Express Confidence with Self-Reflective Rationales"☆109Updated last year
- Framework and toolkits for building and evaluating collaborative agents that can work together with humans.☆110Updated last month
- Official implementation for the paper "DoLa: Decoding by Contrasting Layers Improves Factuality in Large Language Models"☆524Updated 10 months ago
- ☆241Updated last year
- Comprehensive benchmark for RAG☆244Updated 5 months ago
- Reproducible, flexible LLM evaluations☆286Updated 2 weeks ago
- Official code for "MAmmoTH2: Scaling Instructions from the Web" [NeurIPS 2024]☆149Updated last year
- A banchmark list for evaluation of large language models.☆151Updated 2 months ago
- Official Code Repository for the paper "Distilling LLM Agent into Small Models with Retrieval and Code Tools"☆175Updated last month
- Generative Judge for Evaluating Alignment☆248Updated last year
- LOFT: A 1 Million+ Token Long-Context Benchmark☆218Updated 5 months ago
- Critique-out-Loud Reward Models☆70Updated last year