llm-as-a-judge / Awesome-LLM-as-a-judgeLinks
☆388Updated last week
Alternatives and similar repositories for Awesome-LLM-as-a-judge
Users that are interested in Awesome-LLM-as-a-judge are comparing it to the libraries listed below
Sorting:
- ☆591Updated last week
- LLM hallucination paper list☆320Updated last year
- This is the repository for the Tool Learning survey.☆418Updated 2 months ago
- A Survey on Data Selection for Language Models☆245Updated 3 months ago
- Survey on LLM Agents (Published on CoLing 2025)☆358Updated 3 months ago
- Stop Overthinking: A Survey on Efficient Reasoning for Large Language Models☆547Updated this week
- Official implementation for the paper "DoLa: Decoding by Contrasting Layers Improves Factuality in Large Language Models"☆504Updated 6 months ago
- A series of technical report on Slow Thinking with LLM☆713Updated last month
- The repository for the survey paper <<Survey on Large Language Models Factuality: Knowledge, Retrieval and Domain-Specificity>>☆341Updated last year
- ☆128Updated 4 months ago
- A curated list of LLM Interpretability related material - Tutorial, Library, Survey, Paper, Blog, etc..☆262Updated 4 months ago
- [ICML 2024] LESS: Selecting Influential Data for Targeted Instruction Tuning☆474Updated 9 months ago
- This is a collection of research papers for Self-Correcting Large Language Models with Automated Feedback.☆542Updated 9 months ago
- [ACL 2024] A Survey of Chain of Thought Reasoning: Advances, Frontiers and Future☆457Updated 6 months ago
- Collection of training data management explorations for large language models☆329Updated last year
- 😎 A Survey of Efficient Reasoning for Large Reasoning Models: Language, Multimodality, and Beyond☆277Updated last month
- ☆151Updated 10 months ago
- Latest Advances on Long Chain-of-Thought Reasoning☆459Updated 2 weeks ago
- [EMNLP 2024] The official GitHub repo for the survey paper "Knowledge Conflicts for LLMs: A Survey"☆129Updated 10 months ago
- Code for Parametric RAG, SIGIR 2025 Full Paper☆182Updated 3 months ago
- L1: Controlling How Long A Reasoning Model Thinks With Reinforcement Learning☆234Updated 2 months ago
- This is the repository of HaluEval, a large-scale hallucination evaluation benchmark for Large Language Models.☆497Updated last year
- A Survey of Attributions for Large Language Models☆207Updated 11 months ago
- Generative AI Act II: Test Time Scaling Drives Cognition Engineering☆198Updated 3 months ago
- [EMNLP 2024 (Oral)] Leave No Document Behind: Benchmarking Long-Context LLMs with Extended Multi-Doc QA☆139Updated 8 months ago
- A live reading list for LLM-synthetic-data.☆343Updated last week
- Awesome Agent Training☆204Updated last week
- A simple toolkit for benchmarking LLMs on mathematical reasoning tasks. 🧮✨☆239Updated last year
- An up-to-date curated list of Retrieval-Augmented Generation (RAG) for LLMs.☆130Updated this week
- Codebase for reproducing the experiments of the semantic uncertainty paper (short-phrase and sentence-length experiments).☆349Updated last year