wangzx1219 / AgentDropoutLinks
☆29Updated 2 months ago
Alternatives and similar repositories for AgentDropout
Users that are interested in AgentDropout are comparing it to the libraries listed below
Sorting:
- ☆60Updated 2 months ago
- ☆35Updated 5 months ago
- [EMNLP 2024] The official GitHub repo for the survey paper "Knowledge Conflicts for LLMs: A Survey"☆123Updated 8 months ago
- ICML 2025 Spotlight☆68Updated last month
- ☆59Updated last week
- [NeurIPS 2024] The official implementation of paper: Chain of Preference Optimization: Improving Chain-of-Thought Reasoning in LLMs.☆123Updated 2 months ago
- ☆24Updated last month
- This is the official implementation of the paper "S²R: Teaching LLMs to Self-verify and Self-correct via Reinforcement Learning"☆64Updated last month
- ☆32Updated 7 months ago
- [ACL'24] Chain of Thought (CoT) is significant in improving the reasoning abilities of large language models (LLMs). However, the correla…☆46Updated 3 weeks ago
- Implementation for the research paper "Enhancing LLM Reasoning via Critique Models with Test-Time and Training-Time Supervision".☆54Updated 6 months ago
- Implementation of the MATRIX framework (ICML 2024)☆53Updated last year
- ☆32Updated last week
- A Framework for LLM-based Multi-Agent Reinforced Training and Inference☆89Updated this week
- [NeurIPS 2024] Official code of $\beta$-DPO: Direct Preference Optimization with Dynamic $\beta$☆45Updated 7 months ago
- Watch Every Step! LLM Agent Learning via Iterative Step-level Process Refinement (EMNLP 2024 Main Conference)☆57Updated 7 months ago
- ☆22Updated 7 months ago
- A Survey of Personalization: From RAG to Agent☆40Updated last month
- [ICML'25] Multi-agent Architecture Search via Agentic Supernet☆58Updated last month
- [ACL'24] Beyond One-Preference-Fits-All Alignment: Multi-Objective Direct Preference Optimization☆79Updated 9 months ago
- Accepted LLM Papers in NeurIPS 2024☆37Updated 7 months ago
- [ICLR 2025] Benchmarking Agentic Workflow Generation☆94Updated 3 months ago
- ☆60Updated 2 weeks ago
- In-Context Sharpness as Alerts: An Inner Representation Perspective for Hallucination Mitigation (ICML 2024)☆59Updated last year
- Official repository for 'Safety Challenges in Large Reasoning Models: A Survey' - Exploring safety risks, attacks, and defenses for Large…☆36Updated last week
- MARFT stands for Multi-Agent Reinforcement Fine-Tuning. This repository implements an LLM-based multi-agent reinforcement fine-tuning fra…☆35Updated 2 weeks ago
- ☆70Updated this week
- CoT-Valve: Length-Compressible Chain-of-Thought Tuning☆69Updated 3 months ago
- this is an implementation for the paper Improve Mathematical Reasoning in Language Models by Automated Process Supervision from google de…☆32Updated 2 months ago
- Official Implementation of Dynamic LLM-Agent Network: An LLM-agent Collaboration Framework with Agent Team Optimization☆146Updated last year