shizhediao / R-TuningView external linksLinks
[NAACL 2024 Outstanding Paper] Source code for the NAACL 2024 paper entitled "R-Tuning: Instructing Large Language Models to Say 'I Don't Know'"
☆129Jul 10, 2024Updated last year
Alternatives and similar repositories for R-Tuning
Users that are interested in R-Tuning are comparing it to the libraries listed below
Sorting:
- [ICML'2024] Can AI Assistants Know What They Don't Know?☆85Feb 5, 2024Updated 2 years ago
- Code & Data for our Paper "Alleviating Hallucinations of Large Language Models through Induced Hallucinations"☆69Feb 27, 2024Updated last year
- [ACL 2024] LLM2LLM: Boosting LLMs with Novel Iterative Data Enhancement☆193Mar 25, 2024Updated last year
- ☆17Nov 3, 2024Updated last year
- Source code for our paper: "ARIA: Training Language Agents with Intention-Driven Reward Aggregation".☆26Aug 9, 2025Updated 6 months ago
- Code for ACL 2023 paper "A Close Look into the Calibration of Pre-trained Language Models"☆11May 9, 2023Updated 2 years ago
- [ACL'24] Superfiltering: Weak-to-Strong Data Filtering for Fast Instruction-Tuning☆188Jun 25, 2025Updated 7 months ago
- Deita: Data-Efficient Instruction Tuning for Alignment [ICLR2024]☆588Dec 9, 2024Updated last year
- ☆18Feb 20, 2024Updated last year
- CFT-RAG: An Entity Tree Based Retrieval Augmented Generation Algorithm With Cuckoo Filter☆22May 28, 2025Updated 8 months ago
- Repo for Rho-1: Token-level Data Selection & Selective Pretraining of LLMs.☆459Apr 18, 2024Updated last year
- [EMNLP 2024 Findings] Benchmarking Language Model Agents for Data-Driven Science☆34Oct 25, 2024Updated last year
- Control LLM☆22Apr 6, 2025Updated 10 months ago
- 🌟Official code of our AAAI26 paper 🔍WebFilter☆35Nov 9, 2025Updated 3 months ago
- ☆24Oct 14, 2024Updated last year
- [NAACL'24] Self-data filtering of LLM instruction-tuning data using a novel perplexity-based difficulty score, without using any other mo…☆416Jun 25, 2025Updated 7 months ago
- Official code repository for the paper "ToMAP: Training Opponent-Aware LLM Persuaders with Theory of Mind"☆22Sep 25, 2025Updated 4 months ago
- ☆30Aug 21, 2025Updated 5 months ago
- Official implementation for the paper "DoLa: Decoding by Contrasting Layers Improves Factuality in Large Language Models"☆537Jan 17, 2025Updated last year
- Code accompanying the paper "Noise Contrastive Alignment of Language Models with Explicit Rewards" (NeurIPS 2024)☆58Nov 8, 2024Updated last year
- Codebase for Instruction Following without Instruction Tuning☆36Sep 24, 2024Updated last year
- [ACL 2024] Learning to Edit: Aligning LLMs with Knowledge Editing☆36Aug 19, 2024Updated last year
- ☆29Dec 28, 2025Updated last month
- Official implementation of the ACL 2023 paper: "Zero-shot Faithful Factual Error Correction"☆17Aug 14, 2023Updated 2 years ago
- Implementations of online merging optimizers proposed by Online Merging Optimizers for Boosting Rewards and Mitigating Tax in Alignment☆81Jun 19, 2024Updated last year
- Code for the paper "RADAR: Enhancing Radiology Report Generation with Supplementary Knowledge Injection" (ACL'25).☆33Jul 23, 2025Updated 6 months ago
- Do Large Language Models Know What They Don’t Know?☆102Nov 8, 2024Updated last year
- Language models scale reliably with over-training and on downstream tasks☆99Apr 2, 2024Updated last year
- ☆78May 22, 2024Updated last year
- Implementation of MixCE method described in ACL 2023 paper by Zhang et al.☆20May 29, 2023Updated 2 years ago
- Public code repo for paper "SaySelf: Teaching LLMs to Express Confidence with Self-Reflective Rationales"☆112Sep 28, 2024Updated last year
- SELF-GUIDE: Better Task-Specific Instruction Following via Self-Synthetic Finetuning. COLM 2024 Accepted Paper☆32May 29, 2024Updated last year
- Preference Learning for LLaVA☆59Nov 9, 2024Updated last year
- [EMNLP'23] Code for "Non-autoregressive Text Editing with Copy-aware Latent Alignments".☆20Oct 17, 2023Updated 2 years ago
- ☆554Jan 2, 2025Updated last year
- [ACL'24] Beyond One-Preference-Fits-All Alignment: Multi-Objective Direct Preference Optimization☆96Aug 20, 2024Updated last year
- Exchange-of-Thought: Enhancing Large Language Model Capabilities through Cross-Model Communication☆21Mar 21, 2024Updated last year
- We aim to provide the best references to search, select, and synthesize high-quality and large-quantity data for post-training your LLMs.☆61Oct 3, 2024Updated last year
- This repository provides an original implementation of Detecting Pretraining Data from Large Language Models by *Weijia Shi, *Anirudh Aji…☆240Nov 3, 2023Updated 2 years ago