tml-epfl / icl-alignment
Is In-Context Learning Sufficient for Instruction Following in LLMs?
☆25Updated 5 months ago
Related projects ⓘ
Alternatives and complementary repositories for icl-alignment
- ☆19Updated last year
- ☆36Updated 3 months ago
- ☆15Updated 4 months ago
- This is the oficial repository for "Safer-Instruct: Aligning Language Models with Automated Preference Data"☆17Updated 8 months ago
- Long Is More for Alignment: A Simple but Tough-to-Beat Baseline for Instruction Fine-Tuning [ICML 2024]☆15Updated 6 months ago
- Directional Preference Alignment☆50Updated last month
- Long Context Extension and Generalization in LLMs☆39Updated 2 months ago
- Repository for NPHardEval, a quantified-dynamic benchmark of LLMs☆48Updated 7 months ago
- Official Repository for Dataset Inference for LLMs☆23Updated 3 months ago
- ☆24Updated 4 months ago
- ☆44Updated 2 months ago
- ☆12Updated 2 months ago
- Data Valuation on In-Context Examples (ACL23)☆23Updated last month
- ☆14Updated 9 months ago
- Self-Supervised Alignment with Mutual Information☆14Updated 5 months ago
- This repository contains data, code and models for contextual noncompliance.☆18Updated 4 months ago
- Code for "Seeking Neural Nuggets: Knowledge Transfer in Large Language Models from a Parametric Perspective"☆30Updated 6 months ago
- ☆26Updated last year
- Codebase for Instruction Following without Instruction Tuning☆31Updated last month
- Data and code for the preprint "In-Context Learning with Long-Context Models: An In-Depth Exploration"☆31Updated 3 months ago
- [ACL 2024] Self-Training with Direct Preference Optimization Improves Chain-of-Thought Reasoning☆30Updated 3 months ago
- Code for "Universal Adversarial Triggers Are Not Universal."☆16Updated 6 months ago
- Restore safety in fine-tuned language models through task arithmetic☆26Updated 7 months ago
- ☆26Updated 8 months ago
- [EMNLP-2022 Findings] Code for paper “ProGen: Progressive Zero-shot Dataset Generation via In-context Feedback”.☆24Updated last year
- Official implementation of Bootstrapping Language Models via DPO Implicit Rewards☆39Updated 3 months ago
- Offcial Repo of Paper "Eliminating Position Bias of Language Models: A Mechanistic Approach""☆11Updated 2 months ago
- The repository contains code for Adaptive Data Optimization☆18Updated last month
- Code for the arXiv preprint "The Unreasonable Effectiveness of Easy Training Data"☆44Updated 10 months ago
- ☆32Updated last year