Hritikbansal / jpo
☆12Updated 3 months ago
Alternatives and similar repositories for jpo:
Users that are interested in jpo are comparing it to the libraries listed below
- ☆25Updated 2 years ago
- ☆15Updated last year
- ☆29Updated 11 months ago
- Lightweight Adapting for Black-Box Large Language Models☆22Updated last year
- Methods and evaluation for aligning language models temporally☆29Updated last year
- One Network, Many Masks: Towards More Parameter-Efficient Transfer Learning☆39Updated last year
- Code for "Inducer-tuning: Connecting Prefix-tuning and Adapter-tuning" (EMNLP 2022) and "Empowering Parameter-Efficient Transfer Learning…☆12Updated 2 years ago
- Directional Preference Alignment☆56Updated 6 months ago
- ☆41Updated last year
- Code for Paper (Preserving Diversity in Supervised Fine-tuning of Large Language Models)☆17Updated last week
- ☆16Updated 7 months ago
- Official implementation of Bootstrapping Language Models via DPO Implicit Rewards☆43Updated 8 months ago
- ☆10Updated 10 months ago
- SLED: Self Logits Evolution Decoding for Improving Factuality in Large Language Model https://arxiv.org/pdf/2411.02433☆25Updated 4 months ago
- Visual and Embodied Concepts evaluation benchmark☆21Updated last year
- [EMNLP-2022 Findings] Code for paper “ProGen: Progressive Zero-shot Dataset Generation via In-context Feedback”.☆26Updated 2 years ago
- Code for "Can Retriever-Augmented Language Models Reason? The Blame Game Between the Retriever and the Language Model", EMNLP Findings 20…☆28Updated last year
- Restore safety in fine-tuned language models through task arithmetic☆28Updated last year
- ☆39Updated last year
- The code and data for the paper JiuZhang3.0☆43Updated 10 months ago
- On the Effectiveness of Parameter-Efficient Fine-Tuning☆38Updated last year
- [NeurIPS 2024] Official code of $\beta$-DPO: Direct Preference Optimization with Dynamic $\beta$☆41Updated 5 months ago
- Code for paper "Unraveling Cross-Modality Knowledge Conflicts in Large Vision-Language Models."☆42Updated 5 months ago
- This is the oficial repository for "Parameter-Efficient Multi-task Tuning via Attentional Mixtures of Soft Prompts" (EMNLP 2022)☆100Updated 2 years ago
- code for EMNLP 2024 paper: How do Large Language Models Learn In-Context? Query and Key Matrices of In-Context Heads are Two Towers for M…☆11Updated 4 months ago
- [ICLR 2025] Code&Data for the paper "Super(ficial)-alignment: Strong Models May Deceive Weak Models in Weak-to-Strong Generalization"☆13Updated 9 months ago
- Source code for the paper "Prefix Language Models are Unified Modal Learners"☆43Updated last year
- ☆16Updated 8 months ago
- This is the official repo for Towards Uncertainty-Aware Language Agent.☆24Updated 8 months ago
- Extending context length of visual language models☆11Updated 3 months ago