ADaM-BJTU / W2SG
The code of “Improving Weak-to-Strong Generalization with Scalable Oversight and Ensemble Learning”
☆15Updated 11 months ago
Alternatives and similar repositories for W2SG:
Users that are interested in W2SG are comparing it to the libraries listed below
- BeHonest: Benchmarking Honesty in Large Language Models☆31Updated 5 months ago
- GSM-Plus: Data, Code, and Evaluation for Enhancing Robust Mathematical Reasoning in Math Word Problems.☆55Updated 6 months ago
- ☆29Updated 8 months ago
- [NAACL'25] Steering Knowledge Selection Behaviours in LLMs via SAE-Based Representation Engineering☆41Updated 2 months ago
- Code & Data for our Paper "Alleviating Hallucinations of Large Language Models through Induced Hallucinations"☆62Updated 11 months ago
- ☆36Updated last year
- [AAAI 2025 oral] Evaluating Mathematical Reasoning Beyond Accuracy☆44Updated last month
- ☆72Updated 8 months ago
- ☆13Updated 6 months ago
- The official implementation of "ICDPO: Effectively Borrowing Alignment Capability of Others via In-context Direct Preference Optimization…☆14Updated 11 months ago
- ☆30Updated 9 months ago
- Watch Every Step! LLM Agent Learning via Iterative Step-level Process Refinement (EMNLP 2024 Main Conference)☆51Updated 3 months ago
- The repository of the project "Fine-tuning Large Language Models with Sequential Instructions", code base comes from open-instruct and LA…☆29Updated 2 months ago
- Evaluating the Ripple Effects of Knowledge Editing in Language Models☆53Updated 9 months ago
- ☆38Updated last year
- Source codes for "Preference-grounded Token-level Guidance for Language Model Fine-tuning" (NeurIPS 2023).☆15Updated 3 weeks ago
- In-Context Sharpness as Alerts: An Inner Representation Perspective for Hallucination Mitigation (ICML 2024)☆50Updated 10 months ago
- ☆25Updated last year
- Long Is More for Alignment: A Simple but Tough-to-Beat Baseline for Instruction Fine-Tuning [ICML 2024]☆17Updated 8 months ago
- Official code for ICML 2024 paper on Persona In-Context Learning (PICLe)☆23Updated 7 months ago
- Official code for "Decoding-Time Language Model Alignment with Multiple Objectives".☆18Updated 3 months ago
- The official repository of "Improving Large Language Models via Fine-grained Reinforcement Learning with Minimum Editing Constraint"☆34Updated last year
- AbstainQA, ACL 2024☆25Updated 3 months ago
- Evaluate the Quality of Critique☆35Updated 7 months ago
- [AAAI 2024] MELO: Enhancing Model Editing with Neuron-indexed Dynamic LoRA☆24Updated 9 months ago
- ☆12Updated 6 months ago
- Safety-J: Evaluating Safety with Critique☆16Updated 6 months ago
- Analyzing LLM Alignment via Token distribution shift☆15Updated last year
- ☆26Updated last month
- This is an official implementation of the Reward rAnked Fine-Tuning Algorithm (RAFT), also known as iterative best-of-n fine-tuning or re…☆22Updated 4 months ago