shiqichen17 / SPALinks
Github repository for "Internalizing World Models via Self-Play Finetuning for Agentic RL"
☆33Updated 3 months ago
Alternatives and similar repositories for SPA
Users that are interested in SPA are comparing it to the libraries listed below
Sorting:
- ☆29Updated last year
- ☆46Updated 2 years ago
- ☆41Updated 2 years ago
- Code for Representation Bending Paper☆16Updated 6 months ago
- Official implementation of Bootstrapping Language Models via DPO Implicit Rewards☆47Updated 9 months ago
- A Mechanistic Understanding of Alignment Algorithms: A Case Study on DPO and Toxicity.☆85Updated 11 months ago
- [ACL 2024] Code and data for "Machine Unlearning of Pre-trained Large Language Models"☆66Updated last year
- [COLM 2025] SEAL: Steerable Reasoning Calibration of Large Language Models for Free☆51Updated 10 months ago
- [ICLR'25 Spotlight] Min-K%++: Improved baseline for detecting pre-training data of LLMs☆52Updated 8 months ago
- ☆53Updated 10 months ago
- ☆44Updated last year
- ☆72Updated last year
- ☆51Updated 2 years ago
- Test-time-training on nearest neighbors for large language models☆49Updated last year
- Restore safety in fine-tuned language models through task arithmetic☆31Updated last year
- In-Context Sharpness as Alerts: An Inner Representation Perspective for Hallucination Mitigation (ICML 2024)☆62Updated last year
- Röttger et al. (NAACL 2024): "XSTest: A Test Suite for Identifying Exaggerated Safety Behaviours in Large Language Models"☆127Updated 11 months ago
- Code for safety test in "Keeping LLMs Aligned After Fine-tuning: The Crucial Role of Prompt Templates"☆22Updated 4 months ago
- RENT (Reinforcement Learning via Entropy Minimization) is an unsupervised method for training reasoning LLMs.☆41Updated 3 months ago
- Code associated with Tuning Language Models by Proxy (Liu et al., 2024)☆127Updated last year
- ☆69Updated 11 months ago
- [ICLR'24] RAIN: Your Language Models Can Align Themselves without Finetuning☆98Updated last year
- [EMNLP 2025 Main] ConceptVectors Benchmark and Code for the paper "Intrinsic Evaluation of Unlearning Using Parametric Knowledge Traces"☆39Updated 5 months ago
- Function Vectors in Large Language Models (ICLR 2024)☆191Updated 9 months ago
- RWKU: Benchmarking Real-World Knowledge Unlearning for Large Language Models. NeurIPS 2024☆89Updated last year
- ☆37Updated 2 years ago
- This is an official implementation of the Reward rAnked Fine-Tuning Algorithm (RAFT), also known as iterative best-of-n fine-tuning or re…☆39Updated last year
- Long Is More for Alignment: A Simple but Tough-to-Beat Baseline for Instruction Fine-Tuning [ICML 2024]☆21Updated last year
- Is In-Context Learning Sufficient for Instruction Following in LLMs? [ICLR 2025]☆32Updated last year
- [NAACL'25 Oral] Steering Knowledge Selection Behaviours in LLMs via SAE-Based Representation Engineering☆72Updated 3 weeks ago