shivamag125 / EM_PTLinks
☆14Updated 2 months ago
Alternatives and similar repositories for EM_PT
Users that are interested in EM_PT are comparing it to the libraries listed below
Sorting:
- [ACL 2025] Knowledge Unlearning for Large Language Models☆39Updated 2 months ago
- The repository of the paper "REEF: Representation Encoding Fingerprints for Large Language Models," aims to protect the IP of open-source…☆58Updated 6 months ago
- ☆19Updated 5 months ago
- The official implementation of "LightTransfer: Your Long-Context LLM is Secretly a Hybrid Model with Effortless Adaptation"☆20Updated 3 months ago
- Official repository for paper: O1-Pruner: Length-Harmonizing Fine-Tuning for O1-Like Reasoning Pruning☆85Updated 5 months ago
- ☆15Updated last month
- Source code for our paper: "ARIA: Training Language Agents with Intention-Driven Reward Aggregation".☆19Updated last week
- ☆27Updated 3 months ago
- CoT-Valve: Length-Compressible Chain-of-Thought Tuning☆77Updated 5 months ago
- ☆37Updated 3 months ago
- ☆15Updated 7 months ago
- ☆26Updated 3 months ago
- [ACL'25 Oral] What Happened in LLMs Layers when Trained for Fast vs. Slow Thinking: A Gradient Perspective☆70Updated last month
- Code for "CREAM: Consistency Regularized Self-Rewarding Language Models", ICLR 2025.☆22Updated 5 months ago
- SG-Bench: Evaluating LLM Safety Generalization Across Diverse Tasks and Prompt Types☆20Updated 7 months ago
- [ICLR 2025 Workshop] "Landscape of Thoughts: Visualizing the Reasoning Process of Large Language Models"☆31Updated 3 weeks ago
- ☆126Updated 2 months ago
- The official repository of paper "AdaR1: From Long-Cot to Hybrid-CoT via Bi-Level Adaptive Reasoning Optimization"☆18Updated 2 months ago
- ☆43Updated last week
- FeatureAlignment = Alignment + Mechanistic Interpretability☆29Updated 4 months ago
- Official code for SEAL: Steerable Reasoning Calibration of Large Language Models for Free☆30Updated 3 months ago
- [ACL 2025 (Findings)] DEMO: Reframing Dialogue Interaction with Fine-grained Element Modeling☆15Updated 7 months ago
- Optimizing Anytime Reasoning via Budget Relative Policy Optimization☆43Updated last week
- ☆32Updated 3 months ago
- [ICLR 2025] Official codebase for the ICLR 2025 paper "Multimodal Situational Safety"☆21Updated last month
- To Think or Not to Think: Exploring the Unthinking Vulnerability in Large Reasoning Models☆31Updated 2 months ago
- A holistic benchmark for LLM abstention☆40Updated 2 weeks ago
- Test-time preferenece optimization (ICML 2025).☆148Updated 2 months ago
- A Sober Look at Language Model Reasoning☆79Updated last month
- [ACL 2024] Code and data for "Machine Unlearning of Pre-trained Large Language Models"☆59Updated 9 months ago