LeonDiao0427 / SEASLinks
We release our code and data for SEAS in this repository.
☆19Updated last year
Alternatives and similar repositories for SEAS
Users that are interested in SEAS are comparing it to the libraries listed below
Sorting:
- The reinforcement learning codes for dataset SPA-VL☆43Updated last year
- ☆23Updated last year
- ☆55Updated last year
- ☆36Updated last year
- 【ACL 2024】 SALAD benchmark & MD-Judge☆169Updated 10 months ago
- Safety-J: Evaluating Safety with Critique☆16Updated last year
- [ACL2025 Best Paper] Language Models Resist Alignment☆40Updated 6 months ago
- A comprehensive collection of process reward models.☆130Updated 3 months ago
- [2025-TMLR] A Survey on the Honesty of Large Language Models☆63Updated last year
- Repository for Label Words are Anchors: An Information Flow Perspective for Understanding In-Context Learning☆168Updated last year
- Implementation for the research paper "Enhancing LLM Reasoning via Critique Models with Test-Time and Training-Time Supervision".☆55Updated last year
- ☆51Updated 11 months ago
- R-Judge: Benchmarking Safety Risk Awareness for LLM Agents (EMNLP Findings 2024)☆95Updated this week
- Multilingual safety benchmark for Large Language Models☆54Updated last year
- [NeurIPS 2024 Oral] Aligner: Efficient Alignment by Learning to Correct☆191Updated 11 months ago
- [ACL' 25] The official code repository for PRMBench: A Fine-grained and Challenging Benchmark for Process-Level Reward Models.☆85Updated 10 months ago
- Official code for paper "SPA-RL: Reinforcing LLM Agent via Stepwise Progress Attribution"☆61Updated 3 months ago
- 🔍 Awesome Agentic Search is a curated list of papers, tools, and resources on agentic search—where AI agents plan, search, and reason to…☆50Updated 4 months ago
- [EMNLP 2024] The official GitHub repo for the paper "Course-Correction: Safety Alignment Using Synthetic Preferences"☆19Updated last year
- ☆47Updated 9 months ago
- Awesome-Long2short-on-LRMs is a collection of state-of-the-art, novel, exciting long2short methods on large reasoning models. It contains…☆255Updated 4 months ago
- This is the repository of DEER, a Dynamic Early Exit in Reasoning method for Large Reasoning Language Models.☆176Updated 6 months ago
- Flames is a highly adversarial benchmark in Chinese for LLM's harmlessness evaluation developed by Shanghai AI Lab and Fudan NLP Group.☆62Updated last year
- [COLM'25] Missing Premise exacerbates Overthinking: Are Reasoning Models losing Critical Thinking Skill?☆36Updated 7 months ago
- ☆70Updated 6 months ago
- ☆44Updated 6 months ago
- Awesome Large Reasoning Model(LRM) Safety.This repository is used to collect security-related research on large reasoning models such as …☆79Updated this week
- ☆161Updated 2 months ago
- Official repository for "Safety in Large Reasoning Models: A Survey" - Exploring safety risks, attacks, and defenses for Large Reasoning …☆83Updated 4 months ago
- This is the official GitHub repository for our survey paper "Beyond Single-Turn: A Survey on Multi-Turn Interactions with Large Language …☆161Updated 7 months ago