SCIR-SC-Qiaoban-Team / FreeEvalLMLinks
☆10Updated 9 months ago
Alternatives and similar repositories for FreeEvalLM
Users that are interested in FreeEvalLM are comparing it to the libraries listed below
Sorting:
- [COLM 2025] SEAL: Steerable Reasoning Calibration of Large Language Models for Free☆52Updated 9 months ago
- Resources and paper list for 'Scaling Environments for Agents'. This repository accompanies our survey on how environments contribute to …☆56Updated last month
- [ACL'24] Beyond One-Preference-Fits-All Alignment: Multi-Objective Direct Preference Optimization☆94Updated last year
- This is the official code for the paper "Vaccine: Perturbation-aware Alignment for Large Language Models" (NeurIPS2024)☆49Updated last week
- ☆44Updated last year
- [ACL 25] SafeChain: Safety of Language Models with Long Chain-of-Thought Reasoning Capabilities☆27Updated 9 months ago
- AdaRFT: Efficient Reinforcement Finetuning via Adaptive Curriculum Learning☆53Updated 7 months ago
- Official implementation for "ALI-Agent: Assessing LLMs'Alignment with Human Values via Agent-based Evaluation"☆21Updated 5 months ago
- ☆30Updated last year
- This repo is for the safety topic, including attacks, defenses and studies related to reasoning and RL☆59Updated 4 months ago
- [ACL2025 Best Paper] Language Models Resist Alignment☆41Updated 7 months ago
- ☆23Updated 11 months ago
- [ICML 2024] Assessing the Brittleness of Safety Alignment via Pruning and Low-Rank Modifications☆89Updated 9 months ago
- ☆56Updated 3 months ago
- A Sober Look at Language Model Reasoning☆92Updated 2 months ago
- Can Knowledge Editing Really Correct Hallucinations? (ICLR 2025)☆27Updated 5 months ago
- ☆73Updated 9 months ago
- [NeurIPS 2024] Official code of $\beta$-DPO: Direct Preference Optimization with Dynamic $\beta$☆50Updated last year
- ☆37Updated 2 years ago
- [NeurIPS 2024] "Can Language Models Perform Robust Reasoning in Chain-of-thought Prompting with Noisy Rationales?"☆38Updated 6 months ago
- ☆62Updated 6 months ago
- [ICLR 2025] Code and Data Repo for Paper "Latent Space Chain-of-Embedding Enables Output-free LLM Self-Evaluation"☆93Updated last year
- ☆51Updated last year
- [ICLR 25 Oral] RM-Bench: Benchmarking Reward Models of Language Models with Subtlety and Style☆73Updated 6 months ago
- [COLM'25] Missing Premise exacerbates Overthinking: Are Reasoning Models losing Critical Thinking Skill?☆36Updated 7 months ago
- Code for "CREAM: Consistency Regularized Self-Rewarding Language Models", ICLR 2025.☆28Updated 11 months ago
- Model merging is a highly efficient approach for long-to-short reasoning.☆98Updated 3 months ago
- In-Context Sharpness as Alerts: An Inner Representation Perspective for Hallucination Mitigation (ICML 2024)☆62Updated last year
- This is the official implementation of the paper "S²R: Teaching LLMs to Self-verify and Self-correct via Reinforcement Learning"☆73Updated 9 months ago
- [ICML 2025] "From Passive to Active Reasoning: Can Large Language Models Ask the Right Questions under Incomplete Information?"☆49Updated 3 months ago