zwhe99 / X-SIRLinks
[ACL 2024] Can Watermarks Survive Translation? On the Cross-lingual Consistency of Text Watermark for Large Language Models
☆38Updated last year
Alternatives and similar repositories for X-SIR
Users that are interested in X-SIR are comparing it to the libraries listed below
Sorting:
- Recent papers on (1) Psychology of LLMs; (2) Biases in LLMs.☆49Updated last year
- Code & Data for our Paper "Alleviating Hallucinations of Large Language Models through Induced Hallucinations"☆65Updated last year
- ☆19Updated 11 months ago
- 🩺 A collection of ChatGPT evaluation reports on various bechmarks.☆49Updated 2 years ago
- Do Large Language Models Know What They Don’t Know?☆96Updated 7 months ago
- Feeling confused about super alignment? Here is a reading list☆42Updated last year
- [ACL 2023] Code and Data Repo for Paper "Element-aware Summary and Summary Chain-of-Thought (SumCoT)"☆53Updated last year
- [ICLR'24 spotlight] Tool-Augmented Reward Modeling☆50Updated 3 weeks ago
- [ICML'2024] Can AI Assistants Know What They Don't Know?☆81Updated last year
- 🍼 Official implementation of Dynamic Data Mixing Maximizes Instruction Tuning for Mixture-of-Experts☆39Updated 8 months ago
- Semi-Parametric Editing with a Retrieval-Augmented Counterfactual Model☆68Updated 2 years ago
- ☆22Updated 11 months ago
- ☆22Updated last month
- [ACL 2024 Findings] CriticBench: Benchmarking LLMs for Critique-Correct Reasoning☆26Updated last year
- Evaluating the Ripple Effects of Knowledge Editing in Language Models☆55Updated last year
- Logiqa2.0 dataset - logical reasoning in MRC and NLI tasks☆92Updated last year
- ☆26Updated 9 months ago
- Mostly recording papers about models' trustworthy applications. Intending to include topics like model evaluation & analysis, security, c…☆21Updated 2 years ago
- self-adaptive in-context learning☆45Updated 2 years ago
- [NeurIPS'24] Weak-to-Strong Search: Align Large Language Models via Searching over Small Language Models☆61Updated 6 months ago
- Implementation of "Investigating the Factual Knowledge Boundary of Large Language Models with Retrieval Augmentation"☆80Updated last year
- ☆44Updated last year
- Repository for the paper "Cognitive Mirage: A Review of Hallucinations in Large Language Models"☆47Updated last year
- [ACL 2023] Solving Math Word Problems via Cooperative Reasoning induced Language Models (LLMs + MCTS + Self-Improvement)☆49Updated last year
- [ACL 2024] Code for the paper "ALaRM: Align Language Models via Hierarchical Rewards Modeling"☆25Updated last year
- R-Judge: Benchmarking Safety Risk Awareness for LLM Agents (EMNLP Findings 2024)☆77Updated last month
- A Survey on the Honesty of Large Language Models☆57Updated 6 months ago
- [ACL 2024] An Easy-to-use Hallucination Detection Framework for LLMs.☆34Updated 4 months ago
- Our research proposes a novel MoGU framework that improves LLMs' safety while preserving their usability.☆15Updated 5 months ago
- [EMNLP 2024] The official GitHub repo for the survey paper "Knowledge Conflicts for LLMs: A Survey"☆126Updated 9 months ago