hy-zhao23 / Explainability-for-Large-Language-Models
β94Updated 8 months ago
Related projects: β
- A reading list on LLM based Synthetic Data Generation π₯β105Updated last month
- A curated list of LLM Interpretability related material - Tutorial, Library, Survey, Paper, Blog, etc..β109Updated last week
- A Survey on Data Selection for Language Modelsβ148Updated 3 months ago
- A curated reading list for large language model (LLM) alignment. Take a look at our new survey "Large Language Model Alignment: A Survey"β¦β65Updated 11 months ago
- LLM Unlearningβ112Updated 11 months ago
- A Survey of Attributions for Large Language Modelsβ155Updated 3 weeks ago
- Project for the paper entitled `Instruction Tuning for Large Language Models: A Survey`β134Updated 6 months ago
- Official github repo for AutoDetect, an automated weakness detection framework for LLMs.β36Updated 2 months ago
- BeaverTails is a collection of datasets designed to facilitate research on safety alignment in large language models (LLMs).β99Updated 10 months ago
- This repository provides an original implementation of Detecting Pretraining Data from Large Language Models by *Weijia Shi, *Anirudh Ajiβ¦β198Updated 10 months ago
- Model Merging in LLMs, MLLMs, and Beyond: Methods, Theories, Applications and Opportunities. arXiv:2408.07666.β107Updated this week
- The repository for the survey paper <<Survey on Large Language Models Factuality: Knowledge, Retrieval and Domain-Specificity>>β323Updated 4 months ago
- The official GitHub repo for the survey paper "Knowledge Conflicts for LLMs: A Survey"β62Updated 3 weeks ago
- LLM hallucination paper listβ268Updated 6 months ago
- Official github repo for SafetyBench, a comprehensive benchmark to evaluate LLMs' safety.β141Updated 2 months ago
- β170Updated last month
- [ICLR 2024] Evaluating Large Language Models at Evaluating Instruction Followingβ104Updated 2 months ago
- R-Judge: Benchmarking Safety Risk Awareness for LLM Agentsβ57Updated last month
- π An unofficial implementation of Self-Alignment with Instruction Backtranslation.β128Updated 2 months ago
- Code associated with Tuning Language Models by Proxy (Liu et al., 2024)β84Updated 5 months ago
- [ACL'24] Chain of Thought (CoT) is significant in improving the reasoning abilities of large language models (LLMs). However, the correlaβ¦β28Updated last month
- EMNLP'23 survey: a curation of awesome papers and resources on refreshing large language models (LLMs) without expensive retraining.β121Updated 9 months ago
- [NAACL 2024 Outstanding Paper] Source code for the NAACL 2024 paper entitled "R-Tuning: Instructing Large Language Models to Say 'I Don'tβ¦β82Updated 2 months ago
- Official Implementation of Dynamic LLM-Agent Network: An LLM-agent Collaboration Framework with Agent Team Optimizationβ96Updated 4 months ago
- Repository for the paper "Cognitive Mirage: A Review of Hallucinations in Large Language Models"β47Updated 10 months ago
- Must-read Papers on Large Language Model (LLM) Continual Learningβ129Updated 10 months ago
- γACL 2024γ SALAD benchmark & MD-Judgeβ81Updated this week
- Do Large Language Models Know What They Donβt Know?β84Updated 9 months ago
- [ACL 2024] A Survey of Chain of Thought Reasoning: Advances, Frontiers and Futureβ279Updated 2 months ago
- Collection of training data management explorations for large language modelsβ266Updated last month