voidism / Lookback-LensLinks
Code for the EMNLP 2024 paper "Detecting and Mitigating Contextual Hallucinations in Large Language Models Using Only Attention Maps"
☆142Updated 3 months ago
Alternatives and similar repositories for Lookback-Lens
Users that are interested in Lookback-Lens are comparing it to the libraries listed below
Sorting:
- Public code repo for paper "SaySelf: Teaching LLMs to Express Confidence with Self-Reflective Rationales"☆112Updated last year
- ☆161Updated last year
- Code for In-context Vectors: Making In Context Learning More Effective and Controllable Through Latent Space Steering☆198Updated 11 months ago
- Co-LLM: Learning to Decode Collaboratively with Multiple Language Models☆126Updated last year
- LongEmbed: Extending Embedding Models for Long Context Retrieval (EMNLP 2024)☆146Updated last year
- The first dense retrieval model that can be prompted like an LM☆90Updated 8 months ago
- Code for EMNLP 2024 paper "Learn Beyond The Answer: Training Language Models with Reflection for Mathematical Reasoning"☆54Updated last year
- ☆130Updated last year
- [NeurIPS 2023] This is the code for the paper `Large Language Model as Attributed Training Data Generator: A Tale of Diversity and Bias`.☆156Updated 2 years ago
- ☆107Updated last year
- Code for PHATGOOSE introduced in "Learning to Route Among Specialized Experts for Zero-Shot Generalization"☆91Updated last year
- [ACL'25 Oral] What Happened in LLMs Layers when Trained for Fast vs. Slow Thinking: A Gradient Perspective☆75Updated 7 months ago
- ☆75Updated last year
- Official repository for Montessori-Instruct: Generate Influential Training Data Tailored for Student Learning [ICLR 2025]☆50Updated last year
- Codes and datasets for the paper Measuring and Enhancing Trustworthiness of LLMs in RAG through Grounded Attributions and Learning to Ref…☆72Updated 11 months ago
- Official repository for paper "ReasonIR Training Retrievers for Reasoning Tasks".☆217Updated 7 months ago
- ☆52Updated 8 months ago
- [Preprint] Learning to Filter Context for Retrieval-Augmented Generaton☆196Updated last year
- Benchmarking LLMs with Challenging Tasks from Real Users☆245Updated last year
- Official code for "MAmmoTH2: Scaling Instructions from the Web" [NeurIPS 2024]☆148Updated last year
- Official repository for "Scaling Retrieval-Based Langauge Models with a Trillion-Token Datastore".☆224Updated last month
- ☆99Updated last year
- ☆38Updated last year
- [TMLR 2026] When Attention Collapses: How Degenerate Layers in LLMs Enable Smaller, Stronger Models☆122Updated 11 months ago
- Ruler: A Model-Agnostic Method to Control Generated Length for Large Language Models☆40Updated last year
- [COLING'25] Exploring Concept Depth: How Large Language Models Acquire Knowledge at Different Layers?☆82Updated last year
- [IJCAI 2024] FactCHD: Benchmarking Fact-Conflicting Hallucination Detection☆90Updated last year
- PASTA: Post-hoc Attention Steering for LLMs☆134Updated last year
- Evaluating LLMs with fewer examples☆169Updated last year
- [NeurIPS 2024] Knowledge Circuits in Pretrained Transformers☆163Updated 2 months ago