ChicagoHAI / hypothesis-generation
This is the official repository for HypoGeniC (Hypothesis Generation in Context), which is an automated, data-driven tool that leverages large language models to generate hypothesis for open-domain research. For more details, please see the original paper using the link below.
☆23Updated this week
Related projects: ⓘ
- ☆33Updated 9 months ago
- A package to evaluate factuality of long-form generation. Original implementation of our EMNLP 2023 paper "FActScore: Fine-grained Atomic…☆271Updated 4 months ago
- ☆101Updated last year
- ☆113Updated 3 months ago
- Repository for the Bias Benchmark for QA dataset.☆83Updated 8 months ago
- paper list on reasoning in NLP☆169Updated 10 months ago
- [EMNLP 2023] MQuAKE: Assessing Knowledge Editing in Language Models via Multi-Hop Questions☆96Updated last week
- ☆44Updated 5 months ago
- This is a collection of research papers for Self-Correcting Large Language Models with Automated Feedback.☆386Updated 7 months ago
- ☆33Updated last year
- Code and data for "Lost in the Middle: How Language Models Use Long Contexts"☆300Updated 8 months ago
- MAD: The first work to explore Multi-Agent Debate with Large Language Models :D☆242Updated 10 months ago
- Sotopia: an Open-ended Social Learning Environment (ICLR 2024 spotlight)☆146Updated this week
- Representation Engineering: A Top-Down Approach to AI Transparency☆693Updated last month
- Official Repo for ICLR 2024 paper MINT: Evaluating LLMs in Multi-turn Interaction with Tools and Language Feedback by Xingyao Wang*, Ziha…☆100Updated 3 months ago
- The official repo of paper "Self-Control of LLM Behaviors by Compressing Suffix Gradient into Prefix Controller"☆17Updated last month
- [EMNLP 2023] Enabling Large Language Models to Generate Text with Citations. Paper: https://arxiv.org/abs/2305.14627☆442Updated last month
- LLM hallucination paper list☆268Updated 6 months ago
- Multilingual Large Language Models Evaluation Benchmark☆91Updated last month
- A curated list of Human Preference Datasets for LLM fine-tuning, RLHF, and eval.☆300Updated 11 months ago
- Steering Llama 2 with Contrastive Activation Addition☆83Updated 3 months ago
- ☆92Updated 4 months ago
- This is the repository of HaluEval, a large-scale hallucination evaluation benchmark for Large Language Models.☆386Updated 7 months ago
- Source code of our paper MIND, ACL 2024 Long Paper☆19Updated 3 months ago
- Official implementation for the paper "DoLa: Decoding by Contrasting Layers Improves Factuality in Large Language Models"☆405Updated 4 months ago
- ☆267Updated 2 months ago
- Awesome-LLM-Robustness: a curated list of Uncertainty, Reliability and Robustness in Large Language Models☆627Updated 3 months ago
- Awesome papers involving LLMs in Social Science.☆222Updated this week
- ☆160Updated last year
- The implementation of "RQUGE: Reference-Free Metric for Evaluating Question Generation by Answering the Question" [ACL 2023]☆13Updated 5 months ago