EchoseChen / SPA-VL-RLHF
The reinforcement learning codes for dataset SPA-VL
β15Updated 2 months ago
Related projects: β
- π up-to-date & curated list of awesome LMM hallucinations papers, methods & resources.β140Updated 5 months ago
- mPLUG-HalOwl: Multimodal Hallucination Evaluation and Mitigatingβ75Updated 7 months ago
- [CVPR 2024 Highlight] Mitigating Object Hallucinations in Large Vision-Language Models through Visual Contrastive Decodingβ177Updated 2 months ago
- β23Updated 7 months ago
- A RLHF Infrastructure for Vision-Language Modelsβ86Updated 3 months ago
- ChartMimic: Evaluating LMMβs Cross-Modal Reasoning Capability via Chart-to-Code Generationβ80Updated 2 months ago
- Accepted by IJCAI-24 Survey Trackβ117Updated 3 weeks ago
- code for "Strengthening Multimodal Large Language Model with Bootstrapped Preference Optimization"β40Updated 3 weeks ago
- my commonly-used toolsβ46Updated last month
- This is the repo for our paper "Mr-Ben: A Comprehensive Meta-Reasoning Benchmark for Large Language Models"β38Updated 2 months ago
- The official GitHub page for ''Evaluating Object Hallucination in Large Vision-Language Models''β166Updated 5 months ago
- Repository for Label Words are Anchors: An Information Flow Perspective for Understanding In-Context Learningβ144Updated 7 months ago
- An LLM-free Multi-dimensional Benchmark for Multi-modal Hallucination Evaluationβ85Updated 8 months ago
- β28Updated 7 months ago
- [ACL 2024] A Prospector of Long-Dependency Data for Large Language Modelsβ48Updated last month
- [CVPR 2024 Highlight] OPERA: Alleviating Hallucination in Multi-Modal Large Language Models via Over-Trust Penalty and Retrospection-Alloβ¦β255Updated 3 weeks ago
- Code and data for "ConflictBank: A Benchmark for Evaluating the Influence of Knowledge Conflicts in LLM"β14Updated 3 months ago
- [CVPR'24] HallusionBench: You See What You Think? Or You Think What You See? An Image-Context Reasoning Benchmark Challenging for GPT-4V(β¦β220Updated 6 months ago
- β73Updated 8 months ago
- Official implementation of "LOOK-M: Look-Once Optimization in KV Cache for Efficient Multimodal Long-Context Inference"β61Updated 2 months ago
- β71Updated 8 months ago
- Watch Every Step! LLM Agent Learning via Iterative Step-level Process Refinementβ21Updated last month
- Code & Data for our Paper "Alleviating Hallucinations of Large Language Models through Induced Hallucinations"β56Updated 6 months ago
- Paper list and datasets for the paper: A Survey on Data Selection for LLM Instruction Tuningβ27Updated 7 months ago
- A Neural-Symbolic Self-Training Frameworkβ95Updated last month
- β17Updated 4 months ago
- [ACL'2024 Findings] GAOKAO-MM: A Chinese Human-Level Benchmark for Multimodal Models Evaluationβ32Updated 6 months ago
- β22Updated last year
- MoCLE (First MLLM with MoE for instruction customization and generalization!) (https://arxiv.org/abs/2312.12379)β28Updated 5 months ago
- This is the first released survey paper on hallucinations of large vision-language models (LVLMs). To keep track of this field and continβ¦β32Updated last month