PKU-YuanGroup / Peer-review-in-LLMs
Peer-review-in-LLMs: Automatic Evaluation Method for LLMs in Open-environment,https://arxiv.org/pdf/2402.01830.pdf
☆26Updated 7 months ago
Related projects: ⓘ
- ☆13Updated this week
- LLMBind: A Unified Modality-Task Integration Framework☆14Updated 3 months ago
- Repo for the paper `ControlMLLM: Training-Free Visual Prompt Learning for Multimodal Large Language Models'☆44Updated 3 weeks ago
- Implementation of Foundation Model is Efficient Multimodal Multitask Model Selector☆33Updated 6 months ago
- GPT-4V(ision) as A Social Media Analysis Engine☆30Updated 10 months ago
- [EMNLP'23] The official GitHub page for ''Evaluating Object Hallucination in Large Vision-Language Models''☆67Updated 5 months ago
- MMIU: Multimodal Multi-image Understanding for Evaluating Large Vision-Language Models☆35Updated this week
- A Comprehensive Benchmark and Toolkit for Evaluating Video-based Large Language Models!☆114Updated 8 months ago
- Official Dataloader and Evaluation Scripts for LongVideoBench.☆52Updated last month
- 本项目用于Multimodal领域新手的学习路线,包括该领域的经典论文,项目及课程。旨在希望学习者在一定的时间内达到对这个领域有较为深刻的认知,能够自己进行的独立研究。☆14Updated 5 months ago
- HalluciDoctor: Mitigating Hallucinatory Toxicity in Visual Instruction Data (Accepted by CVPR 2024)☆39Updated 2 months ago
- ☆70Updated 4 months ago
- ☆15Updated 6 months ago
- ☆32Updated 3 months ago
- ☕️ CREMA: Generalizable and Efficient Video-Language Reasoning via Multimodal Modular Fusion☆24Updated 3 months ago
- [Arxiv] Calibrated Self-Rewarding Vision Language Models☆35Updated 3 months ago
- Code for paper "AGLA: Mitigating Object Hallucinations in Large Vision-Language Models with Assembly of Global and Local Attention"☆13Updated 2 months ago
- Visual CoT: Advancing Multi-Modal Language Models with a Comprehensive Dataset and Benchmark for Chain-of-Thought Reasoning☆93Updated 2 months ago
- ☆19Updated last month
- The official code of paper "Automated Multi-level Preference for MLLMs"☆15Updated 3 weeks ago
- ☆16Updated this week
- ☆37Updated 3 months ago
- ACL'24 (Oral) Tuning Large Multimodal Models for Videos using Reinforcement Learning from AI Feedback☆39Updated last week
- LLaVA-NeXT-Image-Llama3-Lora, Modified from https://github.com/arielnlee/LLaVA-1.6-ft☆37Updated 2 months ago
- FreeVA: Offline MLLM as Training-Free Video Assistant☆42Updated 3 months ago
- MADAv2: Advanced Multi-Anchor Based Active Domain Adaptation Segmentation☆24Updated last year
- Official implementation of "Why are Visually-Grounded Language Models Bad at Image Classification?"☆32Updated 2 months ago
- ☆52Updated 4 months ago
- Enhancing Large Vision Language Models with Self-Training on Image Comprehension.☆51Updated 3 months ago
- Official repository for CoMM Dataset☆16Updated this week