PKU-YuanGroup / PiCOLinks
[ICLR'25] PiCO: Peer Review in LLMs based on the Consistency Optimization, https://arxiv.org/pdf/2402.01830
☆36Updated 5 months ago
Alternatives and similar repositories for PiCO
Users that are interested in PiCO are comparing it to the libraries listed below
Sorting:
- Code for the paper "AsFT: Anchoring Safety During LLM Fune-Tuning Within Narrow Safety Basin".☆25Updated last week
- V1: Toward Multimodal Reasoning by Designing Auxiliary Task☆34Updated 3 months ago
- GPT as a Monte Carlo Language Tree: A Probabilistic Perspective☆45Updated 6 months ago
- [NeurIPS 2024] Calibrated Self-Rewarding Vision Language Models☆76Updated last year
- Official implement of MIA-DPO☆59Updated 5 months ago
- Doodling our way to AGI ✏️ 🖼️ 🧠☆72Updated last month
- code for "Strengthening Multimodal Large Language Model with Bootstrapped Preference Optimization"☆55Updated 10 months ago
- This repository is the official implementation of "Look-Back: Implicit Visual Re-focusing in MLLM Reasoning".☆29Updated last week
- [ICLR2025] MMIU: Multimodal Multi-image Understanding for Evaluating Large Vision-Language Models☆81Updated 10 months ago
- (ICLR2025 Spotlight) DEEM: Official implementation of Diffusion models serve as the eyes of large language models for image perception.☆35Updated 2 weeks ago
- A Comprehensive Benchmark and Toolkit for Evaluating Video-based Large Language Models!☆130Updated last year
- VideoNIAH: A Flexible Synthetic Method for Benchmarking Video MLLMs☆48Updated 4 months ago
- GPT-4V(ision) as A Social Media Analysis Engine☆37Updated 6 months ago
- NoisyRollout: Reinforcing Visual Reasoning with Data Augmentation☆78Updated last month
- HalluciDoctor: Mitigating Hallucinatory Toxicity in Visual Instruction Data (Accepted by CVPR 2024)☆45Updated last year
- ☆27Updated 8 months ago
- Enhancing Large Vision Language Models with Self-Training on Image Comprehension.☆69Updated last year
- Fast-Slow Thinking for Large Vision-Language Model Reasoning☆16Updated 2 months ago
- ✨✨The Curse of Multi-Modalities (CMM): Evaluating Hallucinations of Large Multimodal Models across Language, Visual, and Audio☆46Updated last week
- The codebase for our EMNLP24 paper: Multimodal Self-Instruct: Synthetic Abstract Image and Visual Reasoning Instruction Using Language Mo…☆79Updated 5 months ago
- ☆45Updated 6 months ago
- VideoHallucer, The first comprehensive benchmark for hallucination detection in large video-language models (LVLMs)☆35Updated 3 months ago
- [EMNLP 2024] mDPO: Conditional Preference Optimization for Multimodal Large Language Models.☆77Updated 8 months ago
- Official code for "What Makes for Good Visual Tokenizers for Large Language Models?".☆58Updated 2 years ago
- TemporalBench: Benchmarking Fine-grained Temporal Understanding for Multimodal Video Models☆33Updated 8 months ago
- [EMNLP'23] The official GitHub page for ''Evaluating Object Hallucination in Large Vision-Language Models''☆85Updated last year
- ☆174Updated 3 weeks ago
- (ICLR 2025 Spotlight) Official code repository for Interleaved Scene Graph.☆22Updated 5 months ago
- [SCIS 2024] The official implementation of the paper "MMInstruct: A High-Quality Multi-Modal Instruction Tuning Dataset with Extensive Di…☆55Updated 8 months ago
- Implementation and dataset for paper "Can MLLMs Perform Text-to-Image In-Context Learning?"☆39Updated last month