PKU-YuanGroup / LLMBind
LLMBind: A Unified Modality-Task Integration Framework
☆18Updated 10 months ago
Alternatives and similar repositories for LLMBind:
Users that are interested in LLMBind are comparing it to the libraries listed below
- FreeVA: Offline MLLM as Training-Free Video Assistant☆61Updated 11 months ago
- This repo contains evaluation code for the paper "AV-Odyssey: Can Your Multimodal LLMs Really Understand Audio-Visual Information?"☆24Updated 4 months ago
- [ECCV 2024] Learning Video Context as Interleaved Multimodal Sequences☆38Updated last month
- Code for "CAFe: Unifying Representation and Generation with Contrastive-Autoregressive Finetuning"☆13Updated last month
- VisRL: Intention-Driven Visual Perception via Reinforced Reasoning☆27Updated last month
- A Massive Multi-Discipline Lecture Understanding Benchmark☆16Updated this week
- ☆25Updated last month
- The official repository for paper "PruneVid: Visual Token Pruning for Efficient Video Large Language Models".☆38Updated 2 months ago
- Official implement of MIA-DPO☆57Updated 3 months ago
- Official repository for paper "Scaling Video-Language Models to 10K Frames via Hierarchical Differential Distillation"☆18Updated this week
- Official implementation of Next Block Prediction: Video Generation via Semi-Autoregressive Modeling☆31Updated 2 months ago
- ImageGen-CoT: Enhancing Text-to-Image In-context Learning with Chain-of-Thought Reasoning☆31Updated last month
- [CVPR 2025] OVO-Bench: How Far is Your Video-LLMs from Real-World Online Video Understanding?☆55Updated last month
- 【NeurIPS 2024】The official code of paper "Automated Multi-level Preference for MLLMs"☆19Updated 7 months ago
- [CVPR 2025 Oral] VideoEspresso: A Large-Scale Chain-of-Thought Dataset for Fine-Grained Video Reasoning via Core Frame Selection☆75Updated 3 weeks ago
- [ICLR 2025] CREMA: Generalizable and Efficient Video-Language Reasoning via Multimodal Modular Fusion☆43Updated 3 months ago
- [NeurIPS'24] Official PyTorch Implementation of Seeing the Image: Prioritizing Visual Correlation by Contrastive Alignment☆56Updated 7 months ago
- ☆30Updated 9 months ago
- MME-Unify: A Comprehensive Benchmark for Unified Multimodal Understanding and Generation Models☆34Updated 3 weeks ago
- [NeurIPS 2024 D&B Track] Official Repo for "LVD-2M: A Long-take Video Dataset with Temporally Dense Captions"☆55Updated 6 months ago
- WISE: A World Knowledge-Informed Semantic Evaluation for Text-to-Image Generation☆81Updated last month
- ☆31Updated last month
- Official repository of "Inst-IT: Boosting Multimodal Instance Understanding via Explicit Visual Prompt Instruction Tuning"☆29Updated 2 months ago
- ☆35Updated 10 months ago
- Official repo of the ICLR 2025 paper "MMWorld: Towards Multi-discipline Multi-faceted World Model Evaluation in Videos"☆25Updated 7 months ago
- LongVALE: Vision-Audio-Language-Event Benchmark Towards Time-Aware Omni-Modal Perception of Long Videos. (CVPR 2025))☆20Updated 3 weeks ago
- Visual Programming for Text-to-Image Generation and Evaluation (NeurIPS 2023)☆56Updated last year
- [CVPR 2024] The official implementation of paper "synthesize, diagnose, and optimize: towards fine-grained vision-language understanding"☆42Updated 2 months ago
- [ICLR2025] γ -MOD: Mixture-of-Depth Adaptation for Multimodal Large Language Models☆36Updated 2 months ago
- VideoHallucer, The first comprehensive benchmark for hallucination detection in large video-language models (LVLMs)☆29Updated last month