inFaaa / EvolverLinks
[COLING 2025π₯] Evolver: Chain-of-Evolution Prompting to Boost Large Multimodal Models for Hateful Meme Detection
β13Updated 5 months ago
Alternatives and similar repositories for Evolver
Users that are interested in Evolver are comparing it to the libraries listed below
Sorting:
- [EMNLP 2024 Findingsπ₯] Official implementation of ": LOOK-M: Look-Once Optimization in KV Cache for Efficient Multimodal Long-Context Inβ¦β97Updated 8 months ago
- A Self-Training Framework for Vision-Language Reasoningβ80Updated 5 months ago
- [ACM MM 2025] TimeChat-online: 80% Visual Tokens are Naturally Redundant in Streaming Videosβ60Updated this week
- Official Code and data for ACL 2024 finding, "An Empirical Study on Parameter-Efficient Fine-Tuning for MultiModal Large Language Models"β19Updated 8 months ago
- β53Updated 2 months ago
- [EMNLP 2024] mDPO: Conditional Preference Optimization for Multimodal Large Language Models.β77Updated 8 months ago
- V1: Toward Multimodal Reasoning by Designing Auxiliary Taskβ34Updated 3 months ago
- [ICLR 2025] The official pytorch implement of "Dynamic-LLaVA: Efficient Multimodal Large Language Models via Dynamic Vision-language Contβ¦β46Updated 7 months ago
- β76Updated last year
- NoisyRollout: Reinforcing Visual Reasoning with Data Augmentationβ78Updated last month
- Code release for VTW (AAAI 2025 Oral)β45Updated this week
- [ICCV 2025] The official code of the paper "Deciphering Cross-Modal Alignment in Large Vision-Language Models with Modality Integration Rβ¦β101Updated last week
- More Thinking, Less Seeing? Assessing Amplified Hallucination in Multimodal Reasoning Modelsβ30Updated last month
- MME-CoT: Benchmarking Chain-of-Thought in LMMs for Reasoning Quality, Robustness, and Efficiencyβ117Updated 3 weeks ago
- VoCoT: Unleashing Visually Grounded Multi-Step Reasoning in Large Multi-Modal Modelsβ69Updated last year
- Less is More: Mitigating Multimodal Hallucination from an EOS Decision Perspective (ACL 2024)β54Updated 8 months ago
- β¨β¨The Curse of Multi-Modalities (CMM): Evaluating Hallucinations of Large Multimodal Models across Language, Visual, and Audioβ46Updated last week
- Official repository of the video reasoning benchmark MMR-V. Can Your MLLMs "Think with Video"?β31Updated 3 weeks ago
- β16Updated 2 months ago
- β129Updated 5 months ago
- β23Updated last month
- A Comprehensive Survey on Evaluating Reasoning Capabilities in Multimodal Large Language Models.β66Updated 4 months ago
- β25Updated last year
- MultiMath: Bridging Visual and Mathematical Reasoning for Large Language Modelsβ29Updated 5 months ago
- Code for DeCo: Decoupling token compression from semanchc abstraction in multimodal large language modelsβ40Updated this week
- [ECCV 2024] Paying More Attention to Image: A Training-Free Method for Alleviating Hallucination in LVLMsβ128Updated 8 months ago
- π Global Compression Commander: Plug-and-Play Inference Acceleration for High-Resolution Large Vision-Language Modelsβ30Updated last month
- β25Updated 5 months ago
- β46Updated 3 months ago
- [Neurips 24' D&B] Official Dataloader and Evaluation Scripts for LongVideoBench.β103Updated 11 months ago