Collection of papers about video-audio understanding
☆22Dec 26, 2025Updated 2 months ago
Alternatives and similar repositories for Awesome-omni-modal-understanding
Users that are interested in Awesome-omni-modal-understanding are comparing it to the libraries listed below
Sorting:
- [CAC2023] Bilateral Network with Residual U-blocks and Dual-Guided Attention for Real-time Semantic Segmentation☆11Nov 28, 2024Updated last year
- a unified and simple codebase for weakly-supervised temporal action localization☆19Sep 30, 2023Updated 2 years ago
- ☆35Mar 20, 2023Updated 2 years ago
- ☆11Dec 6, 2024Updated last year
- [ICCV 2025] Official PyTorch Code for "Describe, Adapt and Combine: Empowering CLIP Encoders for Open-set 3D Object Retrieval"☆17Aug 23, 2025Updated 6 months ago
- [ACL 2025] Official code for ''Learning to Reason from Feedback at Test-Time''.☆13May 16, 2025Updated 9 months ago
- ☆12Jul 4, 2024Updated last year
- ☆13Jul 3, 2024Updated last year
- [CVPR 2025] VASparse: Towards Efficient Visual Hallucination Mitigation via Visual-Aware Token Sparsification☆49Mar 24, 2025Updated 11 months ago
- The first multimodal QA dataset specifically designed for evaluating large TCM language models.☆21Oct 24, 2025Updated 4 months ago
- ☆20Jun 13, 2025Updated 8 months ago
- ☆13May 15, 2025Updated 9 months ago
- [NeurIPS 2025] VLA-Cache: Efficient Vision-Language-Action Manipulation via Adaptive Token Caching☆69Feb 27, 2026Updated last week
- ☆14Sep 11, 2025Updated 5 months ago
- ☆40Jan 16, 2026Updated last month
- open-source Mandarian biased word dataset☆14Sep 21, 2023Updated 2 years ago
- ☆20Nov 21, 2025Updated 3 months ago
- ☆10Jan 26, 2025Updated last year
- ☆14Dec 25, 2024Updated last year
- [ECCV 2024] The first zero-shot setting for spatio-temporal video grounding.☆11Jul 16, 2024Updated last year
- [AAAI 2025] The official repository of our paper "GCD: Advancing Vision-Language Models for Incremental Object Detection via Global Align…☆15Sep 10, 2025Updated 6 months ago
- [ICCV 2025] SparseMM: Head Sparsity Emerges from Visual Concept Responses in MLLMs☆82Jan 17, 2026Updated last month
- [NAACL 2025🔥] MEDA: Dynamic KV Cache Allocation for Efficient Multimodal Long-Context Inference☆18Jun 19, 2025Updated 8 months ago
- Code for paper: Unraveling the Shift of Visual Information Flow in MLLMs: From Phased Interaction to Efficient Inference☆13Jun 7, 2025Updated 9 months ago
- LLaVA-Next for STVG☆18Dec 5, 2025Updated 3 months ago
- ☆11Jun 13, 2024Updated last year
- [ICML 2024] Memory-Space Visual Prompting for Efficient Vision-Language Fine-Tuning☆50May 12, 2024Updated last year
- This is the official repository of Daily-Omni: Towards Audio-Visual Reasoning with Temporal Alignment across Modalities☆38Feb 25, 2026Updated last week
- [WACV 2024 Oral] Rethinking Visibility in Human Pose Estimation: Occluded Pose Reasoning via Transformers☆14Jul 6, 2024Updated last year
- [ICCV 2025] AdsQA: Towards Advertisement Video Understanding Arxiv: https://arxiv.org/abs/2509.08621☆33Oct 30, 2025Updated 4 months ago
- 🔥🔥[NeurIPS2025]Exploring and mitigating semantic hallucinations in scene text perception and reasoning☆26Dec 11, 2025Updated 2 months ago
- An efficient implementation of the FSG seed bank☆11Jan 4, 2022Updated 4 years ago
- https://avocado-captioner.github.io/☆31Oct 16, 2025Updated 4 months ago
- Official PyTorch code for ICLR 2025 paper "Gnothi Seauton: Empowering Faithful Self-Interpretability in Black-Box Models"☆24Mar 4, 2025Updated last year
- Fast, memory-efficient attention column reduction (e.g., sum, mean, max)☆37Feb 10, 2026Updated last month
- Extending context length of visual language models☆12Dec 18, 2024Updated last year
- ☆16Mar 24, 2025Updated 11 months ago
- ☆15Nov 1, 2024Updated last year
- A visual LLM for image region description or QA.☆16Jul 14, 2023Updated 2 years ago