GeWu-Lab / CrabLinks
[CVPR 2025] Crab: A Unified Audio-Visual Scene Understanding Model with Explicit Cooperation
☆42Updated 3 weeks ago
Alternatives and similar repositories for Crab
Users that are interested in Crab are comparing it to the libraries listed below
Sorting:
- ☆30Updated 8 months ago
- [2025 CVPR] Towards Open-Vocabulary Audio-Visual Event Localization☆21Updated 3 months ago
- [CVPR 2024 Highlight] Official implementation of the paper: Cooperation Does Matter: Exploring Multi-Order Bilateral Relations for Audio-…☆39Updated 2 months ago
- Official repository for "Boosting Audio Visual Question Answering via Key Semantic-Aware Cues" in ACM MM 2024.☆16Updated 8 months ago
- [ECCV’24] Official Implementation for CAT: Enhancing Multimodal Large Language Model to Answer Questions in Dynamic Audio-Visual Scenario…☆53Updated 9 months ago
- [AAAI 2024] AVSegFormer: Audio-Visual Segmentation with Transformer☆67Updated 3 months ago
- This repository contains code for AAAI2025 paper "Dense Audio-Visual Event Localization under Cross-Modal Consistency and Multi-Temporal …☆18Updated 5 months ago
- Official code for WACV 2024 paper, "Annotation-free Audio-Visual Segmentation"☆31Updated 8 months ago
- The official repo for "Ref-AVS: Refer and Segment Objects in Audio-Visual Scenes", ECCV 2024☆43Updated 6 months ago
- Research code for NeurIPS 2023 paper "Modality-Independent Teachers Meet Weakly-Supervised Audio-Visual Event Parser"☆18Updated last year
- Dense-Localizing Audio-Visual Events in Untrimmed Videos: A Large-Scale Benchmark and Baseline (CVPR 2023)☆64Updated last year
- Towards Long Form Audio-visual Video Understanding☆15Updated 2 months ago
- [CVPR 2025] 🔥 Official impl. of "Audio-Visual Instance Segmentation".☆24Updated 3 weeks ago
- Official Implementation of "Open-Vocabulary Audio-Visual Semantic Segmentation" [ACM MM 2024 Oral].☆29Updated 7 months ago
- Vision Transformers are Parameter-Efficient Audio-Visual Learners☆99Updated last year
- MUSIC-AVQA, CVPR2022 (ORAL)☆85Updated 2 years ago
- Official codebase for "Unveiling the Power of Audio-Visual Early Fusion Transformers with Dense Interactions through Masked Modeling".☆32Updated 10 months ago
- Unified Audio-Visual Perception for Multi-Task Video Localization☆25Updated last year
- Codebase for the paper: "TIM: A Time Interval Machine for Audio-Visual Action Recognition"☆41Updated 7 months ago
- Official repository of "Prompting Segmentation with Sound is Generalizable Audio-Visual Source Localizer", AAAI 2024☆20Updated last year
- NeurIPS'2023 official implementation code☆64Updated last year
- LongVALE: Vision-Audio-Language-Event Benchmark Towards Time-Aware Omni-Modal Perception of Long Videos. (CVPR 2025))☆34Updated 2 weeks ago
- [TPAMI 2024] Dynamic MDETR: A Dynamic Multimodal Transformer Decoder for Visual Grounding☆26Updated 9 months ago
- ☆14Updated last year
- ☆20Updated 5 months ago
- Official implementation for MGN☆20Updated 2 years ago
- The code repo for ICASSP 2023 Paper "MMCosine: Multi-Modal Cosine Loss Towards Balanced Audio-Visual Fine-Grained Learning"☆21Updated 2 years ago
- VideoChat-R1: Enhancing Spatio-Temporal Perception via Reinforcement Fine-Tuning☆153Updated 2 weeks ago
- This is the official implementation of ReVisionLLM: Recursive Vision-Language Model for Temporal Grounding in Hour-Long Videos☆23Updated this week
- Grounded-VideoLLM: Sharpening Fine-grained Temporal Grounding in Video Large Language Models☆116Updated 3 months ago