art-jang / LiTFiCLinks
[CVPR2025] Official code for Lost in Translation Found in Context
☆23Updated last week
Alternatives and similar repositories for LiTFiC
Users that are interested in LiTFiC are comparing it to the libraries listed below
Sorting:
- [ECCV2024] The official implementation of "Listen to Look into the Future: Audio-Visual Egocentric Gaze Anticipation".☆13Updated 10 months ago
- Official code for the paper "Understanding Co-speech Gestures in-the-wild"☆20Updated 2 months ago
- ☆40Updated 9 months ago
- [CVPR'23 Highlight] AutoAD: Movie Description in Context.☆102Updated last year
- Question-Aware Gaussian Experts for Audio-Visual Question Answering -- Official Pytorch Implementation (CVPR'25, Highlight)☆25Updated 7 months ago
- [NAACL'24] Repository for "SMILE: Multimodal Dataset for Understanding Laughter in Video with Language Models"☆15Updated last year
- [CVPR 2025] Official Repository of the paper "On the Consistency of Video Large Language Models in Temporal Comprehension"☆15Updated 3 months ago
- ☆37Updated 6 months ago
- [🏆 IJCV 2025 & ACCV 2024 Best Paper Honorable Mention] Official pytorch implementation of the paper "High-Quality Visually-Guided Sound …☆25Updated 2 months ago
- Code for "Modeling Multimodal Social Interactions: New Challenges and Baselines with Densely Aligned Representations" (CVPR 2024 Oral)☆18Updated last year
- Code implementation for the paper "Large-scale Pre-training for Grounded Video Caption Generation" (ICCV 2025)☆27Updated 2 months ago
- Codebase for the paper: "TIM: A Time Interval Machine for Audio-Visual Action Recognition"☆50Updated last year
- [ECCV2024, Oral, Best Paper Finalist] This is the official implementation of the paper "LEGO: Learning EGOcentric Action Frame Generation…☆39Updated 10 months ago
- ☆19Updated last year
- [AAAI-24] VVS : Video-to-Video Retrieval With Irrelevant Frame Suppression☆20Updated last year
- Code and data for the paper "Emergent Visual-Semantic Hierarchies in Image-Text Representations" (ECCV 2024)☆32Updated last year
- [ICLR 2025] CREMA: Generalizable and Efficient Video-Language Reasoning via Multimodal Modular Fusion☆55Updated 6 months ago
- [CVPR 2023] iQuery: Instruments as Queries for Audio-Visual Sound Separation☆71Updated 2 years ago
- Official implementation of EgoHOD at ICLR 2025; 14 EgoVis Challenge Winners in CVPR 2024☆31Updated last month
- [CVPR 2024] Code and datasets for 'Learning Spatial Features from Audio-Visual Correspondence in Egocentric Videos'☆13Updated last year
- ☆27Updated 2 years ago
- Official code for WACV 2024 paper, "Annotation-free Audio-Visual Segmentation"☆37Updated last year
- [EMNLP 2024] Preserving Multi-Modal Capabilities of Pre-trained VLMs for Improving Vision-Linguistic Compositionality☆21Updated last year
- Code implementation for paper titled "HOI-Ref: Hand-Object Interaction Referral in Egocentric Vision"☆29Updated last year
- ☆36Updated 7 months ago
- [ICCV 2025] Official Implementation of "Shot-by-Shot: Film-Grammar-Aware Training-Free Audio Description Generation". Junyu Xie, Tengda H…☆20Updated 5 months ago
- Official Codebase of "Localizing Visual Sounds the Easy Way" (ECCV 2022)☆37Updated 3 years ago
- [ACCV 2024] Official Implementation of "AutoAD-Zero: A Training-Free Framework for Zero-Shot Audio Description". Junyu Xie, Tengda Han, M…☆27Updated 11 months ago
- Official repository of PanoAVQA: Grounded Audio-Visual Question Answering in 360° Videos (ICCV 2021)☆17Updated 4 years ago
- Action Scene Graphs for Long-Form Understanding of Egocentric Videos (CVPR 2024)☆44Updated 9 months ago