edsonroteia / cav-mae-syncLinks
[CVPR25] Official Implementation of CAV-MAE Sync
☆20Updated last month
Alternatives and similar repositories for cav-mae-sync
Users that are interested in cav-mae-sync are comparing it to the libraries listed below
Sorting:
- Official implementation of USR (NeurIPS 2024)☆31Updated 6 months ago
- Offical code for the CVPR 2024 Paper: Separating the "Chirp" from the "Chat": Self-supervised Visual Grounding of Sound and Language☆83Updated last year
- ☆31Updated last week
- This repo contains the official PyTorch implementation of AudioToken: Adaptation of Text-Conditioned Diffusion Models for Audio-to-Image …☆84Updated last year
- Official implementation of RAVEn (ICLR 2023) and BRAVEn (ICASSP 2024)☆67Updated 4 months ago
- Action2Sound: Ambient-Aware Generation of Action Sounds from Egocentric Videos☆22Updated 9 months ago
- ☆18Updated last year
- Code for the IEEE Signal Processing Letters 2022 paper "UAVM: Towards Unifying Audio and Visual Models".☆55Updated 2 years ago
- [CVPR'23 Highlight] AutoAD: Movie Description in Context.☆100Updated 8 months ago
- The repo host the code and model of MAViL.☆44Updated last year
- Splits for epic-sounds dataset☆76Updated 7 months ago
- Source code for "Sparse in Space and Time: Audio-visual Synchronisation with Trainable Selectors." (Spotlight at the BMVC 2022)☆51Updated last year
- Code and Pretrained Models for ICLR 2023 Paper "Contrastive Audio-Visual Masked Autoencoder".☆264Updated last year
- Official implementation for AVGN☆35Updated 2 years ago
- Official PyTorch implementation of ReWaS (AAAI'25) "Read, Watch and Scream! Sound Generation from Text and Video"☆42Updated 7 months ago
- Code for the C2KD paper (ICASSP 2023)☆18Updated 2 years ago
- Official Implementation of the work "Audio Mamba: Bidirectional State Space Model for Audio Representation Learning"☆145Updated 7 months ago
- ☆34Updated last month
- Vision Transformers are Parameter-Efficient Audio-Visual Learners☆100Updated last year
- [ECCV’24] Official Implementation for CAT: Enhancing Multimodal Large Language Model to Answer Questions in Dynamic Audio-Visual Scenario…☆54Updated 10 months ago
- Ego4DSounds: A diverse egocentric dataset with high action-audio correspondence☆18Updated last year
- Source code for the paper 'Audio Captioning Transformer'☆54Updated 3 years ago
- Official implementation for MGN☆20Updated 2 years ago
- code for A Large-scale Dataset for Audio-Language Representation Learning☆13Updated 10 months ago
- [AAAI 2023 (Oral)] CrissCross: Self-Supervised Audio-Visual Representation Learning with Relaxed Cross-Modal Synchronicity☆25Updated 2 years ago
- NeurIPS'2023 official implementation code☆64Updated last year
- ☆65Updated 2 years ago
- Pytorch implementation for “V2C: Visual Voice Cloning”☆32Updated 2 years ago
- Source code for "Synchformer: Efficient Synchronization from Sparse Cues" (ICASSP 2024)☆70Updated 5 months ago
- ACAV100M: Automatic Curation of Large-Scale Datasets for Audio-Visual Video Representation Learning. In ICCV, 2021.☆57Updated 3 years ago