facebookresearch / AVID-CMA
Audio Visual Instance Discrimination with Cross-Modal Agreement
☆128Updated 3 years ago
Alternatives and similar repositories for AVID-CMA:
Users that are interested in AVID-CMA are comparing it to the libraries listed below
- Self-Supervised Learning by Cross-Modal Audio-Video Clustering (NeurIPS 2020)☆90Updated 2 years ago
- Unified Multisensory Perception: Weakly-Supervised Audio-Visual Video Parsing, ECCV, 2020. (Spotlight)☆83Updated 6 months ago
- This repo covers the implementation for Labelling unlabelled videos from scratch with multi-modal self-supervision, which learns clusters…☆115Updated 3 years ago
- Audio-Visual Event Localization in Unconstrained Videos, ECCV 2018☆178Updated 3 years ago
- Listen to Look: Action Recognition by Previewing Audio (CVPR 2020)☆128Updated 3 years ago
- Implementation for ECCV20 paper "Self-Supervised Learning of audio-visual objects from video"☆111Updated 4 years ago
- Localizing Visual Sounds the Hard Way☆78Updated 2 years ago
- Code for CVPR 2021 paper Exploring Heterogeneous Clues for Weakly-Supervised Audio-Visual Video Parsing☆24Updated 3 years ago
- Cross-model active contrastive coding☆22Updated 3 years ago
- PyTorch GPU distributed training code for MIL-NCE HowTo100M☆215Updated 2 years ago
- Official implementation of ACMMM'20 paper 'Self-supervised Video Representation Learning Using Inter-intra Contrastive Framework'☆111Updated 3 years ago
- Code for the AVLnet (Interspeech 2021) and Cascaded Multilingual (Interspeech 2021) papers.☆50Updated 2 years ago
- ☆28Updated 2 years ago
- Code for Discriminative Sounding Objects Localization (NeurIPS 2020)☆57Updated 3 years ago
- Unofficial Implementation of Google Deepmind's paper `Objects that Sound`☆83Updated 6 years ago
- [NeurIPS'20] Self-supervised Co-Training for Video Representation Learning. Tengda Han, Weidi Xie, Andrew Zisserman.☆288Updated 3 years ago
- VGGSound: A Large-scale Audio-Visual Dataset☆303Updated 3 years ago
- ☆88Updated 3 years ago
- [2021 CVPR] Positive Sample Propagation along the Audio-Visual Event Line☆41Updated 2 years ago
- EgoCom: A Multi-person Multi-modal Egocentric Communications Dataset☆55Updated 4 years ago
- Download scripts for EPIC-KITCHENS☆129Updated 6 months ago
- Datasets, transforms and samplers for video in PyTorch☆87Updated last year
- Deep Audio-Visual Embedding network (DAVEnet) implementation in PyTorch☆65Updated 6 years ago
- PyTorch Implementation on Paper [CVPR2021]Distilling Audio-Visual Knowledge by Compositional Contrastive Learning☆85Updated 3 years ago
- Co-Separating Sounds of Visual Objects (ICCV 2019)☆94Updated last year
- Implementation of "EPIC-Fusion: Audio-Visual Temporal Binding for Egocentric Action Recognition, ICCV, 2019" in PyTorch☆110Updated 4 years ago
- MAD: A Scalable Dataset for Language Grounding in Videos from Movie Audio Descriptions☆158Updated last year
- A Dataset for Grounded Video Description☆160Updated 3 years ago
- Official Codebase of "Localizing Visual Sounds the Easy Way" (ECCV 2022)☆31Updated 2 years ago
- S3D Text-Video model trained on HowTo100M using MIL-NCE☆195Updated 4 years ago