rohitrango / objects-that-soundLinks
Unofficial Implementation of Google Deepmind's paper `Objects that Sound`
☆83Updated 7 years ago
Alternatives and similar repositories for objects-that-sound
Users that are interested in objects-that-sound are comparing it to the libraries listed below
Sorting:
- Deep Audio-Visual Embedding network (DAVEnet) implementation in PyTorch☆65Updated 6 years ago
- Learning to Separate Object Sounds by Watching Unlabeled Video (ECCV 2018)☆51Updated 5 years ago
- Code for the paper: Audio-Visual Scene Analysis with Self-Supervised Multisensory Features☆220Updated 6 years ago
- 2.5D visual sound dataset☆99Updated 3 years ago
- MUSIC Dataset from The Sound of Pixels (ECCV '18)☆129Updated 2 years ago
- Audio Visual Instance Discrimination with Cross-Modal Agreement☆129Updated 3 years ago
- Listen to Look: Action Recognition by Previewing Audio (CVPR 2020)☆130Updated 3 years ago
- Keras Implementation of "Look, Listen and Learn" Model☆21Updated 7 years ago
- Implementation for ECCV20 paper "Self-Supervised Learning of audio-visual objects from video"☆113Updated 4 years ago
- Audio-Visual Event Localization in Unconstrained Videos, ECCV 2018☆185Updated 4 years ago
- Co-Separating Sounds of Visual Objects (ICCV 2019)☆96Updated last year
- Torch code for using Residual Networks with LSTMs for Lipreading☆98Updated 6 years ago
- Code for Discriminative Sounding Objects Localization (NeurIPS 2020)☆58Updated 3 years ago
- TensorFlow implementation of "SoundNet".☆145Updated 7 years ago
- Content-Based Video-Music Retrieval using Soft Intra-Modal Structure Constraint☆61Updated 7 years ago
- Codebase and Dataset for the paper: Learning to Localize Sound Source in Visual Scenes☆92Updated 7 months ago
- SVHF-Net for Cross-modal binary matching☆32Updated 6 years ago
- Adversarial Unsupervised Domain Adaptation for Acoustic Scene Classification☆35Updated 6 years ago
- Learn and L3 embedding from audio/video pairs☆87Updated 3 years ago
- Localizing Visual Sounds the Hard Way☆80Updated 3 years ago
- Self-Supervised Learning by Cross-Modal Audio-Video Clustering (NeurIPS 2020)☆90Updated 2 years ago
- VGGSound: A Large-scale Audio-Visual Dataset☆322Updated 3 years ago
- This repo covers the implementation for Labelling unlabelled videos from scratch with multi-modal self-supervision, which learns clusters…☆116Updated 4 years ago
- Pytorch implementation of 'See, Hear, and Read: Deep Aligned Representations'☆33Updated 6 years ago
- 2.5D visual sound☆114Updated last year
- converting the pretrained tensorflow SoundNet model to pytorch☆13Updated 3 years ago
- ☆29Updated 3 years ago
- EgoCom: A Multi-person Multi-modal Egocentric Communications Dataset☆57Updated 4 years ago
- Implementation of "Slow-Fast Auditory Streams for Audio Recognition, ICASSP, 2021" in PyTorch☆74Updated 3 years ago
- ☆59Updated 7 years ago