yanbeic / CCL
PyTorch Implementation on Paper [CVPR2021]Distilling Audio-Visual Knowledge by Compositional Contrastive Learning
☆84Updated 3 years ago
Alternatives and similar repositories for CCL:
Users that are interested in CCL are comparing it to the libraries listed below
- Official implementation of "Everything at Once - Multi-modal Fusion Transformer for Video Retrieval". CVPR 2022☆98Updated 2 years ago
- Unified Multisensory Perception: Weakly-Supervised Audio-Visual Video Parsing, ECCV, 2020. (Spotlight)☆83Updated 6 months ago
- Self-Supervised Learning by Cross-Modal Audio-Video Clustering (NeurIPS 2020)☆90Updated 2 years ago
- [2021 CVPR] Positive Sample Propagation along the Audio-Visual Event Line☆41Updated 2 years ago
- Code for the AVLnet (Interspeech 2021) and Cascaded Multilingual (Interspeech 2021) papers.☆50Updated 2 years ago
- This repository contains the code for our CVPR 2022 paper on "Audio-visual Generalised Zero-shot Learning with Cross-modal Attention and …☆35Updated 2 years ago
- ☆31Updated 3 years ago
- MUSIC-AVQA, CVPR2022 (ORAL)☆75Updated 2 years ago
- Vision Transformers are Parameter-Efficient Audio-Visual Learners☆96Updated last year
- The repo for "Class-aware Sounding Objects Localization", TPAMI 2021.☆29Updated 2 years ago
- CrossCLR: Cross-modal Contrastive Learning For Multi-modal Video Representations, ICCV 2021☆61Updated 3 years ago
- Official implementation of AdaMML. https://arxiv.org/abs/2105.05165.☆50Updated 2 years ago
- ☆72Updated 2 years ago
- Code on selecting an action based on multimodal inputs. Here in this case inputs are voice and text.☆69Updated 3 years ago
- Some papers about *diverse* image (a few videos) captioning☆26Updated last year
- ICCV 2021☆33Updated 2 years ago
- ☆31Updated 4 years ago
- Cross Modal Retrieval with Querybank Normalisation☆55Updated last year
- ☆14Updated last year
- Code for Discriminative Sounding Objects Localization (NeurIPS 2020)☆57Updated 3 years ago
- This repository contains the code for our ECCV 2022 paper "Temporal and cross-modal attention for audio-visual zero-shot learning"☆24Updated 2 years ago
- https://layer6ai-labs.github.io/xpool/☆118Updated last year
- CVPR2022☆20Updated 2 years ago
- Code for ACM MM2020 paper: Jointly Cross- and Self-Modal Graph Attention Network for Query-Based Moment Localization☆34Updated 4 years ago
- Code for CVPR 2021 paper Exploring Heterogeneous Clues for Weakly-Supervised Audio-Visual Video Parsing☆24Updated 3 years ago
- ☆24Updated 2 years ago
- Cross-Modal Relation-Aware Networks for Audio-Visual Event Localization, ACM MM 2020☆33Updated 4 years ago
- Cross-model active contrastive coding☆22Updated 3 years ago
- Learning phrase grounding from captioned images through InfoNCE bound on mutual information☆72Updated 4 years ago
- Code for the paper "Zero-shot Natural Language Video Localization" (ICCV2021, Oral).☆47Updated last year