my-yy / sl_icmr2022
Code for "Self-Lifting: A Novel Framework For Unsupervised Voice-Face Association Learning,ICMR,2022"
☆11Updated 3 months ago
Alternatives and similar repositories for sl_icmr2022:
Users that are interested in sl_icmr2022 are comparing it to the libraries listed below
- PyTorch implementation of "Multi-modality Associative Bridging through Memory: Speech Sound Recollected from Face Video" (ICCV2021)☆20Updated 2 years ago
- FG2021: Cross Attentional AV Fusion for Dimensional Emotion Recognition☆26Updated 2 months ago
- ☆9Updated 3 years ago
- Multimodal Variational Auto-encoder based Audio-Visual Segmentation [ICCV2023].☆19Updated 5 months ago
- [IJCAI2022] Unsupervised Voice-Face Representation Learning by Cross-Modal Prototype Contrast☆20Updated last year
- ☆14Updated 3 years ago
- Vision Transformers are Parameter-Efficient Audio-Visual Learners☆97Updated last year
- The code repo for ICASSP 2023 Paper "MMCosine: Multi-Modal Cosine Loss Towards Balanced Audio-Visual Fine-Grained Learning"☆19Updated last year
- ☆20Updated 4 months ago
- [CVPR 2024] EmoVIT: Revolutionizing Emotion Insights with Visual Instruction Tuning☆24Updated 5 months ago
- [CVPR 2023] Code for "Learning Emotion Representations from Verbal and Nonverbal Communication"☆45Updated this week
- ☆17Updated 3 years ago
- The official implementation of OpenSR (ACL2023 Oral)☆15Updated last year
- The repo for "Class-aware Sounding Objects Localization", TPAMI 2021.☆29Updated 2 years ago
- PyTorch Implementation on Paper [CVPR2021]Distilling Audio-Visual Knowledge by Compositional Contrastive Learning☆84Updated 3 years ago
- [2021 CVPR] Positive Sample Propagation along the Audio-Visual Event Line☆41Updated 2 years ago
- Official implementation of FOP method as described in "Fusion and Orthogonal Projection for Improved Face-Voice Association"☆17Updated last year
- PyTorch code for "Self-Supervised Predictive Learning: A Negative-Free Method for Sound Source Localization in Visual Scenes" (CVPR, 2022…☆31Updated 7 months ago
- Code on selecting an action based on multimodal inputs. Here in this case inputs are voice and text.☆69Updated 3 years ago
- PyTorch implementation of "Distinguishing Homophenes using Multi-Head Visual-Audio Memory" (AAAI2022)☆26Updated 11 months ago
- Cyclic Co-Learning of Sounding Object Visual Grounding and Sound Separation☆25Updated 3 years ago
- Official Codebase of "Localizing Visual Sounds the Easy Way" (ECCV 2022)☆31Updated 2 years ago
- ABAW3 (CVPRW): A Joint Cross-Attention Model for Audio-Visual Fusion in Dimensional Emotion Recognition☆41Updated last year
- MUSIC-AVQA, CVPR2022 (ORAL)☆75Updated 2 years ago
- Official Codebase of "A Unified Audio-Visual Learning Framework for Localization, Separation, and Recognition" (ICML 2023)☆9Updated last year
- ☆18Updated last month
- Pytorch implementation for Tailor Versatile Multi-modal Learning for Multi-label Emotion Recognition☆56Updated 2 years ago
- ☆12Updated 11 months ago
- GCNet, official pytorch implementation of our paper "GCNet: Graph Completion Network for Incomplete Multimodal Learning in Conversation"☆75Updated last year
- Official implementation for AVGN☆35Updated last year