joannahong / Lip2Wav-pytorch
a PyTorch implementation of Lip2Wav
☆50Updated 2 years ago
Alternatives and similar repositories for Lip2Wav-pytorch:
Users that are interested in Lip2Wav-pytorch are comparing it to the libraries listed below
- PyTorch implementation of "Lip to Speech Synthesis with Visual Context Attentional GAN" (NeurIPS2021)☆23Updated last year
- ☆55Updated last year
- The speaker-labeled information of LRW dataset, which is the outcome of the paper "Speaker-adaptive Lip Reading with User-dependent Paddi…☆10Updated last year
- Official implementation of RAVEn (ICLR 2023) and BRAVEn (ICASSP 2024)☆63Updated 2 months ago
- [Interspeech 2023] Intelligible Lip-to-Speech Synthesis with Speech Units☆35Updated 6 months ago
- A pipeline to read lips and generate speech for the read content, i.e Lip to Speech Synthesis.☆83Updated 3 years ago
- Official PyTorch implementation of paper Leveraging Unimodal Self Supervised Learning for Multimodal Audio-Visual Speech Recognition (ACL…☆65Updated 2 years ago
- Tools for downloading VoxCeleb2 dataset☆29Updated last year
- Audio-Visual Corruption Modeling of our paper "Watch or Listen: Robust Audio-Visual Speech Recognition with Visual Corruption Modeling an…☆33Updated last year
- Speech Representation Disentanglement with Adversarial Mutual Information Learning for One-shot Voice Conversion (Interspeech 2022)☆116Updated last year
- ☆22Updated last year
- Baseline system for CNVSRC2023 (Chinese Continuous Visual Speech Recognition Challenge 2023)☆22Updated last year
- ☆46Updated 2 years ago
- [CVPR 2024] AV2AV: Direct Audio-Visual Speech to Audio-Visual Speech Translation with Unified Audio-Visual Speech Representation☆35Updated 8 months ago
- PyTorch implementation of "Lip to Speech Synthesis in the Wild with Multi-task Learning" (ICASSP2023)☆66Updated last year
- This is the implementation of the paper "Emotion Intensity and its Control for Emotional Voice Conversion".☆88Updated 3 years ago
- ☆162Updated 10 months ago
- Look Who’s Talking: Active Speaker Detection in the Wild☆72Updated last year
- PyTorch implementation of "Distinguishing Homophenes using Multi-Head Visual-Audio Memory" (AAAI2022)☆27Updated last year
- ☆16Updated 5 months ago
- Pytorch implementation for “V2C: Visual Voice Cloning”☆32Updated 2 years ago
- INTERSPEECH2023: Target Active Speaker Detection with Audio-visual Cues☆51Updated last year
- Official repository for the paper VocaLiST: An Audio-Visual Synchronisation Model for Lips and Voices☆66Updated last year
- Disentangled Speech Embeddings using Cross-Modal Self-Supervision☆159Updated 5 years ago
- [Interspeech 2024] SyncVSR: Data-Efficient Visual Speech Recognition with End-to-End Crossmodal Audio Token Synchronization☆49Updated last month
- Emotion Rendering for Conversational Speech Synthesis with Heterogeneous Graph-Based Context Modeling (Accepted by AAAI'2024)☆56Updated 10 months ago
- [INTERSPEECH 2022] This dataset is designed for multi-modal speaker diarization and lip-speech synchronization in the wild.☆52Updated last year
- This is the code for controllable EVC framework for seen and unseen emotion generation.☆44Updated 3 years ago
- An unofficial (PyTorch) implementation for the paper Deep Lip Reading: A comparison of models and an online application.☆10Updated 4 years ago
- The Official Implementation of “Content-Dependent Fine-Grained Speaker Embedding for Zero-Shot Speaker Adaptation in Text-to-Speech Synth…☆85Updated 2 years ago