mpc001 / Visual_Speech_Recognition_for_Multiple_Languages
Visual Speech Recognition for Multiple Languages
☆385Updated last year
Alternatives and similar repositories for Visual_Speech_Recognition_for_Multiple_Languages:
Users that are interested in Visual_Speech_Recognition_for_Multiple_Languages are comparing it to the libraries listed below
- ICASSP'22 Training Strategies for Improved Lip-Reading; ICASSP'21 Towards Practical Lipreading with Distilled and Efficient Models; ICASS…☆408Updated last year
- The PyTorch Code and Model In "Learn an Effective Lip Reading Model without Pains", (https://arxiv.org/abs/2011.07557), which reaches the…☆157Updated last year
- A PyTorch implementation of the Deep Audio-Visual Speech Recognition paper.☆219Updated last year
- A self-supervised learning framework for audio-visual speech☆874Updated last year
- Official Implementation of Visual Transformer Pooling for Lip reading☆39Updated 2 years ago
- Auto-AVSR: Lip-Reading Sentences Project☆310Updated last month
- Audio-Visual Speech Separation with Cross-Modal Consistency☆227Updated last year
- Code and models for evaluating a state-of-the-art lip reading network☆194Updated last year
- A pipeline to read lips and generate speech for the read content, i.e Lip to Speech Synthesis.☆79Updated 3 years ago
- The state-of-art PyTorch implementation of the method described in the paper "LipNet: End-to-End Sentence-level Lipreading" (https://arxi…☆218Updated 2 years ago
- Out of time: automated lip sync in the wild☆723Updated last year
- Disentangled Speech Embeddings using Cross-Modal Self-Supervision☆156Updated 4 years ago
- ACM MM 2021: 'Is Someone Speaking? Exploring Long-term Temporal Features for Audio-visual Active Speaker Detection'☆344Updated last year
- Phoneme Recognition using pre-trained models Wav2vec2, HuBERT and WavLM. Throughout this project, we compared specifically three differen…☆220Updated 2 years ago
- Pytorch code for End-to-End Audiovisual Speech Recognition☆174Updated 2 years ago
- MEAD: A Large-scale Audio-visual Dataset for Emotional Talking-face Generation [ECCV2020]☆256Updated 7 months ago
- [Interspeech 2024] SyncVSR: Data-Efficient Visual Speech Recognition with End-to-End Crossmodal Audio Token Synchronization☆46Updated 2 months ago
- This is the GitHub page for publicly available emotional speech data.☆336Updated 3 years ago
- The repository for IEEE CVPR 2023 (A Light Weight Model for Active Speaker Detection)☆116Updated 10 months ago
- A collection of datasets for the purpose of emotion recognition/detection in speech.☆311Updated 4 months ago
- [WACV 2023] Audio-Visual Efficient Conformer (AVEC) for Robust Speech Recognition☆92Updated 2 years ago
- [CVPR] MARLIN: Masked Autoencoder for facial video Representation LearnINg☆241Updated 2 months ago
- ☆416Updated last year
- PyTorch implementation of "Lip to Speech Synthesis in the Wild with Multi-task Learning" (ICASSP2023)☆65Updated 11 months ago
- Crowd Sourced Emotional Multimodal Actors Dataset (CREMA-D)☆399Updated 2 years ago
- MuAViC: A Multilingual Audio-Visual Corpus for Robust Speech Recognition and Robust Speech-to-Text Translation☆378Updated last year
- This is the main repository of open-sourced speech technology by Huawei Noah's Ark Lab.☆572Updated last year
- [ACL 2024] Official PyTorch code for extracting features and training downstream models with emotion2vec: Self-Supervised Pre-Training fo…☆736Updated last month
- Code for the ICASSP 2022 paper "Transformer-Based Multi-Aspect Multi-Granularity Non-native English Speaker Pronunciation Assessment".☆164Updated 2 years ago
- Official code for the paper "Visual Speech Enhancement Without A Real Visual Stream" published at WACV 2021☆104Updated 8 months ago