CheyneyComputerScience / CREMA-D
Crowd Sourced Emotional Multimodal Actors Dataset (CREMA-D)
☆366Updated 2 years ago
Related projects ⓘ
Alternatives and complementary repositories for CREMA-D
- A collection of datasets for the purpose of emotion recognition/detection in speech.☆290Updated last month
- Python package for openSMILE☆245Updated 7 months ago
- [ICASSP 2023] Official Tensorflow implementation of "Temporal Modeling Matters: A Novel Temporal Emotional Modeling Approach for Speech E…☆159Updated 5 months ago
- Official implementation for the paper Exploring Wav2vec 2.0 fine-tuning for improved speech emotion recognition☆142Updated 3 years ago
- Official implementation of INTERSPEECH 2021 paper 'Emotion Recognition from Speech Using Wav2vec 2.0 Embeddings'☆126Updated 2 years ago
- Auto-AVSR: Lip-Reading Sentences Project☆173Updated 6 months ago
- This is the GitHub page for publicly available emotional speech data.☆320Updated 2 years ago
- A PyTorch implementation of the Deep Audio-Visual Speech Recognition paper.☆207Updated 8 months ago
- A multimodal approach on emotion recognition using audio and text.☆158Updated 4 years ago
- A Cooperative Voice Analysis Repository for Speech Technologies☆350Updated 4 years ago
- TensorFlow implementation of "Multimodal Speech Emotion Recognition using Audio and Text," IEEE SLT-18☆258Updated 4 months ago
- Audio-Visual Speech Separation with Cross-Modal Consistency☆220Updated last year
- Wav2Vec for speech recognition, classification, and audio classification☆249Updated 2 years ago
- The code for our INTERSPEECH 2020 paper - Jointly Fine-Tuning "BERT-like'" Self Supervised Models to Improve Multimodal Speech Emotion R…☆113Updated 3 years ago
- Lightweight and Interpretable ML Model for Speech Emotion Recognition and Ambiguity Resolution (trained on IEMOCAP dataset)☆399Updated 10 months ago
- ☆97Updated 2 years ago
- Code for the InterSpeech 2023 paper: MMER: Multimodal Multi-task learning for Speech Emotion Recognition☆65Updated 7 months ago
- ACM MM 2021: 'Is Someone Speaking? Exploring Long-term Temporal Features for Audio-visual Active Speaker Detection'☆312Updated last year
- ☆129Updated 2 months ago
- Multilingual datasets with raw audio for speech emotion recognition☆19Updated 3 years ago
- Deep-Learning-Based Audio-Visual Speech Enhancement and Separation☆203Updated last year
- ☆46Updated 11 months ago
- Multi-modal Speech Emotion Recogniton on IEMOCAP dataset☆84Updated last year
- This repository contains PyTorch implementation of 4 different models for classification of emotions of the speech.☆191Updated last year
- Repository with the code of the paper: A proposal for Multimodal Emotion Recognition using auraltransformers and Action Units on RAVDESS …☆96Updated 7 months ago
- Code and models for evaluating a state-of-the-art lip reading network☆189Updated last year
- Speaker embedding (d-vector) trained with GE2E loss☆273Updated 10 months ago
- Official implement of SpeechFormer written in Python (PyTorch).☆75Updated last year
- Code for Speech Emotion Recognition with Co-Attention based Multi-level Acoustic Information☆125Updated 11 months ago
- INTERSPEECH 2023-2024 Papers: A complete collection of influential and exciting research papers from the INTERSPEECH 2023-24 conference. …☆638Updated 3 months ago