razvan404 / multimodal-speech-emotion-recognition
Multimodal SER Model meant to be trained on recognising emotions from speech (text + acoustic data). Fine-tuned the DeBERTaV3 model, respectively the Wav2Vec2 model to extract the features and classify the emotions from the text, respectively audio data, then passed their features and their classification through an MLP to achieve better results…
☆10Updated 10 months ago
Alternatives and similar repositories for multimodal-speech-emotion-recognition:
Users that are interested in multimodal-speech-emotion-recognition are comparing it to the libraries listed below
- Scripts used in the research described in the paper "Multimodal Emotion Recognition with High-level Speech and Text Features" accepted in…☆53Updated 3 years ago
- TensorFlow code implementation of "MultiMAE-DER: Multimodal Masked Autoencoder for Dynamic Emotion Recognition"☆18Updated last month
- ☆11Updated last year
- Chinese BERT classification with tf2.0 and audio classification with mfcc☆14Updated 4 years ago
- IEEE T-BIOM : "Audio-Visual Fusion for Emotion Recognition in the Valence-Arousal Space Using Joint Cross-Attention"☆38Updated 4 months ago
- The code for Multi-Scale Receptive Field Graph Model for Emotion Recognition in Conversations☆10Updated 2 years ago
- A Fully End2End Multimodal System for Fast Yet Effective Video Emotion Recognition☆38Updated 8 months ago
- Code for the InterSpeech 2023 paper: MMER: Multimodal Multi-task learning for Speech Emotion Recognition☆73Updated last year
- This is the official code for paper "Speech Emotion Recognition with Global-Aware Fusion on Multi-scale Feature Representation" published…☆46Updated 3 years ago
- ABAW6 (CVPR-W) We achieved second place in the valence arousal challenge of ABAW6☆19Updated 11 months ago
- The code repository for NAACL 2021 paper "Multimodal End-to-End Sparse Model for Emotion Recognition".☆102Updated 2 years ago
- This repository contains the code for our ICASSP paper `Speech Emotion Recognition using Semantic Information` https://arxiv.org/pdf/2103…☆24Updated 4 years ago
- Group Gated Fusion on Attention-based Bidirectional Alignment for Multimodal Emotion Recognition☆13Updated 2 years ago
- AuxFormer: Robust Approach to Audiovisual Emotion Recognition☆14Updated 2 years ago
- FRAME-LEVEL EMOTIONAL STATE ALIGNMENT METHOD FOR SPEECH EMOTION RECOGNITION☆20Updated 4 months ago
- PyTorch implementation for Audio-Visual Domain Adaptation Feature Fusion for Speech Emotion Recognition☆12Updated 3 years ago
- 多模态,语音和文本结合的情感识别,大模型finetune☆17Updated last year
- [ICASSP 2023] Mingling or Misalignment? Temporal Shift for Speech Emotion Recognition with Pre-trained Representations☆37Updated last year
- [ICASSP 2024] Emotion Neural Transducer for Fine-Grained Speech Emotion Recognition☆23Updated last year
- MultiModal Sentiment Analysis architectures for CMU-MOSEI.☆43Updated 2 years ago
- ☆13Updated 10 months ago
- Multimodal (text, acoustic, visual) Sentiment Analysis and Emotion Recognition on CMU-MOSEI dataset.☆26Updated 4 years ago
- TensorFlow implementation of "Attentive Modality Hopping for Speech Emotion Recognition," ICASSP-20☆32Updated 4 years ago
- ☆59Updated 9 months ago
- Source code for ICASSP 2022 paper "MM-DFN: Multimodal Dynamic Fusion Network For Emotion Recognition in Conversations".☆86Updated 2 years ago
- A survey of deep multimodal emotion recognition.☆52Updated 2 years ago
- This repository provides implementation for the paper "Self-attention fusion for audiovisual emotion recognition with incomplete data".☆133Updated 7 months ago
- The PyTorch code for paper: "CONSK-GCN: Conversational Semantic- and Knowledge-Oriented Graph Convolutional Network for Multimodal Emotio…☆11Updated 2 years ago
- Implementation of the paper "Multimodal Transformer With Learnable Frontend and Self Attention for Emotion Recognition" submitted to ICAS…☆24Updated 3 years ago
- Multimodal Transformer for Korean Sentiment Analysis with Audio and Text Features☆28Updated 3 years ago