razvan404 / multimodal-speech-emotion-recognitionLinks
Multimodal SER Model meant to be trained on recognising emotions from speech (text + acoustic data). Fine-tuned the DeBERTaV3 model, respectively the Wav2Vec2 model to extract the features and classify the emotions from the text, respectively audio data, then passed their features and their classification through an MLP to achieve better results…
☆10Updated last year
Alternatives and similar repositories for multimodal-speech-emotion-recognition
Users that are interested in multimodal-speech-emotion-recognition are comparing it to the libraries listed below
Sorting:
- The code repository for NAACL 2021 paper "Multimodal End-to-End Sparse Model for Emotion Recognition".☆104Updated 2 years ago
- Scripts used in the research described in the paper "Multimodal Emotion Recognition with High-level Speech and Text Features" accepted in…☆53Updated 3 years ago
- The code for our INTERSPEECH 2020 paper - Jointly Fine-Tuning "BERT-like'" Self Supervised Models to Improve Multimodal Speech Emotion R…☆120Updated 4 years ago
- Source code for ICASSP 2022 paper "MM-DFN: Multimodal Dynamic Fusion Network For Emotion Recognition in Conversations".☆89Updated 2 years ago
- ☆13Updated last year
- [IEEE ICPRS 2024 Oral] TensorFlow code implementation of "MultiMAE-DER: Multimodal Masked Autoencoder for Dynamic Emotion Recognition"☆19Updated this week
- The code for Multi-Scale Receptive Field Graph Model for Emotion Recognition in Conversations☆10Updated 2 years ago
- CM-BERT: Cross-Modal BERT for Text-Audio Sentiment Analysis(MM2020)☆113Updated 4 years ago
- AuxFormer: Robust Approach to Audiovisual Emotion Recognition☆14Updated 2 years ago
- ABAW6 (CVPR-W) We achieved second place in the valence arousal challenge of ABAW6☆25Updated last year
- Code for the InterSpeech 2023 paper: MMER: Multimodal Multi-task learning for Speech Emotion Recognition☆75Updated last year
- IEEE T-BIOM : "Audio-Visual Fusion for Emotion Recognition in the Valence-Arousal Space Using Joint Cross-Attention"☆41Updated 8 months ago
- Multimodal (text, acoustic, visual) Sentiment Analysis and Emotion Recognition on CMU-MOSEI dataset.☆27Updated 4 years ago
- Code for Speech Emotion Recognition with Co-Attention based Multi-level Acoustic Information☆148Updated last year
- ☆17Updated last year
- Group Gated Fusion on Attention-based Bidirectional Alignment for Multimodal Emotion Recognition☆14Updated 3 years ago
- A Fully End2End Multimodal System for Fast Yet Effective Video Emotion Recognition☆40Updated 11 months ago
- ☆65Updated last year
- Chinese BERT classification with tf2.0 and audio classification with mfcc☆14Updated 4 years ago
- This repository provides implementation for the paper "Self-attention fusion for audiovisual emotion recognition with incomplete data".☆143Updated 10 months ago
- This is the official code for paper "Speech Emotion Recognition with Global-Aware Fusion on Multi-scale Feature Representation" published…☆47Updated 3 years ago
- The PyTorch code for paper: "CONSK-GCN: Conversational Semantic- and Knowledge-Oriented Graph Convolutional Network for Multimodal Emotio…☆13Updated 2 years ago
- Implementation of the paper "Multimodal Transformer With Learnable Frontend and Self Attention for Emotion Recognition" submitted to ICAS…☆24Updated 3 years ago
- This repository contains the code for our ICASSP paper `Speech Emotion Recognition using Semantic Information` https://arxiv.org/pdf/2103…☆24Updated 4 years ago
- Detect emotion from audio signals of IEMOCAP dataset using multi-modal approach. Utilized acoustic features, mel-spectrogram and text as …☆39Updated last year
- A multimodal approach on emotion recognition using audio and text.☆182Updated 5 years ago
- The code for our IEEE ACCESS (2020) paper Multimodal Emotion Recognition with Transformer-Based Self Supervised Feature Fusion.☆120Updated 3 years ago
- MultiModal Sentiment Analysis architectures for CMU-MOSEI.☆46Updated 2 years ago
- [EMNLP2023] Conversation Understanding using Relational Temporal Graph Neural Networks with Auxiliary Cross-Modality Interaction☆63Updated last year
- ☆92Updated 2 years ago