toxtli / AutomEditor
AutomEditor is an AI based video editor that helps video bloggers to remove bloopers automatically. It uses multimodal spatio-temporal blooper recognition and localization approaches. The models were trained in keras and integrate feature fusion techniques from face, body gestures (skelethon), emotions progression, and audio features
☆48Updated 6 years ago
Alternatives and similar repositories for AutomEditor
Users that are interested in AutomEditor are comparing it to the libraries listed below
Sorting:
- You Said That?: Synthesising Talking Faces from Audio☆69Updated 7 years ago
- Final Project for Stanford Deep Generative Modeling Class CS236.☆14Updated 5 years ago
- Learning Lip Sync of Obama from Speech Audio☆67Updated 4 years ago
- A Python Library for Multimodal Analysis of Movies and Content-based Movie Recommendation☆29Updated 3 years ago
- Supporting code for "Emotion Recognition in Speech using Cross-Modal Transfer in the Wild"☆102Updated 5 years ago
- ☆34Updated 6 years ago
- ☆21Updated 4 years ago
- mirror of VoxCeleb dataset - a large-scale speaker identification dataset☆71Updated 5 years ago
- Speech-conditioned face generation using Generative Adversarial Networks☆88Updated 2 years ago
- Repository for th OMG Emotion Challenge☆89Updated 4 months ago
- Accompany code to reproduce the baselines of the International Multimodal Sentiment Analysis Challenge (MuSe 2020).☆16Updated 2 years ago
- ☆110Updated 2 years ago
- This is the repository containing the solution for FG-2020 ABAW Competition☆117Updated last year
- This repository is a repository for the paper, "Irgun: Improved residue based gradual up-scaling network for single image super resolutio…☆14Updated 4 years ago
- Human Emotion Understanding using multimodal dataset.☆97Updated 4 years ago
- Repository with the code of the paper: A proposal for Multimodal Emotion Recognition using auraltransformers and Action Units on RAVDESS …☆104Updated last year
- Official code for the paper "Visual Speech Enhancement Without A Real Visual Stream" published at WACV 2021☆107Updated 11 months ago
- Multi-modality Hierarchical Recall based on GBDTs for Bipolar Disorder Classification☆10Updated last year
- The code for our INTERSPEECH 2020 paper - Jointly Fine-Tuning "BERT-like'" Self Supervised Models to Improve Multimodal Speech Emotion R…☆120Updated 4 years ago
- ☆89Updated 6 years ago
- Generating Talking Face Landmarks from Speech☆159Updated 2 years ago
- AI Talking Head: create video from plain text or audio file in minutes, support up to 100+ languages and 350+ voice models.☆35Updated 2 years ago
- An automatic movie trailer generator.☆41Updated 2 years ago
- Live demo for speech emotion recognition using Keras and Tensorflow models☆39Updated 9 months ago
- A Pytorch implementation of emotion recognition from videos☆18Updated 4 years ago
- Unsupervised Any-to-many Audiovisual Synthesis via Exemplar Autoencoders☆121Updated 2 years ago
- Predicting various emotion in human speech signal by detecting different speech components affected by human emotion.☆46Updated 9 months ago
- Turning films into structured data to unlock the vast wealth of emotional knowledge within.☆30Updated 3 years ago
- Implementation of the paper "Improved End-to-End Speech Emotion Recognition Using Self Attention Mechanism and Multitask Learning" From I…☆57Updated 4 years ago
- Real-Time Lip Sync for Live 2D Animation☆139Updated 5 years ago