toxtli / AutomEditorLinks
AutomEditor is an AI based video editor that helps video bloggers to remove bloopers automatically. It uses multimodal spatio-temporal blooper recognition and localization approaches. The models were trained in keras and integrate feature fusion techniques from face, body gestures (skelethon), emotions progression, and audio features
☆48Updated 6 years ago
Alternatives and similar repositories for AutomEditor
Users that are interested in AutomEditor are comparing it to the libraries listed below
Sorting:
- Final Project for Stanford Deep Generative Modeling Class CS236.☆13Updated 6 years ago
- Deepstory turns a text/generated text into a video where the character is animated to speak your story using his/her voice.☆102Updated 2 months ago
- Neural Voice Cloning with a few voice samples, using the speaker adaptation method. Speaker adaptation is based on fine-tuning a multi-sp…☆58Updated 6 years ago
- This repository is a repository for the paper, "Irgun: Improved residue based gradual up-scaling network for single image super resolutio…☆15Updated 5 years ago
- This repository is an implementation of Transfer Learning from Speaker Verification to Multispeaker Text-To-Speech Synthesis (SV2TTS) wit…☆171Updated 5 years ago
- Video chat apps with computer vision filters built on top of Streamlit☆50Updated 2 years ago
- Identify the emotion of multiple speakers in an Audio Segment☆178Updated 2 years ago
- ☆25Updated 7 years ago
- You Said That?: Synthesising Talking Faces from Audio☆70Updated 7 years ago
- Automated Lip reading from real-time videos in tensorflow in python☆162Updated 7 years ago
- A deep learning model to lip-sync a given video with any given audio. It uses GAN architecture to orchestrate loss reconstruction or trai…☆123Updated 2 years ago
- Learning Lip Sync of Obama from Speech Audio☆66Updated 5 years ago
- Speech Emotion Detection using SVM, Decision Tree, Random Forest, MLP, CNN with different architectures☆39Updated 2 years ago
- Generalized cross-modal NNs; new audiovisual benchmark (IEEE TNNLS 2019)☆30Updated 5 years ago
- Turning films into structured data to unlock the vast wealth of emotional knowledge within.☆30Updated 3 years ago
- Speech to Facial Animation using GANs☆40Updated 4 years ago
- Real-time speech to text with specific language translation.☆47Updated 5 years ago
- ☆34Updated 7 years ago
- Code to detect scenes and transitions in videos and compose a video to visualize the data.☆28Updated 7 years ago
- AI Talking Head: create video from plain text or audio file in minutes, support up to 100+ languages and 350+ voice models.☆37Updated 3 years ago
- Official code for the paper "Visual Speech Enhancement Without A Real Visual Stream" published at WACV 2021☆108Updated last year
- Converter Video to Anime. Based on "AnimeGAN" repository.☆38Updated 5 years ago
- Emotion recognition library for PyTorch☆22Updated 5 years ago
- Predicting various emotion in human speech signal by detecting different speech components affected by human emotion.☆47Updated last year
- AI-generated talking head video of fake people responding to your input question text.☆68Updated 4 years ago
- Speech-conditioned face generation using Generative Adversarial Networks☆88Updated 3 years ago
- Gaze estimation from 2D image☆12Updated last year
- Unsupervised Any-to-many Audiovisual Synthesis via Exemplar Autoencoders☆122Updated 3 years ago
- Automatically generate a lip-synced avatar based off of a transcript and audio☆14Updated 2 years ago
- Allows you to edit videos automatically☆45Updated 8 months ago