ERISHA is a mulitilingual multispeaker expressive speech synthesis framework. It can transfer the expressivity to the speaker's voice for which no expressive speech corpus is available.
☆43Dec 17, 2020Updated 5 years ago
Alternatives and similar repositories for ERISHA
Users that are interested in ERISHA are comparing it to the libraries listed below
Sorting:
- ☆69Mar 31, 2021Updated 4 years ago
- LVCNet: Efficient Condition-Dependent Modeling Network for Waveform Generation☆80Feb 24, 2021Updated 5 years ago
- Official PyTorch implementation of Speaker Conditional WaveRNN☆110Jun 22, 2022Updated 3 years ago
- Compressed version of Tacotron 2 using Tensor Train + Waveglow.☆22Dec 26, 2019Updated 6 years ago
- ☆45Dec 16, 2019Updated 6 years ago
- This repository provides a multi-mode and multi-speaker expressive speech synthesis framework, including multi-attentive Tacotron, DurIAN…☆74Sep 21, 2022Updated 3 years ago
- ☆37May 8, 2021Updated 4 years ago
- This is the implementation of the Speaker Odyssey 2020 paper " Transforming spectrum and prosody for emotional voice conversion with non-…☆125Dec 14, 2020Updated 5 years ago
- (R&D) Text to speech using phonemes as inputs and audio codec codes as outputs. Loosely based on MegaByte, VALL-E and Encodec.