🎯 Speech Recognition Challenge by Speech Lab - IIT Madras
☆11Nov 5, 2020Updated 5 years ago
Alternatives and similar repositories for Hindi-ASR-Challenge
Users that are interested in Hindi-ASR-Challenge are comparing it to the libraries listed below
Sorting:
- The project is related to the development of labs for the ITMO Speaker Recognition Course.☆15Feb 3, 2026Updated last month
- This repo contains the baseline model recipes and pre-trained model for GramVanni hindi ASR challenge☆15Mar 26, 2022Updated 3 years ago
- EC499: Major Project☆10Jun 25, 2023Updated 2 years ago
- English ASR Challenge organized by Speech Lab, IIT Madras☆11Feb 3, 2021Updated 5 years ago
- A pipeline to isolate and transcribe one language in mixed-language speech☆20Oct 25, 2022Updated 3 years ago
- A set of tools for working with accent data in Mozilla's Common Voice dataset☆14Nov 3, 2023Updated 2 years ago
- Indic-Conformer models for ASR☆21Jul 19, 2024Updated last year
- ☆13Dec 15, 2022Updated 3 years ago
- The first Dialectal Arabic Code Switching - DACS corpus from broadcast speech. Annotated at the token-level, considering both the linguis…☆15Apr 3, 2022Updated 3 years ago
- Hosts text-to-speech corpus and speech synthesizers for African languages.☆18May 31, 2023Updated 2 years ago
- ☆45Dec 15, 2022Updated 3 years ago
- Breaks a word into syllables using an LSTM-based neural network.☆20Aug 14, 2023Updated 2 years ago
- Long audio alignment using Kaldi☆23Apr 22, 2021Updated 4 years ago
- SylNet: An Adaptable End-to-End Syllable Count Estimator for Speech☆27May 25, 2023Updated 2 years ago
- Vecna is a Python chatbot which recommends songs and movies depending upon your feelings☆12Jun 28, 2022Updated 3 years ago
- Compendium for the paper "Transparent pronunciation scoring using articulatorily weighted phoneme edit distance" by Karhila, Smolander, Y…☆25May 6, 2019Updated 6 years ago
- Transcribe your videos and translate it into Indic languages.☆31Feb 13, 2026Updated 3 weeks ago
- Syllable Segmentation and Cross-Lingual Generalization in a Visually Grounded, Self-Supervised Speech Model