An advanced singing voice synthesis system with high fidelity, expressiveness, controllability and flexibility based on DiffSinger: Singing Voice Synthesis via Shallow Diffusion Mechanism
-
Updated
Feb 17, 2025 - Python
An advanced singing voice synthesis system with high fidelity, expressiveness, controllability and flexibility based on DiffSinger: Singing Voice Synthesis via Shallow Diffusion Mechanism
Command line utility for forced alignment using Kaldi
My-Voice Analysis is a Python library for the analysis of voice (simultaneous speech, high entropy) without the need of a transcription. It breaks utterances and detects syllable boundaries, fundamental frequency contours, and formants.
A Python library for measuring the acoustic features of speech (simultaneous speech, high entropy) compared to ones of native speech.
PyTorch implementation of the Factorized TDNN (TDNN-F) from "Semi-Orthogonal Low-Rank Matrix Factorization for Deep Neural Networks" and Kaldi
Foreign Accent Conversion by Synthesizing Speech from Phonetic Posteriorgrams (Interspeech'19)
Official code for "Learning Neural Acoustic Fields" (NeurIPS 2022)
[AAAI 2024] CTX-txt2vec, the acoustic model in UniCATS
The official code for the SALMon🍣 benchmark (ICASSP 2025)
🎵 A repository for manually annotating files to create labeled acoustic datasets for machine learning.
SC-CNN: Effective Speaker Conditioning Method for Zero-Shot Multi-Speaker Text-to-Speech Systems
PyTorch implementation of automatic speech recognition models.
Code for the paper: Audio to Score Matching by Combining Phonetic and Duration Information
Automated, end-to-end wakeword model maker using the Precise Wakeword Engine
A Simple Automatic Speech Recognition (ASR) Model in Tensorflow, which only needs to focus on Deep Neural Network. It's easy to test popular cells (most are LSTM and its variants) and models (unidirectioanl RNN, bidirectional RNN, ResNet and so on). Moreover, you are welcome to play with self-defined cells or models.
It is an algorithm analysed the acoustic features of a voice and creates an acoustic classifier - USEFUL for auto-speech-rater
Code for: "Leveraging Sound and Wrist Motion to Detect Activities of Daily Living with Commodity Smartwatches"
It is an algorithm analysed the acoustic features of a voice and creates an acoustic classifier - USEFUL for auto-speech-rater
Repository of an implementation of the matrix method for acoustic levitation simulations.
This is a sub-repository in building to create acoustic model in Mandarin speech recognition.
Add a description, image, and links to the acoustic-model topic page so that developers can more easily learn about it.
To associate your repository with the acoustic-model topic, visit your repo's landing page and select "manage topics."