Editing
VITS
(section)
Jump to navigation
Jump to search
Warning:
You are not logged in. Your IP address will be publicly visible if you make any edits. If you
log in
or
create an account
, your edits will be attributed to your username, along with other benefits.
Anti-spam check. Do
not
fill this in!
== Technical Architecture == === VITS (2021) === VITS employs a conditional variational autoencoder (VAE) framework combined with several advanced techniques: '''Core Components:''' * '''Posterior Encoder:''' Processes linear-scale spectrograms during training to learn latent representations * '''Prior Encoder:''' Contains a text encoder and normalizing flows to model the prior distribution of latent variables * '''Decoder:''' Based on HiFi-GAN V1 generator, converts latent variables to raw waveforms * '''Discriminator:''' Multi-period discriminator from HiFi-GAN for adversarial training '''Key Innovations:''' * '''Monotonic Alignment Search (MAS):''' Automatically learns alignments between text and speech without external annotations by finding alignments that maximize the likelihood of target speech. * '''Stochastic Duration Predictor:''' Uses normalizing flows to model the distribution of phoneme durations, enabling synthesis of speech with diverse rhythms from the same text input. * '''Adversarial Training:''' Improves waveform quality through generator-discriminator competition The model addresses the one-to-many relationship in speech synthesis, where a single text input can be spoken in multiple ways with different pitches, rhythms, and prosodic patterns. === VITS2 (2023) === VITS2 introduced several improvements over the original model to address issues including intermittent unnaturalness, computational efficiency, and strong dependence on phoneme conversion. '''Major Improvements:''' * '''Adversarial Duration Predictor:''' Replaced the flow-based stochastic duration predictor with one trained through adversarial learning, using a time step-wise conditional discriminator to improve efficiency and naturalness. * '''Enhanced Normalizing Flows:''' Added transformer blocks to normalizing flows to capture long-term dependencies when transforming distributions, addressing limitations of convolution-only approaches. * '''Improved Alignment Search:''' Modified Monotonic Alignment Search by adding Gaussian noise to calculated probabilities, giving the model additional opportunities to explore alternative alignments during early training. * '''Speaker-Conditioned Text Encoder:''' For multi-speaker models, conditioned the speaker vector on the third transformer block of the text encoder to better capture speaker-specific pronunciation and intonation characteristics.
Summary:
Please note that all contributions to TTS Wiki are considered to be released under the Creative Commons Attribution 4.0 (see
Project:Copyrights
for details). If you do not want your writing to be edited mercilessly and redistributed at will, then do not submit it here.
You are also promising us that you wrote this yourself, or copied it from a public domain or similar free resource.
Do not submit copyrighted work without permission!
Cancel
Editing help
(opens in new window)
Navigation menu
Personal tools
Not logged in
Talk
Contributions
Create account
Log in
Namespaces
Page
Discussion
English
Views
Read
Edit
Edit source
View history
More
Search
Navigation
Main page
Recent changes
Random page
Help about MediaWiki
Tools
What links here
Related changes
Special pages
Page information