TRANSLATOTRON 2
Translatotron 2 is a robust direct speech-to-speech translation (S2ST) model from Google AI. It is an improved version of Translatotron, the world's first S2ST model that can translate speech directly between two languages and also retained the source speaker's voice in the translated speech
Translatotron 2 uses a novel method for transferring the source speakers’ voices to the translated speech, which works well even for input speech containing multiple speakers speaking in turns and also reduced the potential for misuse. Translatotron 2 outperforms Translatotron by a large margin on translation quality, speech naturalness, and speech robustness.
Need for a better S2ST
Speech-to-speech translation (S2ST) helps solve the language barrier problems between people all over the world. Automatic S2ST systems consist of a cascade of speech recognition, machine translation, and speech synthesis subsystems. But they suffered from longer latency, loss of information ( paralinguistic and non-linguistic information), and compounding errors between subsystems.
Translatotron introduced by Google AI in 2019 consisted of a single attentive sequence-to-sequence model for S2ST, trained end to end and was independent of the intermediate text representation. But Translatotron's performance was not on par with a strong baseline cascade S2ST system (e.g., composed of S2ST model-1, model-2, followed by a Tacotron 2 TTS model).
Translatotron 2 - Components
Translatotron 2 has 4 components :
Speech encoder
Target phoneme decoder
Target speech synthesizer
Attention module that connects them together
The combination of the encoder, the attention module, and the decoder is similar to a direct speech-to-text translation (ST) model.
The synthesizer is conditioned on the output from both the decoder and the attention.
3 Key Improvements made in Translatotron 2
1.The output from the target phoneme decoder is :
Used only as an auxiliary loss in Translatotron
Used as one of the inputs to the spectrogram synthesizer in Translatotron 2 making it easier to train with better performance
2. The spectrogram synthesizer in the original is:
Attention-based (like Tacotron 2 TTS model) in Translatotron and thus suffers from the robustness issues exhibited by Tacotron 2.
Duration-based (like Non-Attentive Tacotron) in Translatotron 2, and thus the robustness of the synthesized speech is improved.
3. Attention-based connection to the encoded source speech is driven by the:
Spectrogram synthesizer in Translatotron
Phoneme decoder in Translatotron 2. Thus, the acoustic information that the spectrogram synthesizer sees is aligned with the translated content that it’s synthesizing and hence each speaker’s voice is retained across speaker turns.
VOICE RETENTION
Responsible Voice Retention
Translatotron could retain the source speaker's voice in the translated speech, by conditioning its decoder on a speaker embedding, generated from a separately trained speaker encoder. This enabled Translatotron to generate the translated speech in a different speaker's voice (if a clip of the target speaker's recording were used as the reference audio to the speaker encoder, or if the embedding of the target speaker were directly available). Though this was a powerful approach, it could be misused to spoof audio with arbitrary content. Hence it was a major concern to deploy the model in production.
But Translatotron 2 uses only a single speech encoder, which is responsible for both linguistic understanding and voice capture. Thus, the model cannot be misused to reproduce non-source voices. This approach can also be applied to Translatotron.
Voice Retention Dataset
To retain speakers' voices across translation, researchers usually train S2ST models on parallel utterances with the same speaker's voice on both sides. This type of dataset with human recordings on both sides requires a large number of fluent bilingual speakers & hence difficult to collect
PnG NAT is a TTS (Text to Speech Synthesis) model which can do cross-lingual voice transferring to synthesize such training targets. The modified PnG NAT model used in Translatotron 2 incorporates a separately trained speaker encoder in the same way as in Translatotron and hence capable of zero-shot voice transference and thus the bilingual dataset problem is solved.
Individual Voice Retention with ConcatAug
For input speech with multiple speakers speaking in turns, each speaker’s voice in the translated speech can be retained by the S2ST models by a novel technique called ConcatAug.
ConcatAug augments the training data on the fly by:
1.Randomly sampling pairs of training examples and
2. Forming new training examples by concatenating the
source speech
target speech
target phoneme sequences
The resulting samples contain two speakers’ voices in both the source and the target speech.Thus the model can learn on examples with speaker turns.
Performance
Translatotron 2 outperforms Translatotron by large margins, such as :
Higher translation quality (measured by BLEU, where higher is better)
Speech naturalness (measured by MOS, higher is better)
Speech robustness (measured by UDR, lower is better)
It excelled on the more difficult Fisher corpus.
The translation quality and speech quality of Translatotron 2 is on par with strong baseline cascade system
Its speech robustness is better than the cascade baseline
Translation quality (measured by BLEU, where higher is better) evaluated on 2 Spanish-English corpora
Speech naturalness (measured by MOS, where higher is better) evaluated on 2 Spanish-English corpora
Speech robustness (measured by UDR, where lower is better) evaluated on 2 Spanish-English corpora
Multilingual S2ST
In a multilingual set-up, Translatotron 2 model took speech input from 4 languages and translated them into English. The language of the input speech was intentionally not provided, forcing the model to detect the language by itself. It outperformed Translatotron by a large margin. The results are not directly comparable between S2ST and ST, but the close numbers suggest that its translation quality is comparable to a baseline speech-to-text translation model. Thus, Translatotron 2 performs well on multilingual S2ST.
Hope you are doing well ... Pleasure meeting you online ...
I am Sri Lakshmi , AI Practitioner, Developer & Technical Content Producer
Liked this article ??? If you want me to : Write articles that give simple explanations of complex topics / Design outstanding presentations / Develop cool AI apps / Launch and popularize your products to the target audience / Manage social media and digital presence / Partner or Collaborate with me, feel free to discuss with me your ideas & requirements by clicking the button below
Further Reading: View the Translatotron 2 paper here
Update your knowledge with my other interesting Articles