Spoken language identification refers to the task of automatically predi...
We introduce MADLAD-400, a manually audited, general domain 3T token
mon...
We introduce AudioPaLM, a large language model for speech understanding ...
Speech representation learning approaches for non-semantic tasks such as...
This paper introduces a new speech dataset called “LibriTTS-R” designed ...
Recently, a number of approaches to train speech models by incorpo-ratin...
Speech restoration (SR) is a task of converting degraded speech signals ...
We introduce the Universal Speech Model (USM), a single large model that...
We present Mu^2SLAM, a multilingual sequence-to-sequence model
pre-train...
This paper proposes Virtuoso, a massively multilingual speech-text joint...
Training state-of-the-art Automated Speech Recognition (ASR) models typi...
We present JOIST, an algorithm to train a streaming, cascaded, encoder
e...
Much of text-to-speech research relies on human evaluation, which incurs...
We introduce FLEURS, the Few-shot Learning Evaluation of Universal
Repre...
In this paper we share findings from our effort to build practical machi...
We present Maestro, a self-supervised training method to unify
represent...
End-to-end speech-to-speech translation (S2ST) without relying on
interm...
We introduce XTREME-S, a new benchmark to evaluate universal cross-lingu...
Multilingual neural machine translation models are trained to maximize t...
We present mSLAM, a multilingual Speech and LAnguage Model that learns
c...
Natural language understanding and generation models follow one of the t...
Achieving universal translation between all human language pairs is the
...
Self-supervised training has shown promising gains in pretraining models...
Unsupervised pre-training is now the predominant approach for both text ...
Document-level neural machine translation (DocNMT) delivers coherent
tra...
We present an empirical study of scaling properties of encoder-decoder
T...
With the success of large-scale pre-training and multilingual modeling i...
To mitigate the negative effect of low quality training data on the
perf...
Large text corpora are increasingly important for a wide variety of Natu...
Over the last few years two promising research directions in low-resourc...
Most neural networks utilize the same amount of compute for every exampl...
Motivated by the fact that most of the information relevant to the predi...
Most neural machine translation systems still translate sentences in
iso...
Fine-tuning pre-trained Neural Machine Translation (NMT) models is the
d...
Multilingual end-to-end (E2E) models have shown great promise in expansi...
Multilingual Neural Machine Translation (NMT) models have yielded large
...
The recently proposed massively multilingual neural machine translation ...
We introduce our efforts towards building a universal neural machine
tra...
Multilingual Neural Machine Translation (NMT) models are capable of
tran...
Neural Networks trained with gradient descent are known to be susceptibl...
Lingvo is a Tensorflow framework offering a complete solution for
collab...
Translating characters instead of words or word-fragments has the potent...
While current state-of-the-art NMT models, such as RNN seq2seq and
Trans...
The past year has witnessed rapid advances in sequence-to-sequence (seq2...
We propose Machines Talking To Machines (M2M), a framework combining
aut...
State-of-the-art slot filling models for goal-oriented human/machine
con...
Spoken Language Understanding (SLU) is a key component of goal oriented
...