Unspeech: Unsupervised Speech Context Embeddings

04/18/2018
by   Benjamin Milde, et al.
0

We introduce "Unspeech" embeddings, which are based on unsupervised learning of context feature representations of spoken language. The embeddings were trained on up to 9500 hours of crawled English speech data without transcriptions or speaker information, by using a straightforward learning objective based on context and non-context discrimination with negative sampling. We use a Siamese convolutional neural network architecture to train Unspeech embeddings and evaluate them on speaker comparison, utterance clustering and as a context feature in TDNN-HMM acoustic models trained on TED-LIUM, comparing it to i-vector baselines. Particularly decoding out-of-domain speech data from the recently released Common Voice corpus shows consistent WER reductions. We release our source code and pre-trained Unspeech models under a permissive open source license.

READ FULL TEXT

Please sign up or login with your details

Forgot password? Click here to reset