Generating Sentences from a Continuous Space

11/19/2015
by   Samuel R. Bowman, et al.
0

The standard recurrent neural network language model (RNNLM) generates sentences one word at a time and does not work from an explicit global sentence representation. In this work, we introduce and study an RNN-based variational autoencoder generative model that incorporates distributed latent representations of entire sentences. This factorization allows it to explicitly model holistic properties of sentences such as style, topic, and high-level syntactic features. Samples from the prior over these sentence representations remarkably produce diverse and well-formed sentences through simple deterministic decoding. By examining paths through this latent space, we are able to generate coherent novel sentences that interpolate between known sentences. We present techniques for solving the difficult learning problem presented by this model, demonstrate its effectiveness in imputing missing words, explore many interesting properties of the model's latent sentence space, and present negative results on the use of the model in language modeling.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
02/09/2018

Recurrent Neural Network-Based Semantic Variational Autoencoder for Sequence-to-Sequence Learning

Sequence-to-sequence (Seq2seq) models have played an import role in the ...
research
06/13/2018

Generating Sentences Using a Dynamic Canvas

We introduce the Attentive Unsupervised Text (W)riter (AUTR), which is a...
research
09/26/2017

Generating Sentences by Editing Prototypes

We propose a new generative model of sentences that first samples a prot...
research
06/03/2022

Latent Topology Induction for Understanding Contextualized Representations

In this work, we study the representation space of contextualized embedd...
research
05/07/2017

Generating Memorable Mnemonic Encodings of Numbers

The major system is a mnemonic system that can be used to memorize seque...
research
06/15/2021

Unsupervised Abstractive Opinion Summarization by Generating Sentences with Tree-Structured Topic Guidance

This paper presents a novel unsupervised abstractive summarization metho...
research
01/29/2018

Discrete Autoencoders for Sequence Models

Recurrent models for sequences have been recently successful at many tas...

Please sign up or login with your details

Forgot password? Click here to reset