Fine-grained Analysis of Sentence Embeddings Using Auxiliary Prediction Tasks

08/15/2016
by   Yossi Adi, et al.
0

There is a lot of research interest in encoding variable length sentences into fixed length vectors, in a way that preserves the sentence meanings. Two common methods include representations based on averaging word vectors, and representations based on the hidden states of recurrent neural networks such as LSTMs. The sentence vectors are used as features for subsequent machine learning tasks or for pre-training in the context of deep learning. However, not much is known about the properties that are encoded in these sentence representations and about the language information they capture. We propose a framework that facilitates better understanding of the encoded representations. We define prediction tasks around isolated aspects of sentence structure (namely sentence length, word content, and word order), and score representations by the ability to train a classifier to solve each prediction task when using the representation as input. We demonstrate the potential contribution of the approach by analyzing different sentence representation mechanisms. The analysis sheds light on the relative strengths of different sentence embedding methods with respect to these low level prediction tasks, and on the effect of the encoded vector's dimensionality on the resulting representations.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
10/25/2019

Evaluation of Sentence Representations in Polish

Methods for learning sentence representations have been actively develop...
research
06/13/2019

A Comparison of Word-based and Context-based Representations for Classification Problems in Health Informatics

Distributed representations of text can be used as features when trainin...
research
01/23/2017

dna2vec: Consistent vector representations of variable-length k-mers

One of the ubiquitous representation of long DNA sequence is dividing it...
research
09/18/2018

Analysis of Bag-of-n-grams Representation's Properties Based on Textual Reconstruction

Despite its simplicity, bag-of-n-grams sen- tence representation has bee...
research
11/05/2019

Language coverage and generalization in RNN-based continuous sentence embeddings for interacting agents

Continuous sentence embeddings using recurrent neural networks (RNNs), w...
research
02/20/2020

Contextual Lensing of Universal Sentence Representations

What makes a universal sentence encoder universal? The notion of a gener...
research
04/04/2017

Interpretation of Semantic Tweet Representations

Research in analysis of microblogging platforms is experiencing a renewe...

Please sign up or login with your details

Forgot password? Click here to reset