Neural-based Noise Filtering from Word Embeddings

10/06/2016
by   Kim Anh Nguyen, et al.
0

Word embeddings have been demonstrated to benefit NLP tasks impressively. Yet, there is room for improvement in the vector representations, because current word embeddings typically contain unnecessary information, i.e., noise. We propose two novel models to improve word embeddings by unsupervised learning, in order to yield word denoising embeddings. The word denoising embeddings are obtained by strengthening salient information and weakening noise in the original word embeddings, based on a deep feed-forward neural network filter. Results from benchmark tasks show that the filtered word denoising embeddings outperform the original word embeddings.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
10/08/2018

Word Embeddings from Large-Scale Greek Web content

Word embeddings are undoubtedly very useful components in many NLP tasks...
research
11/23/2017

SPINE: SParse Interpretable Neural Embeddings

Prediction without justification has limited utility. Much of the succes...
research
04/25/2020

All Word Embeddings from One Embedding

In neural network-based models for natural language processing (NLP), th...
research
05/18/2021

WOVe: Incorporating Word Order in GloVe Word Embeddings

Word vector representations open up new opportunities to extract useful ...
research
04/21/2018

Context-Attentive Embeddings for Improved Sentence Representations

While one of the first steps in many NLP systems is selecting what embed...
research
05/09/2018

Incorporating Subword Information into Matrix Factorization Word Embeddings

The positive effect of adding subword information to word embeddings has...
research
10/09/2015

Controlled Experiments for Word Embeddings

An experimental approach to studying the properties of word embeddings i...

Please sign up or login with your details

Forgot password? Click here to reset