Less is More: Simplifying Feature Extractors Prevents Overfitting for Neural Discourse Parsing Models

10/18/2022
by   Ming Li, et al.
0

Complex feature extractors are widely employed for text representation building. However, these complex feature extractors can lead to severe overfitting problems especially when the training datasets are small, which is especially the case for several discourse parsing tasks. Thus, we propose to remove additional feature extractors and only utilize self-attention mechanism to exploit pretrained neural language models in order to mitigate the overfitting problem. Experiments on three common discourse parsing tasks (News Discourse Profiling, Rhetorical Structure Theory based Discourse Parsing and Penn Discourse Treebank based Discourse Parsing) show that powered by recent pretrained language models, our simplied feature extractors obtain better generalizabilities and meanwhile achieve comparable or even better system performance. The simplified feature extractors have fewer learnable parameters and less processing time. Codes will be released and this simple yet effective model can serve as a better baseline for future research.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
04/13/2021

Discourse Probing of Pretrained Language Models

Existing work on probing of pretrained language models (LMs) has predomi...
research
10/15/2022

A Simple and Strong Baseline for End-to-End Neural RST-style Discourse Parsing

To promote and further develop RST-style discourse parsing models, we ne...
research
05/10/2015

Fast Rhetorical Structure Theory Discourse Parsing

In recent years, There has been a variety of research on discourse parsi...
research
05/15/2023

Uncovering the Potential of ChatGPT for Discourse Analysis in Dialogue: An Empirical Study

Large Language Models (LLMs) like ChatGPT have proven a great shallow un...
research
11/25/2017

Complex Structure Leads to Overfitting: A Structure Regularization Decoding Method for Natural Language Processing

Recent systems on structured prediction focus on increasing the level of...
research
06/07/2023

Cross-Genre Argument Mining: Can Language Models Automatically Fill in Missing Discourse Markers?

Available corpora for Argument Mining differ along several axes, and one...
research
06/30/2021

A Conditional Splitting Framework for Efficient Constituency Parsing

We introduce a generic seq2seq parsing framework that casts constituency...

Please sign up or login with your details

Forgot password? Click here to reset