Leveraging Pre-trained Models for Failure Analysis Triplets Generation

10/31/2022
by   Kenneth Ezukwoke, et al.
0

Pre-trained Language Models recently gained traction in the Natural Language Processing (NLP) domain for text summarization, generation and question-answering tasks. This stems from the innovation introduced in Transformer models and their overwhelming performance compared with Recurrent Neural Network Models (Long Short Term Memory (LSTM)). In this paper, we leverage the attention mechanism of pre-trained causal language models such as Transformer model for the downstream task of generating Failure Analysis Triplets (FATs) - a sequence of steps for analyzing defected components in the semiconductor industry. We compare different transformer models for this generative task and observe that Generative Pre-trained Transformer 2 (GPT2) outperformed other transformer model for the failure analysis triplet generation (FATG) task. In particular, we observe that GPT2 (trained on 1.5B parameters) outperforms pre-trained BERT, BART and GPT3 by a large margin on ROUGE. Furthermore, we introduce Levenshstein Sequential Evaluation metric (LESE) for better evaluation of the structured FAT data and show that it compares exactly with human judgment than existing metrics.

READ FULL TEXT

page 3

page 23

page 24

page 25

page 32

page 33

research
10/07/2021

A Comparative Study of Transformer-Based Language Models on Extractive Question Answering

Question Answering (QA) is a task in natural language processing that ha...
research
03/31/2022

Leveraging pre-trained language models for conversational information seeking from text

Recent advances in Natural Language Processing, and in particular on the...
research
04/27/2023

Neural Keyphrase Generation: Analysis and Evaluation

Keyphrase generation aims at generating topical phrases from a given tex...
research
08/16/2022

BERT(s) to Detect Multiword Expressions

Multiword expressions (MWEs) present groups of words in which the meanin...
research
01/21/2023

REDAffectiveLM: Leveraging Affect Enriched Embedding and Transformer-based Neural Language Model for Readers' Emotion Detection

Technological advancements in web platforms allow people to express and ...
research
04/10/2022

Pushing on Personality Detection from Verbal Behavior: A Transformer Meets Text Contours of Psycholinguistic Features

Research at the intersection of personality psychology, computer science...
research
09/16/2021

MeLT: Message-Level Transformer with Masked Document Representations as Pre-Training for Stance Detection

Much of natural language processing is focused on leveraging large capac...

Please sign up or login with your details

Forgot password? Click here to reset