Distilling BERT for low complexity network training

05/13/2021
by   Bansidhar Mangalwedhekar, et al.
0

This paper studies the efficiency of transferring BERT learnings to low complexity models like BiLSTM, BiLSTM with attention and shallow CNNs using sentiment analysis on SST-2 dataset. It also compares the complexity of inference of the BERT model with these lower complexity models and underlines the importance of these techniques in enabling high performance NLP models on edge devices like mobiles, tablets and MCU development boards like Raspberry Pi etc. and enabling exciting new applications.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
01/10/2022

BERT for Sentiment Analysis: Pre-trained and Fine-Tuned Alternatives

BERT has revolutionized the NLP field by enabling transfer learning with...
research
11/20/2020

Fine-Tuning BERT for Sentiment Analysis of Vietnamese Reviews

Sentiment analysis is an important task in the field ofNature Language P...
research
09/06/2019

2-Local Hamiltonian with Low Complexity is QCMA

We prove that 2-Local Hamiltonian (2-LH) with Low Complexity problem is ...
research
03/05/2021

Fine-tuning Pretrained Multilingual BERT Model for Indonesian Aspect-based Sentiment Analysis

Although previous research on Aspect-based Sentiment Analysis (ABSA) for...
research
02/19/2023

Can ChatGPT Understand Too? A Comparative Study on ChatGPT and Fine-tuned BERT

Recently, ChatGPT has attracted great attention, as it can generate flue...
research
07/19/2018

Improving Simple Models with Confidence Profiles

In this paper, we propose a new method called ProfWeight for transferrin...
research
07/06/2022

Low complexity, low probability patterns and consequences for algorithmic probability applications

Developing new ways to estimate probabilities can be valuable for scienc...

Please sign up or login with your details

Forgot password? Click here to reset