Training a T5 Using Lab-sized Resources

08/25/2022
by   Manuel R. Ciosici, et al.
0

Training large neural language models on large datasets is resource- and time-intensive. These requirements create a barrier to entry, where those with fewer resources cannot build competitive models. This paper presents various techniques for making it possible to (a) train a large language model using resources that a modest research lab might have, and (b) train it in a reasonable amount of time. We provide concrete recommendations for practitioners, which we illustrate with a case study: a T5 model for Danish, the first for this language.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
07/11/2023

SuryaKiran at MEDIQA-Sum 2023: Leveraging LoRA for Clinical Dialogue Summarization

Finetuning Large Language Models helps improve the results for domain-sp...
research
04/25/2023

Unstructured and structured data: Can we have the best of both worlds with large language models?

This paper presents an opinion on the potential of using large language ...
research
08/05/2020

Efficient MDI Adaptation for n-gram Language Models

This paper presents an efficient algorithm for n-gram language model ada...
research
12/01/2021

DPRK-BERT: The Supreme Language Model

Deep language models have achieved remarkable success in the NLP domain....
research
01/11/2020

A Continuous Space Neural Language Model for Bengali Language

Language models are generally employed to estimate the probability distr...
research
10/22/2020

UniCase – Rethinking Casing in Language Models

In this paper, we introduce a new approach to dealing with the problem o...
research
07/08/2022

XR Hackathon Going Online: Lessons Learned from a Case Study with Goethe-Institut

In this article we report a case study of a Language and Culture-oriente...

Please sign up or login with your details

Forgot password? Click here to reset