Multi-Task Structural Learning using Local Task Similarity induced Neuron Creation and Removal

04/30/2023
by   Naresh Kumar Gurulingan, et al.
8

Multi-task learning has the potential to improve generalization by maximizing positive transfer between tasks while reducing task interference. Fully achieving this potential is hindered by manually designed architectures that remain static throughout training. On the contrary, learning in the brain occurs through structural changes that are in tandem with changes in synaptic strength. Thus, we propose Multi-Task Structural Learning (MTSL) that simultaneously learns the multi-task architecture and its parameters. MTSL begins with an identical single-task network for each task and alternates between a task-learning phase and a structural-learning phase. In the task learning phase, each network specializes in the corresponding task. In each of the structural learning phases, starting from the earliest layer, locally similar task layers first transfer their knowledge to a newly created group layer before being removed. MTSL then uses the group layer in place of the corresponding removed task layers and moves on to the next layers. Our empirical results show that MTSL achieves competitive generalization with various baselines and improves robustness to out-of-distribution data.

READ FULL TEXT

page 3

page 13

research
05/02/2020

Understanding and Improving Information Transfer in Multi-Task Learning

We investigate multi-task learning approaches that use a shared feature ...
research
05/15/2018

Cross-connected Networks for Multi-task Learning of Detection and Segmentation

Multi-task learning improves generalization performance by sharing knowl...
research
09/11/2020

Towards Interpretable Multi-Task Learning Using Bilevel Programming

Interpretable Multi-Task Learning can be expressed as learning a sparse ...
research
06/08/2021

One Semantic Parser to Parse Them All: Sequence to Sequence Multi-Task Learning on Semantic Parsing Datasets

Semantic parsers map natural language utterances to meaning representati...
research
02/18/2023

MaxGNR: A Dynamic Weight Strategy via Maximizing Gradient-to-Noise Ratio for Multi-Task Learning

When modeling related tasks in computer vision, Multi-Task Learning (MTL...
research
08/19/2022

Curbing Task Interference using Representation Similarity-Guided Multi-Task Feature Sharing

Multi-task learning of dense prediction tasks, by sharing both the encod...
research
03/08/2023

Better Together: Using Multi-task Learning to Improve Feature Selection within Structural Datasets

There have been recent efforts to move to population-based structural he...

Please sign up or login with your details

Forgot password? Click here to reset