Despite the superior performance, Large Language Models (LLMs) require
s...
In this paper, we propose a highly parameter-efficient approach to scali...
The state-of-the-art Mixture-of-Experts (short as MoE) architecture has
...
This paper presents a novel pre-trained language models (PLM) compressio...
Long Short-Term Memory (LSTM) models are the building blocks of many
sta...
The deep neural network (DNN) based speech enhancement approaches have
a...
A deep neural network is a parameterization of a multi-layer mapping of
...