Knowledge Distillation (KD) is a promising technique for reducing the hi...
In-context learning, where pre-trained language models learn to perform ...
Large language models have exhibited intriguing in-context learning
capa...
Training language models to learn from human instructions for zero-shot
...
In this work, we formulate Text Classification as a
Matching problem bet...
Large-scale pre-training has shown remarkable performance in building
op...
Prompts for pre-trained language models (PLMs) have shown remarkable
per...
Although pre-trained language models have remarkably enhanced the genera...
Pre-trained Language Models (PLMs) have proven to be beneficial for vari...
Recently, pre-trained language models mostly follow the
pre-training-the...
Multi-hop knowledge graph (KG) reasoning is an effective and explainable...