Recent studies have discovered that Chain-of-Thought prompting (CoT) can...
Current endpointing (EP) solutions learn in a supervised framework, whic...
Recently, subgraph GNNs have emerged as an important direction for devel...
Structure-based drug design, i.e., finding molecules with high affinitie...
We propose a new class of linear Transformers called
FourierLearner-Tran...
Designing expressive Graph Neural Networks (GNNs) is a central topic in
...
Designing an efficient yet deployment-friendly 3D backbone to handle spa...
Spiking neural networks (SNNs) are promising brain-inspired energy-effic...
Designing neural networks with bounded Lipschitz constant is a promising...
Unlike vision and language data which usually has a unique format, molec...
Adversarial examples, which are usually generated for specific inputs wi...
The Physics-Informed Neural Network (PINN) approach is a new and promisi...
Relative Positional Encoding (RPE), which encodes the relative distance
...
We present an efficient method of pretraining large-scale autoencoding
l...
This technical note describes the recent updates of Graphormer, includin...
This technical note describes the recent updates of Graphormer, includin...
While end-to-end models have shown great success on the Automatic Speech...
The effectiveness of knowledge graph embedding (KGE) largely depends on ...
Several recent studies have demonstrated that attention-based networks, ...
Recently, Zhang et al. (2021) developed a new neural network architectur...
The attention module, which is a crucial component in Transformer, canno...
In this technical report, we present our solution of KDD Cup 2021 OGB
La...
The Transformer architecture has become a dominant choice in many domain...
Adversarial training (AT) is one of the most effective strategies for
pr...
Semantic understanding of programs is a fundamental problem for programm...
Wav2vec-C introduces a novel representation learning technique combining...
An important development in deep learning from the earliest MLPs has bee...
Improving the efficiency of Transformer-based language pre-training is a...
Many real-world applications use Siamese networks to efficiently match t...
Transformer has demonstrated its great power to learn contextual word
re...
It is well-known that standard neural networks, even with a high
classif...
Accurate and robust prediction of patient's response to drug treatments ...
An unsolved fundamental problem in biology and ecology is to predict
obs...
Normalization plays an important role in the optimization of deep neural...
How to make unsupervised language pre-training more efficient and less
r...
Understanding what information neural networks capture is an essential
p...
How to explicitly encode positional information into neural networks is ...
How to explicitly encode positional information into neural networks is ...
Pre-trained contextual representations (e.g., BERT) have become the
foun...
High-resolution digital images are usually downscaled to fit various dis...
The recently proposed BERT has shown great power on a variety of natural...
The Transformer is widely used in natural language processing tasks. To ...
Adversarial training is one of the most popular ways to learn robust mod...
Robustness of convolutional neural networks has recently been highlighte...
Autoregressive sequence models achieve state-of-the-art performance in
d...
Generative models, especially Generative Adversarial Networks (GANs), ha...
Due to the unparallelizable nature of the autoregressive factorization,
...
Multilingual neural machine translation (NMT), which translates multiple...
We study an interesting problem in training neural network-based models ...
The Transformer architecture is widely used in natural language processi...