Subword tokenization is a key part of many NLP pipelines. However, littl...
Byte-Pair Encoding (BPE) is a popular algorithm used for tokenizing data...
Causal inference is one of the hallmarks of human intelligence. While th...
Solving math story problems is a complex task for students and NLP model...
The Abstraction and Reasoning Corpus (ARC) (Chollet, 2019) and its most
...
Adaptive learning aims to provide customized educational activities (e.g...
Membership Inference attacks (MIAs) aim to predict whether a data sample...
We show that most structured prediction problems can be solved in linear...
Mathematical reasoning in large language models (LLMs) has garnered atte...
Transformer models bring propelling advances in various NLP tasks, thus
...
Although automatic dialogue tutors hold great potential in making educat...
Multi-task learning (MTL) aims at achieving a better model by leveraging...
The fixed-size context of Transformer makes GPT models incapable of
gene...
Topic models are used to make sense of large text collections. However,
...
The primary way of building AI applications is shifting from training
sp...
Several recent papers claim human parity at sentence-level Machine
Trans...
We present a novel extension of the traditional neural network approach ...
With the recent advances in natural language processing (NLP), a vast nu...
NLP datasets are richer than just input-output pairs; rather, they carry...
Large language models generate fluent texts and can follow natural langu...
Textbooks are the primary vehicle for delivering quality education to
st...
Word embeddings that map words into a fixed-dimensional vector space are...
Ideally, dialogue systems should generate responses that are faithful to...
Conversational tutoring systems (CTSs) aim to help students master
educa...
Designing dialog tutors has been challenging as it involves modeling the...
Machine translation quality estimation (QE) predicts human judgements of...
Generated texts from large pretrained language models have been shown to...
Step-by-step reasoning approaches like chain-of-thought (CoT) have prove...
Socratic questioning is an educational method that allows students to
di...
Recent work has demonstrated that pre-trained language models (PLMs) are...
Recent years have seen a paradigm shift in NLP towards using pretrained
...
Centering theory (CT; Grosz et al., 1995) provides a linguistic analysis...
Machine translation (MT) has almost achieved human parity at sentence-le...
To protect the privacy of individuals whose data is being shared, it is ...
Large language models appear to learn facts from the large text corpora ...
We have recently witnessed a number of impressive results on hard
mathem...
Ontonotes has served as the most important benchmark for coreference
res...
AI systems are becoming increasingly intertwined with human life. In ord...
Probing is a popular method to discern what linguistic information is
co...
Many natural language processing tasks, e.g., coreference resolution and...
Human-translated text displays distinct features from naturally written ...
In typical machine learning systems, an estimate of the probability of t...
Languages are continuously undergoing changes, and the mechanisms that
u...
Reasoning is central to human intelligence. However, fallacious argument...
Pretrained language models (LMs) do not capture factual knowledge very w...
Text-based games (TBG) have emerged as promising environments for drivin...
In this work we introduce KERNELIZED TRANSFORMER, a generic, scalable, d...
The principle of independent causal mechanisms (ICM) states that generat...
When reading a literary piece, readers often make inferences about vario...
Multi-head attention, a collection of several attention mechanisms that
...