Most interpretability research in NLP focuses on understanding the behav...
The impact of randomness on model training is poorly understood. How do
...
Most works on transformers trained with the Masked Language Modeling (ML...
At NeurIPS, American and Chinese institutions cite papers from each othe...
Recombining known primitive concepts into larger novel combinations is a...
It is widely accepted in the mode connectivity literature that when two
...
Experiments with pretrained models such as BERT are often based on a sin...
Probes are models devised to investigate the encoding of knowledge – e.g...
Recent work in NLP shows that LSTM language models capture hierarchical
...
The question of how to probe contextual word representations in a way th...
Recent work in NLP shows that LSTM language models capture compositional...
Diverse word representations have surged in most state-of-the-art natura...
Concerns about interpretability, computational resources, and principled...
Recent work has demonstrated that neural language models encode linguist...
We describe DyNet, a toolkit for implementing neural network models base...
Existing corpora for intrinsic evaluation are not targeted towards tasks...