Rising computational demands of modern natural language processing (NLP)...
The integration of multi-document pre-training objectives into language
...
Diffusion models have emerged as a powerful paradigm for generation,
obt...
Pretrained language models (PLMs) are trained on massive corpora, but of...
Rationalization is fundamental to human reasoning and learning. NLP mode...
This work introduces ATTEMPT (Attentional Mixture of Prompt Tuning), a n...
Prior work on controllable text generation has focused on learning how t...
We investigate input-conditioned hypernetworks for multi-tasking in NLP,...
Generative language models are trained on diverse, general domain corpor...
Self-rationalization models that predict task labels and generate free-t...
Making controlled perturbations is essential for various tasks (e.g., da...
Much recent work in NLP has documented dataset artifacts, bias, and spur...
We introduce a new pretraining approach for language models that are gea...
Humans give contrastive explanations that explain why an observed event
...
Typically, machine learning systems solve new tasks by training on thous...
Transformer-based models are unable to process long sequences due to the...
Large neural models have demonstrated human-level performance on languag...
Contextual word representations, typically trained on unstructured, unla...
Contextual word representations derived from pre-trained bidirectional
l...
We introduce a new type of deep contextualized word representation that
...
Pre-trained word embeddings learned from unlabeled text have become a
st...