Despite their impressive capabilities, large language models (LLMs) are ...
Large language models (LLMs) can learn to perform a wide range of natura...
Entailment has been recognized as an important metric for evaluating nat...
Masked language models (MLM) do not explicitly define a distribution ove...
Prompt tuning, in which a base pretrained model is adapted to each task ...
Scaling transformers has led to significant breakthroughs in many domain...
The canonical formulation of federated learning treats it as a distribut...
We study grammar induction with mildly context-sensitive grammars for
un...
Large pre-trained models decay over long-term deployment as input
distri...
Next-word predictions from autoregressive neural language models show
re...
We describe a neural transducer that maintains the flexibility of standa...
Designing better machine translation systems by considering auxiliary in...
We show that large language models, such as GPT-3, perform well at zero-...
Transition-based parsers for Abstract Meaning Representation (AMR) rely ...
The finetuning of pretrained transformer-based language generation model...
We demonstrate that co-training (Blum Mitchell, 1998) can improve th...
Sequence-to-sequence learning with neural networks has become the de fac...
The developmental process of embryos follows a monotonic order. An embry...
While vector-based language representations from pretrained language mod...
While task-specific finetuning of pretrained networks has led to signifi...
Despite their empirical success, neural networks still have difficulty
c...
Deep neural networks (DNNs) have shown much empirical success in solving...
We study a formalization of the grammar induction problem that models
se...
We propose to learn deep undirected graphical models (i.e., MRFs), with ...
Recurrent neural network grammars (RNNG) are generative models of langua...
There has been much recent, exciting work on combining the complementary...
Variational autoencoders (VAEs) learn distributions of high-dimensional ...
Neural attention has become central to many state-of-the-art models in
n...
OpenNMT is an open-source toolkit for neural machine translation (NMT). ...
Amortized variational inference (AVI) replaces instance-specific local
i...
In a controlled experiment of sequence-to-sequence approaches for the ta...
While autoencoders are a key technique in representation learning for
co...
Attention networks have proven to be an effective approach for embedding...
We describe an open-source toolkit for neural machine translation (NMT)....
Neural machine translation (NMT) offers a novel alternative formulation ...
We demonstrate that an attention-based encoder-decoder model can be used...
We describe a simple neural language model that relies only on
character...
We report on a series of experiments with convolutional neural networks ...