Large language models (LLMs) have made significant strides in various ta...
As the capabilities of large language models (LLMs) continue to advance,...
Large language models are powerful text processors and reasoners, but ar...
Open-domain question answering is a crucial task that often requires
acc...
Large Language Models (LLMs) play a powerful Reader of the
Retrieve-then...
Diffusion models have gained significant attention in the realm of image...
Based on the remarkable achievements of pre-trained language models in
a...
Large language models (LLMs) can achieve highly effective performance on...
Many natural language processing (NLP) tasks rely on labeled data to tra...
Large language models can perform various reasoning tasks by using
chain...
In this paper, we propose a large-scale language pre-training for text
G...
The dual-encoder has become the de facto architecture for dense retrieva...
Dense retrieval aims to map queries and passages into low-dimensional ve...
Long-form numerical reasoning in financial analysis aims to generate a
r...
Knowledge distillation is often used to transfer knowledge from a strong...
We introduce GENIUS: a conditional text generation model using sketches ...
Sampling proper negatives from a large document pool is vital to effecti...
Commonsense generation aims to generate a realistic sentence describing ...
Most existing pre-trained language representation models (PLMs) are
sub-...
Code contrastive pre-training has recently achieved significant progress...
Knowledge distillation is an effective way to transfer knowledge from a
...
Due to exposure bias, most existing natural language generation (NLG) mo...
Non-Autoregressive generation is a sequence generation paradigm, which
r...
Dialog response generation in open domain is an important research topic...
Vector quantization (VQ) based ANN indexes, such as Inverted File System...
In this paper, we propose the CodeRetriever model, which combines the
un...
Current dense text retrieval models face two typical challenges. First, ...
We study the problem of coarse-grained response selection in retrieval-b...
Pre-trained language models have led to substantial gains over a broad r...
Transformer-based models have made tremendous impacts in natural languag...
Transformer model with multi-head attention requires caching intermediat...
In this paper, we introduce a two-level attention schema, Poolingformer,...
Now, the pre-training technique is ubiquitous in natural language proces...
Transformer is an attention-based neural network, which consists of two
...
Conditional random fields (CRF) for label decoding has become ubiquitous...
Commonsense generation aims at generating plausible everyday scenario
de...
In a sponsored search engine, generative retrieval models are recently
p...
In this paper, we propose a novel data augmentation method, referred to ...
Reading long documents to answer open-domain questions remains challengi...
News headline generation aims to produce a short sentence to attract rea...
In this paper, we introduce XGLUE, a new benchmark dataset to train
larg...
In this paper, we present a new sequence-to-sequence pre-training model
...
Neural semantic parsing has achieved impressive results in recent years,...
In this paper, we propose a novel pretraining-based encoder-decoder
fram...