Foundational large language models (LLMs) can be instruction-tuned to de...
Large language models have shown surprising performances in understandin...
Getting the most out of limited resources allows advances in natural lan...
Driven by the goal of eradicating language barriers on a global scale,
m...
Back translation is one of the most widely used methods for improving th...
We present the first direct simultaneous speech-to-speech translation
(S...
Every day, millions of people sacrifice their privacy and browsing habit...
Is bias amplified when neural machine translation (NMT) models are optim...
We explore two types of monolingual data that can be included in knowled...
Supervised Chinese word segmentation has been widely approached as seque...
We present Sockeye 2, a modernized and streamlined version of the Sockey...
Neural Machine Translation (NMT) is resource intensive. We design a
quan...
Asynchronous stochastic gradient descent (SGD) is attractive from a spee...
We introduce a novel multi-source technique for incorporating source syn...
In order to extract the best possible performance from asynchronous
stoc...
This paper describes the submissions of the "Marian" team to the WNMT 20...
This paper describes the submissions to the efficiency track for GPUs by...
Neural machine translation (NMT) has been accelerated by deep learning n...
Previously, neural methods in grammatical error correction (GEC) did not...
We present Marian, an efficient and self-contained Neural Machine Transl...
This paper describes the University of Edinburgh's submissions to the WM...
We make distributed stochastic gradient descent faster by exchanging spa...