Conformer, a convolution-augmented Transformer variant, has become the d...
Self-supervised speech representation learning (SSL) has shown to be
eff...
Self-supervised pre-trained transformers have improved the state of the ...
Conformer, combining convolution and self-attention sequentially to capt...
We introduce Wav2Seq, the first self-supervised approach to pre-train bo...
The Transformer architecture has been well adopted as a dominant archite...
This paper is a study of performance-efficiency trade-offs in pre-traine...
Automatic speech recognition (ASR) models make fewer errors when more
su...
Capsule networks (CapsNets) have recently gotten attention as alternativ...
In this paper, we present a Small Energy Masking (SEM) algorithm, which ...
In this paper, we present a new on-device automatic speech recognition (...
In this paper, we propose a refined multi-stage multi-task training stra...
In this paper, we describe the Maximum Uniformity of Distribution (MUD)
...
In this paper, we present an end-to-end training framework for building
...