Self-attention based transformer models have been dominating many comput...
Large pretrained language models (PLMs) are often domain- or task-adapte...
Vision transformers (ViT) have recently attracted considerable attention...
We study the few-shot learning (FSL) problem, where a model learns to
re...
Network pruning has become the de facto tool to accelerate deep neural
n...
Kernel approximation methods have been popular techniques for scalable k...