The large-scale visual pretraining has significantly improve the perform...
The tremendous success of large models trained on extensive datasets
dem...
At the heart of foundation models is the philosophy of "more is differen...
Masked Image Modeling (MIM) achieves outstanding success in self-supervi...
The combination of transformers and masked image modeling (MIM) pre-trai...
Light-weight convolutional neural networks (CNNs) are specially designed...
Knowledge graphs (KGs) are known for their large scale and knowledge
inf...
Network architecture plays a key role in the deep learning-based compute...
Recently, Multilayer Perceptron (MLP) becomes the hotspot in the field o...
Deploying convolutional neural networks (CNNs) on mobile devices is diff...
Transformer networks have achieved great progress for computer vision ta...
Different from traditional convolutional neural network (CNN) and vision...
This paper presents Hire-MLP, a simple yet competitive vision MLP
archit...
Vision transformers have been successfully applied to image recognition ...
This paper studies the model compression problem of vision transformers....
This paper studies the efficiency problem for visual transformers by
exc...
Knowledge distillation is a widely used paradigm for inheriting informat...
Transformer is a type of self-attention-based neural networks originally...
Transformer is a type of deep neural network mainly based on self-attent...
Neural Architecture Search (NAS) refers to automatically design the
arch...
Neural Architecture Search (NAS) has achieved great success in image
cla...
Deploying convolutional neural networks (CNNs) on embedded devices is
di...
Person re-identification is a challenging task due to various complex
fa...
In this paper, we present a so-called interlaced sparse self-attention
a...
How to learn a discriminative fine-grained representation is a key point...