Instruction tuning has significantly advanced large language models (LLM...
This work proposes POMP, a prompt pre-training method for vision-languag...
Contrastive Language-Image Pre-training (CLIP) has demonstrated great
po...
Knowledge distillation (KD) has been proved effective for compressing
la...
Data augmentation aims to enrich training samples for alleviating the
ov...
Despite the achievements of large-scale multimodal pre-training approach...
Dynamic early exiting aims to accelerate pre-trained language models' (P...
We focus on the task of Automatic Live Video Commenting (ALVC), which ai...