Knowledge distillation (KD) has been extensively studied in single-label...
Vision-Language Pre-training (VLP) with large-scale image-text pairs has...
Multi-label learning in the presence of missing labels (MLML) is a
chall...
This paper investigates the feasibility of learning good representation ...
It is a consensus that small models perform quite poorly under the parad...
Knowledge distillation(KD) aims to improve the performance of a student
...