Prompt-tuning has become an increasingly popular parameter-efficient met...
Regularization techniques are crucial to improving the generalization
pe...
Deep models are dominating the artificial intelligence (AI) industry sin...
Fine-tuning a Pre-trained Language Model (PLM) on a specific downstream ...
Knowledge Distillation (KD) has been extensively used for natural langua...
Knowledge Distillation (KD) is a commonly used technique for improving t...
With the ever-growing size of pre-trained models (PMs), fine-tuning them...
We propose a general deep architecture for learning functions on multipl...
Knowledge Distillation (KD) is a prominent neural model compression tech...
At its core, generative modeling seeks to uncover the underlying factors...
Learning disentangled representations of real world data is a challengin...
State-of-the-art neural dialogue systems excel at syntactic and semantic...
Normalizing Flows are generative models which produce tractable distribu...
Triangular maps are a construct in probability theory that allows the
tr...
Graphs arise naturally in many real-world applications including social
...