Representation based meta-learning for few-shot spoken intent recognition

06/29/2021
by   Ashish Mittal, et al.
4

Spoken intent detection has become a popular approach to interface with various smart devices with ease. However, such systems are limited to the preset list of intents-terms or commands, which restricts the quick customization of personal devices to new intents. This paper presents a few-shot spoken intent classification approach with task-agnostic representations via meta-learning paradigm. Specifically, we leverage the popular representation-based meta-learning learning to build a task-agnostic representation of utterances, that then use a linear classifier for prediction. We evaluate three such approaches on our novel experimental protocol developed on two popular spoken intent classification datasets: Google Commands and the Fluent Speech Commands dataset. For a 5-shot (1-shot) classification of novel classes, the proposed framework provides an average classification accuracy of 88.6 Speech Commands dataset. The performance is comparable to traditionally supervised classification models with abundant training samples.

READ FULL TEXT

Please sign up or login with your details

Forgot password? Click here to reset