Publications

Peft-ser: On the use of parameter efficient transfer learning approaches for speech emotion recognition using pre-trained speech models

Abstract

Many recent studies have focused on fine-tuning pretrained models for speech emotion recognition (SER), resulting in promising performance compared to traditional methods that rely largely on low-level, knowledge-inspired acoustic features. These pre-trained speech models learn general-purpose speech representations using self-supervised or weakly-supervised learning objectives from large-scale datasets. Despite the significant advances made in SER through the use of pre-trained architecture, fine-tuning these large pre-trained models for different datasets requires saving copies of entire weight parameters, rendering them impractical to deploy in real-world settings. As an alternative, this work explores parameter-efficient fine-tuning (PEFT) approaches for adapting pre-trained speech models for emotion recognition. Specifically, we evaluate the efficacy of adapter tuning, embedding prompt tuning, and …

Date
September 10, 2023
Authors
Tiantian Feng, Shrikanth Narayanan
Conference
2023 11th International Conference on Affective Computing and Intelligent Interaction (ACII)
Pages
1-8
Publisher
IEEE