Murdoch University Research Repository

Welcome to the Murdoch University Research Repository

The Murdoch University Research Repository is an open access digital collection of research
created by Murdoch University staff, researchers and postgraduate students.

Learn more

Multi-Task learning for acoustic event detection using event and frame position information

Xia, X., Togneri, R., Sohel, F., Zhao, Y. and Huang, D. (2020) Multi-Task learning for acoustic event detection using event and frame position information. IEEE Transactions on Multimedia, 22 (3). pp. 569-578.

Link to Published Version: https://doi.org/10.1109/TMM.2019.2933330
*Subscription may be required

Abstract

Acoustic event detection deals with the acoustic signals to determine the sound type and to estimate the audio event boundaries. Multi-label classification based approaches are commonly used to detect the frame wise event types with a median filter applied to determine the happening acoustic events. However, the multi-label classifiers are trained only on the acoustic event types ignoring the frame position within the audio events. To deal with this, this paper proposes to construct a joint learning based multi-task system. The first task performs the acoustic event type detection and the second task is to predict the frame position information. By sharing representations between the two tasks, we can enable the acoustic models to generalize better than the original classifier by averaging respective noise patterns to be implicitly regularized. Experimental results on the monophonic UPC-TALP and the polyphonic TUT Sound Event datasets demonstrate the superior performance of the joint learning method by achieving lower error rate and higher F-score compared to the baseline AED system.

Item Type: Journal Article
Murdoch Affiliation: College of Science, Health, Engineering and Education
Publisher: IEEE
Copyright: © 2020 IEEE.
URI: http://researchrepository.murdoch.edu.au/id/eprint/55058
Item Control Page Item Control Page