计算机科学
语音识别
分类器(UML)
事件(粒子物理)
帧(网络)
人工智能
多任务学习
任务(项目管理)
模式识别(心理学)
量子力学
电信
物理
经济
管理
作者
Xianjun Xia,Roberto Togneri,Ferdous Sohel,Yuanjun Zhao,Defeng Huang
标识
DOI:10.1109/tmm.2019.2933330
摘要
Acoustic event detection deals with the acoustic signals to determine the sound type and to estimate the audio event boundaries. Multi-label classification based approaches are commonly used to detect the frame wise event types with a median filter applied to determine the happening acoustic events. However, the multi-label classifiers are trained only on the acoustic event types ignoring the frame position within the audio events. To deal with this, this paper proposes to construct a joint learning based multi-task system. The first task performs the acoustic event type detection and the second task is to predict the frame position information. By sharing representations between the two tasks, we can enable the acoustic models to generalize better than the original classifier by averaging respective noise patterns to be implicitly regularized. Experimental results on the monophonic UPC-TALP and the polyphonic TUT Sound Event datasets demonstrate the superior performance of the joint learning method by achieving lower error rate and higher F-score compared to the baseline AED system.
科研通智能强力驱动
Strongly Powered by AbleSci AI