Weakly Labeled Sound Event Detection using Tri-training and Adversarial Learning
|Citation:||H. Park, S. Yun, J. Eum, J. Cho & K. Hwang, "Weakly Labeled Sound Event Detection using Tri-training and Adversarial Learning", Proceedings of the Detection and Classification of Acoustic Scenes and Events 2019 Workshop (DCASE2019), pages 184–188, New York University, NY, USA, Oct. 2019|
|Abstract:||This paper considers a semi-supervised learning framework for weakly labeled polyphonic sound event detection problems for the DCASE 2019 challenge's task4 by combining both the tri-training and adversarial learning. The goal of the task4 is to detect onsets and offsets of multiple sound events in a single audio clip. The entire dataset consists of the synthetic data with a strong label (sound event labels with boundaries) and real data with weakly labeled (sound event labels) and unlabeled dataset. Given this dataset, we apply the tri-training where two different classifiers are used to obtain pseudo labels on the weakly labeled and unlabeled dataset, and the final classifier is trained using the strongly labeled dataset and weakly/unlabeled dataset with pseudo labels. Also, we apply the adversarial learning to reduce the domain gap between the real and synthetic dataset. We evaluated our learning framework using the validation set of the task4 dataset, and in the experiments, our learning framework shows a considerable performance improvement over the baseline model.|
|Appears in Collections:||Proceedings of the Detection and Classification of Acoustic Scenes and Events 2019 Workshop (DCASE2019)|
Files in This Item:
|DCASE2019Workshop_Park_80.pdf||772.87 kB||Adobe PDF||View/Open|
Items in FDA are protected by copyright, with all rights reserved, unless otherwise indicated.