A Novel Discriminative Feature Extraction for Acoustic Scene Classification Using RNN Based Source Separation

Seongkyu MUN  Suwon SHON  Wooil KIM  David K. HAN  Hanseok KO  

IEICE TRANSACTIONS on Information and Systems   Vol.E100-D   No.12   pp.3041-3044
Publication Date: 2017/12/01
Online ISSN: 1745-1361
DOI: 10.1587/transinf.2017EDL8132
Type of Manuscript: LETTER
Category: Artificial Intelligence, Data Mining
acoustic scene classification,  transfer learning,  recurrent neural network,  bottleneck feature,  

Full Text: PDF>>
Buy this Article

Various types of classifiers and feature extraction methods for acoustic scene classification have been recently proposed in the IEEE Detection and Classification of Acoustic Scenes and Events (DCASE) 2016 Challenge Task 1. The results of the final evaluation, however, have shown that even top 10 ranked teams, showed extremely low accuracy performance in particular class pairs with similar sounds. Due to such sound classes being difficult to distinguish even by human ears, the conventional deep learning based feature extraction methods, as used by most DCASE participating teams, are considered facing performance limitations. To address the low performance problem in similar class pair cases, this letter proposes to employ a recurrent neural network (RNN) based source separation for each class prior to the classification step. Based on the fact that the system can effectively extract trained sound components using the RNN structure, the mid-layer of the RNN can be considered to capture discriminative information of the trained class. Therefore, this letter proposes to use this mid-layer information as novel discriminative features. The proposed feature shows an average classification rate improvement of 2.3% compared to the conventional method, which uses additional classifiers for the similar class pair issue.