For Full-Text PDF, please login, if you are a member of IEICE,|
or go to Pay Per View on menu list, if you are a nonmember of IEICE.
A Novel Discriminative Feature Extraction for Acoustic Scene Classification Using RNN Based Source Separation
Seongkyu MUN Suwon SHON Wooil KIM David K. HAN Hanseok KO
IEICE TRANSACTIONS on Information and Systems
Publication Date: 2017/12/01
Online ISSN: 1745-1361
Type of Manuscript: LETTER
Category: Artificial Intelligence, Data Mining
acoustic scene classification, transfer learning, recurrent neural network, bottleneck feature,
Full Text: PDF>>
Various types of classifiers and feature extraction methods for acoustic scene classification have been recently proposed in the IEEE Detection and Classification of Acoustic Scenes and Events (DCASE) 2016 Challenge Task 1. The results of the final evaluation, however, have shown that even top 10 ranked teams, showed extremely low accuracy performance in particular class pairs with similar sounds. Due to such sound classes being difficult to distinguish even by human ears, the conventional deep learning based feature extraction methods, as used by most DCASE participating teams, are considered facing performance limitations. To address the low performance problem in similar class pair cases, this letter proposes to employ a recurrent neural network (RNN) based source separation for each class prior to the classification step. Based on the fact that the system can effectively extract trained sound components using the RNN structure, the mid-layer of the RNN can be considered to capture discriminative information of the trained class. Therefore, this letter proposes to use this mid-layer information as novel discriminative features. The proposed feature shows an average classification rate improvement of 2.3% compared to the conventional method, which uses additional classifiers for the similar class pair issue.