Environment Sound Classification Using a Two-Stream CNN Based on Decision-Level Fusion - Université Paris-Est-Créteil-Val-de-Marne Access content directly
Journal Articles Sensors Year : 2019

Environment Sound Classification Using a Two-Stream CNN Based on Decision-Level Fusion

Yu Su
Ke Zhang
  • Function : Author
Jingyu Wang
  • Function : Author

Abstract

With the popularity of using deep learning-based models in various categorization problems and their proven robustness compared to conventional methods, a growing number of researchers have exploited such methods in environment sound classification tasks in recent years. However, the performances of existing models use auditory features like log-mel spectrogram (LM) and mel frequency cepstral coefficient (MFCC), or raw waveform to train deep neural networks for environment sound classification (ESC) are unsatisfactory. In this paper, we first propose two combined features to give a more comprehensive representation of environment sounds Then, a fourfour-layer convolutional neural network (CNN) is presented to improve the performance of ESC with the proposed aggregated features. Finally, the CNN trained with different features are fused using the Dempster–Shafer evidence theory to compose TSCNN-DS model. The experiment results indicate that our combined features with the four-layer CNN are appropriate for environment sound taxonomic problems and dramatically outperform other conventional methods. The proposed TSCNN-DS model achieves a classification accuracy of 97.2%, which is the highest taxonomic accuracy on UrbanSound8K datasets compared to existing models.

Dates and versions

hal-04318247 , version 1 (01-12-2023)

Identifiers

Cite

Yu Su, Ke Zhang, Jingyu Wang, Kurosh Madani. Environment Sound Classification Using a Two-Stream CNN Based on Decision-Level Fusion. Sensors, 2019, 19 (7), pp.1733. ⟨10.3390/s19071733⟩. ⟨hal-04318247⟩

Collections

LISSI UPEC
7 View
0 Download

Altmetric

Share

Gmail Mastodon Facebook X LinkedIn More