Towards a Neural-Based Single Channel Speech Enhancement Model for Hearing-Aids

Main Authors: Padinjaru Veettil, Muhammed Shifas, Santelli, Claudio, Stylianou, Yannis
Format: Proceeding
Bahasa: eng
Terbitan: Deutsche Gesellschaft für Akustik (DEGA e. V.) & RWTH Publications , 2019
Subjects:
Online Access: https://zenodo.org/record/3529378
Daftar Isi:
  • Advancements in machine learning techniques have promoted the use of deep neural networks (DNNs) forsupervised speech enhancement. However, the DNN’s benefits of non-explicit noise statistics and nonlinearmodeling capacity come at the expense of increased computational complexity for training and inference whichis an issue for real-time restricted applications, like hearing aids. Contrary to the conventional approach whichseparately models the feature extraction and temporal dependency through a sequence of convolutional layersfollowed by a fully-connected recurrent layer, this work promotes the use of convolutional recurrent network lay-ers for single-channel speech enhancement. Thereby, temporal correlations among inherently extracted spectralfeature vectors are exploited, while further reducing the parameter set to be estimated relative to the conven-tional method. The proposed method is compared to a recent low algorithmic delay architecture. The modelswere trained in a speaker independent fashion on the NSDTSEA data set composed of different environmen-tal noises. While objective speech quality and intelligibility measures of the two architectures are similar, thenumber of network parameters in the suggested enhancement method being reduced by 66%. This reduction ishighly beneficial for storage and computation constraint applications.