Index

Predicting Hearing Aid Fittings Based on Audiometric and Subject-Related Data: A Machine Learning Approach

Hearing aids (HA) are configured to the wearer’s individual needs, which might vary greatly from user to user. Currently, it is common practice, that the initial HA gain settings are based on generic fitting formulas that link a user’s pure-tone hearing threshold to amplification characteristics. Subsequently, a time-consuming fine-tuning process follows, in which a hearing care professional (HCP) adjusts the HA settings to the user’s individual demands. An advanced, more personalized gain prescription procedure could support HCPs by reducing fine-tuning effort and facilitate over-the-counter HAs. We propose a machine learning based prediction for HA gain to minimize subsequent fine-tuning effort. The data-driven approach takes audiometic and personal variables into account, such as age, gender, and the user’s acoustical environment.

A random forest regression model was trained on real-world HA fittings from the Connexx database (fitting software provided by Sivantos GmbH). Three months of data from Connexx version 9.1.0.364 were used. A data cleaning framework was implemented to extract a representative data set based on a list of machine learning and audiological criteria. These criteria include, for instance, using only ‘informative’ HCPs who perform fine-tuning for at least some patients. Furthermore, ‘informative’ HCPs are those who perform diagnostics beyond air conduction audiograms, use new technologies and special features. The resulting training data comprised 20,000 HA fittings and a 10-fold cross validation was used to train the random forest.

Deep Learning-based Spectral Noise Reduction for Hearing Aids

The great success of deep learning-based noise reduction algorithms makes it desirable to also use them for hearing aid applications. However, neural networks are both computationally intensive and memory intensive, making it challenging to deploy on an embedded system with limited hardware resources. Thus, in this work, we propose an efficient deep learning-based noise reduction method for hearing aid applications. Compared to a previous study, where a fully-connected neural network was used to estimate Wiener filter gains, we focus on using Recurrent Neural Networks (RNNs). Additionally, convolutional layers were integrated. The neural networks were trained to predict real-valued Wiener filter gains to denoise the noisy speech spectrum. Normalizing the input of the neural network is essential. Therefore, various normalization methods were analyzed, allowing low-cost real-time processing. The presented methods were tested and evaluated on unseen noise and speech data. In comparison to the previous study, the computational complexity and the memory requirements of the neural network were reduced by a factor of more than 400, the complexity of the normalization method by a factor of over 200, while even reaching a higher denoising quality.

Multi-Task Learning for Speech Enhancement and Phoneme Recognition

For speech intelligibility, consonants have a fundamental importance. Unfortunately, when reducing noise in speech, consonants are often also degraded while vocals are easier to preserve/enhance. To improve the detection and enhancement of consonants, we want to use multi-task learning to reduce the noise in the signal and furthermore detect phonemes (smallest acoustic unit in speech).

Development of a deep learning-based phoneme recognizer for noisy speech

For speech intelligibility, consonants have a fundamental importance. Thus, the assumption can be made that automatic phoneme and especially consonant recognition correlates well with human speech intelligibility.
In noisy environments however, speech and especially consonants may be degraded.

In this project, we want to study the effect of noise on speech intelligibility. Therefore, we train a neural network to recognize phoneme based on the TIMIT dataset. We will add diffent noise types and noise levels to the speech signal and study the effect on the recognition rate.

This project requires no preliminary knowledge in deep learning, although may be beneficial. Basic signal processing concepts like sampling theorem and FFT should be present.

Development of a pre-processing/simulation Framework for Multi-Channel Audio Signals

The goal of this thesis is to develop a framework that simulates multi-channel audio signals in a 2D/3D environment for hearing aids. For this purpose, existing head related transfer functions (HRTFs) will be used to simulate direction and hearing aid microphone characteristics. Furthermore, source movement as well as microphone movement and rotation will be implemented. The latter is mandatory for hearing aids, since especially head rotation might change the relative direction of the different sources significantly. The framework will be able to simulate multiple speakers as well as multiple noise sources. To calculate a clean speech target, a provided reference beamformer will be used on the target speech only, neglecting noise and non target speakers. Optionally, an opening angle that defines the target directions can be used to extract the clean speech targets. As a second optional aspect, room environment including absorption and reverberation will be simulated. Therefore, a reference implementation can be used.

Weakly-supervised segmentation of defects on solar cells

Using Deep Learning for segmentation of localized defects on SiC wafers

Learning From Weak Label Sources In Industrial Sensor Systems

Diagnosis of Gasoline Direct Injection Systems with Neural Networks

Semantic Segmentation of the Human Eye for Driver Monitoring