This project prototypes a machine learning model to detect audio artifacts from hearing aid test measurements.
An inference API for this model is available on other repo.
Audio artifacts are undesired sounds originating from within the device that appear at the output signal independently of the input audio. They are generated by the device’s internal electronics as a byproduct of their operation. A primary cause of these artifacts is the use of a shared power supply between the audio engine and other electronic components, which introduces interference into the audio signal path.
Figure 1: Example of no artifact recording.
Figure 1: Example of clearly visible artifact in recording.
The final training approach involved fine-tuning the models on 5-second audio segments with a 0.25 overlap, using an augmented training dataset. Due to the periodic nature of the audio artifacts, a 5-second window was sufficient to capture the full temporal context.
Training was performed on the AzureML platform using an NVIDIA A100 GPU. The models were trained on real production audio samples (private data); therefore, the original dataset cannot be publicly released.
The obtained results are promising and the model will be further evaluated in a production environment for upcoming hearing aid platforms.
| Model | F1-score | Precision | Recall | AUC |
|---|---|---|---|---|
| Wavegram-logmel | 0.9532 | 0.9641 | 0.9426 | 0.9862 |
| Resnet54 | 0.9466 | 0.9918 | 0.9053 | 0.9900 |
Future work will focus on exploring attention-based detection mechanisms, and investigating a hybrid architecture combining the wavegram approach with attention modules.
Jedrzej Ogrodowski

