AcouPipe is a Python toolbox for generating unique acoustical source localization and characterization datasets with Acoular that can be used for training of deep neural networks and machine learning.
AcouPipe supports distributed computation with Ray and comes with two default datasets, which enables to create data for machine learning on the fly! A pre-built Docker container can be downloaded from DockerHub.
The toolbox documentation can be found here.
Currently, AcouPipe provides two default classes to generate microphone array datasets:
- DatasetSynthetic is a simple and fast method that relies on synthetic white noise signals and spatially stationary sources in anechoic conditions.
- DatasetMIRACLE relies on a large-scale set of measured spatial room impulse responses from the MIRACLE dataset, acquired at the TU Berlin anechoic chamber, and synthetic source signals resulting in a realistic and quasi-infinite dataset.
Instead of raw time-data, only the necessary input features for machine learning are stored.
This allows the user to create data sets of manageable size that are portable, or even to create data on the fly, and facilitate reproducible research.
See the latest performance benchmarks on DatasetSynthetic for the most computational demanding features:
Users can cite the package in their contributions by referring to Kujawski and Sarradj, (2023):
@article{Kujawski2023, author = {Kujawski,Adam and Pelling, Art J. R. and Jekosch, Simon and Sarradj,Ennes}, title = {A framework for generating large-scale microphone array data for machine learning}, journal = {Multimedia Tools and Applications}, year = {2023}, doi = {10.1007/s11042-023-16947-w} }
AcouPipe is licensed under the terms of the BSD license. See the file "LICENSE" for more information.