Replies: 1 comment 1 reply
-
Hi @TheFilS,
I would recommend using the efficient parameters (if you don't need the high computational cost features) or reducing the chunk size first and only then, if it still takes too long, to try using parallelization on the cloud. |
Beta Was this translation helpful? Give feedback.
1 reply
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
-
Hi all,
this is my first go at using tsfresh, and although everything seems to be working, its quite slow and I'm looking for some helpful tips as to how I can speed up the process. Ultimately, I'm looking to run feature extraction on wav files (sampling rate ~48kHz, 20 second files) on a growing database that currently has about 50 files. Due to the number of data points, I'm reducing the streams into chunks using:
Unfortunately, this take a lot of time, specifically 2+ hours per file.
Trying to run feature extraction without splitting it like this fails due to memory issues due to matrix size.
Does anyone have a suggestion as to how I can speed up the process? Should I do some sort of parallelization w Azure/AWS or is my strategy just terrible?
Beta Was this translation helpful? Give feedback.
All reactions