Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

collector memory problems #7

Open
megies opened this issue Jul 19, 2017 · 2 comments
Open

collector memory problems #7

megies opened this issue Jul 19, 2017 · 2 comments

Comments

@megies
Copy link
Contributor

megies commented Jul 19, 2017

When I run the collector on a large number of files (e.g. 120k, some with high sampling rates like 2000Hz for a full day with up to 400-500 MB), it fills memory and swap and once both is full the script gets killed (supposedly by the OS itself, since it only shows "Killed" on command line and no pythonic MemoryError).

@Jollyfant
Copy link
Collaborator

Jollyfant commented Jul 19, 2017

The number of files that are being processed should not matter (well the names are kept in memory). However, the collector needs to keep all samples of the processed file and its neighbors in memory. With 2000Hz it does not surprise me that your machine runs out of memory (depending on how much you have). I can look in to making some optimizations but the bulk of the memory consumption will be necessary as each sample is 32-bit in the end (excluded whatever overhead is required for the actual computation of all metrics). For now I'd say skip 2000 Hz data..

@megies
Copy link
Contributor Author

megies commented Jul 20, 2017

the collector needs to keep all samples of the processed file and its neighbors in memory

Is it really calculating metrics (like stdev) over the full day?

For now I'd say skip 2000 Hz data..

Yes, skipping for now..

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants