Replies: 3 comments
-
Hey! Thanks for reaching out!
Hope this helps! |
Beta Was this translation helpful? Give feedback.
0 replies
-
Beta Was this translation helpful? Give feedback.
0 replies
-
Great, thanks a lot for the answer! I could imagine that the issue with memory is solved by avoiding the computation tree to be tracked here (i.e. by wrapping this in a Michael |
Beta Was this translation helpful? Give feedback.
0 replies
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
-
Hi all,
I've been working on using NPE with an embedding net instead of summary statistics as input to the NDE. Addionally, my dataset (multivariate time-series) that is now the input to the embedding net is too large to keep in memory all at once.
With the new training interface, I got the training up and running with an own dataloader (I am using the SBI release v.0.23.3). However, I still have questions about aspects of the pipeline:
build_maf
). But I cannot pass all my data of course. Is there any elegant method to still incorporate z-scoring in the sbi-pipeline based on what's in the toolbox? Or should I perform z-scoring as a form of preprocessing beforehand and then make sure I keep that info to z-score all observations on which I want to run inference later (can I save the parameters in the MAF somehow after pre-computing them)? Do you have any insights in how large the impact would be if z-scoring would be computed based on just a small subset of the data (I guess passing only a subset the size of what fits into memory during construction would be a solution, though not the prefered one as the computation of the z-score parameters would be based on very little data)?.map
procedure is also very high. I'm guessing the slowness is just a "normal" consequence of the larger NN (embedding net+NDE) that has to be evaluated? But I find the memory usage increase harder to grasp, could anybody help me with understanding this? I did find that the peak in memory usage seems to decrease if I lower the num_init samples, but I can't really explain why this has such a large influence (Is the complete neural network saved for each initial sample somehow?). It might be related to the answer in this discussion? Any thoughts on this?And maybe even better: is there anything I can do to mitigate either the runtime or memory usage increase? I could compute the MAP on GPU if there is a way around the memory usage?
This became quite an overload in questions it seems, but I would be grateful for any insights or help this community can provide.
Beta Was this translation helpful? Give feedback.
All reactions