Replies: 4 comments 1 reply
-
Hey! Thanks for reaching out!
Hope this helps! |
Beta Was this translation helpful? Give feedback.
-
Beta Was this translation helpful? Give feedback.
-
Great, thanks a lot for the answer! I could imagine that the issue with memory is solved by avoiding the computation tree to be tracked here (i.e. by wrapping this in a Michael |
Beta Was this translation helpful? Give feedback.
-
Hi @Aranka-S, did you find a solution to this problem in the end? |
Beta Was this translation helpful? Give feedback.
Uh oh!
There was an error while loading. Please reload this page.
Uh oh!
There was an error while loading. Please reload this page.
-
Hi all,
I've been working on using NPE with an embedding net instead of summary statistics as input to the NDE. Addionally, my dataset (multivariate time-series) that is now the input to the embedding net is too large to keep in memory all at once.
With the new training interface, I got the training up and running with an own dataloader (I am using the SBI release v.0.23.3). However, I still have questions about aspects of the pipeline:
build_maf
). But I cannot pass all my data of course. Is there any elegant method to still incorporate z-scoring in the sbi-pipeline based on what's in the toolbox? Or should I perform z-scoring as a form of preprocessing beforehand and then make sure I keep that info to z-score all observations on which I want to run inference later (can I save the parameters in the MAF somehow after pre-computing them)? Do you have any insights in how large the impact would be if z-scoring would be computed based on just a small subset of the data (I guess passing only a subset the size of what fits into memory during construction would be a solution, though not the prefered one as the computation of the z-score parameters would be based on very little data)?.map
procedure is also very high. I'm guessing the slowness is just a "normal" consequence of the larger NN (embedding net+NDE) that has to be evaluated? But I find the memory usage increase harder to grasp, could anybody help me with understanding this? I did find that the peak in memory usage seems to decrease if I lower the num_init samples, but I can't really explain why this has such a large influence (Is the complete neural network saved for each initial sample somehow?). It might be related to the answer in this discussion? Any thoughts on this?And maybe even better: is there anything I can do to mitigate either the runtime or memory usage increase? I could compute the MAP on GPU if there is a way around the memory usage?
This became quite an overload in questions it seems, but I would be grateful for any insights or help this community can provide.
Beta Was this translation helpful? Give feedback.
All reactions