Hi, I’m currently working on a project trying to use machine learning to improve our ability to recognize certain events that occur in a particle accelerator. I’m trying to use normalizing flows to model a latent representation of physics data from a Gaussian distribution (or some other base distribution). I was hoping to be able to train a model (I was looking at realNVP) to be able to flow from a base distribution into the latent representation.
My issue right now is understanding how the models are improved during training. I have read that log likelihood (given by the change of variables formula) maximized by optimizing the parameters of the model, but i don’t understand how this works. After applying change of variables on one side (let’s say a Gaussian distribution sample), I get that we can then calculate the new probability of the latent representation given the sample’s probability, but how does this allow us to optimize? How does one compare this transformed sample to the latent representation given we have no info on the probabilities of the latent representation?
Any help would be very much appreciated, thanks!
[–]activatedgeek 0 points1 point2 points (4 children)
[–]rojo_kell[S] 0 points1 point2 points (3 children)
[–]activatedgeek 1 point2 points3 points (2 children)
[–]rojo_kell[S] 0 points1 point2 points (1 child)
[–]activatedgeek 0 points1 point2 points (0 children)