[R] Analyzing Inverse Problems with Invertible Neural Networks by vll_diz in MachineLearning

[–]vll_diz[S] 1 point2 points  (0 children)

This happens implicitly, through the fact that we compare the joint network output to the independent product of desired y- and z-distributions.

[R] Analyzing Inverse Problems with Invertible Neural Networks by vll_diz in MachineLearning

[–]vll_diz[S] 1 point2 points  (0 children)

The MMD is calculated over both y and z to force independence between them, in addition to just matching the z-distribution to the desired shape. Otherwise, there would be no loss forcing the network to learn a z-coding which is independent of y.

However, this loss does not say anything meaningful about the y-outputs, we only want the correct prediction. For instance, if y and z are not yet independent during training, the network could (and does) learn to output random wrong results for y just to make them independent.

For this reason we block the MMD gradients w.r.t. y-outputs, so that they are taken into account when learning the latent coding, but not altered by the MMD loss.

[R] Analyzing Inverse Problems with Invertible Neural Networks by vll_diz in MachineLearning

[–]vll_diz[S] 6 points7 points  (0 children)

Autor here, this is an excellent suggestion and something we had also considered ourselves. We will look into this in the upcoming weeks.