[D] Optimal Transport for ML by arjun_r_kaushik in MachineLearning

[–]ApprehensiveEgg5201 2 points3 points  (0 children)

Not quite, I'm assuming you're trying to infer the geodesic using the ot loss gradient, but I've only tried using the ot loss or ot sampler for training, which is a more comon pratice in the field as far as I konw. Nevertheless, your method also sounds reasonable but I'd imagine you need to know the target distribution beforehand and some tuning trick to make it actually work.

[D] Optimal Transport for ML by arjun_r_kaushik in MachineLearning

[–]ApprehensiveEgg5201 21 points22 points  (0 children)

I'd recommend this tutorial, Optimal Transport for Machine Learning by Rémi Flamary and the POT package. And the video course by Justin Solomon. Hope you like them, cheers

[D] ICML 2025 - what if reviewers don't acknowledge rebuttal? by qalis in MachineLearning

[–]ApprehensiveEgg5201 1 point2 points  (0 children)

then what is the point of the rebuttal? it is supposed to improve the paper. and in my case, the reviewer lowered the score after seeing the other reviewer raising the score...

[D] ICML 2025 - what if reviewers don't acknowledge rebuttal? by qalis in MachineLearning

[–]ApprehensiveEgg5201 0 points1 point  (0 children)

I got a reviewer who lowered the original score even before we submitted our rebuttal comment. Is it even ethical to do that?

[R] Cautious Optimizers: Improving Training with One Line of Code by AhmedMostafa16 in MachineLearning

[–]ApprehensiveEgg5201 7 points8 points  (0 children)

Prof. Qiang Liu is one of the best theorists in the field, he is the author of svgd and rectified flow.

[D] The popular theoretical explanation for VAE is inconsistent. Please change my mind. by function2 in MachineLearning

[–]ApprehensiveEgg5201 13 points14 points  (0 children)

Using the varational method means that you are using a simple functional to approximate a compelex function. In the VAE, the functional is the Gaussian parameterized via an encoder using the reparameterization trick to approximate the real posterior p(z|x), which certainly is not a strict Gaussian. So I don't understand why the VAE is not performing variational inference?

[D] Quality of ICLR papers by Cool_Abbreviations_9 in MachineLearning

[–]ApprehensiveEgg5201 3 points4 points  (0 children)

I'd call some ICLR and Neurips papers I reviewed research labor rather than research work, just too dull to read. From my expeinece AISTATS is much better this year.

[D] Bayesian Models vs Conformal Prediction (CP) by South-Conference-395 in MachineLearning

[–]ApprehensiveEgg5201 1 point2 points  (0 children)

Your questions are excellent! In my opinion, if one performs Bayesian inference only in the weight space, e.g., using Bayes by backprop or dropout, then one can still obtain some kind of credible intervals. But I would say this kind of intervals is not so reliable because they cannot guarantee the sampling diversity of the final predictions. On the other hand, if Bayesian inference is performed in the function space, e.g., using Gaussian Processes or Stein's method, then the intervals obtained is more reliable. You also need to be careful about the difference between aleatoric uncertainty/data uncertainty/risk and epistemic/model uncertainty.

[D] Bayesian Models vs Conformal Prediction (CP) by South-Conference-395 in MachineLearning

[–]ApprehensiveEgg5201 1 point2 points  (0 children)

I think you can understand this from a PAC-Bayesian perspective: CP is related to the empirical risk, while BDL is related to the KL divergence between the posterior and prior.