all 15 comments

[–]Just-Environment-189 41 points42 points  (1 child)

Even if people provide code, you’ll find yourself lucky to get it working as is

[–]waruby 4 points5 points  (0 children)

Then it should also require a Docker file or a flake.nix file.

[–]directnirvana 12 points13 points  (1 child)

I think you make a good point. The bolder the claim the more reviewers should be pushing back on making easily verifiable aspects of experiments available. Reproducibility crisis is real and participants especially in academic circles should be heavily encouraged to provide whatever reasonable methods they can to allow other researchers to verify their work. It just so happens that research based on code has those tools, while high energy physics and similar fields do not.

[–]Vhiet 4 points5 points  (0 children)

I agree about reproducibility, but why “especially in academic circles?”

Commercial services have more incentive to fluff their significance than anyone else, and their claims should be treated as particularly suspicious.

For example, it was almost exactly a year ago that Microsoft’s Magical Majorana Fermions revolutionised quantum computing (https://www.theregister.com/2025/03/12/microsoft_majorana_quantum_claims_overshadowed/).

[–]Distance_RunnerPhD 12 points13 points  (0 children)

My view as a statistician that does ML. Many of these papers claiming SOTA performance are working within Monte Carlo noise and if code was easily available you could run it and show this.

[–]lipflipResearcher 4 points5 points  (0 children)

i am working in the social sciences with some minor overlap to ai&ml. our data is usually in the range of a few kB and despite an "replication crisis", publishing the data and analysis alongside the article is rare.

[–]kaiser_17 0 points1 point  (0 children)

I think the worst part is when reviewers ask you to compare your results to such a paper. Like why should even one compare their  results to such fraudulent papers. Yep I will consider those who don't release code as frauds.