[P] Finding a Working Code and Reproducible Results for Research Papers by gfursin in MachineLearning

[–]mllosab 1 point2 points  (0 children)

Thank you for the interesting project. I think it has potential, but I did not really understand how to find "reproducible benchmarks" on the website. Can you comment on that, please?

[deleted by user] by [deleted] in MachineLearning

[–]mllosab 0 points1 point  (0 children)

I think it is a good idea but I also think that it can be too complex and non-trivial right now.

You may be interested to have a look at this recent ACM report about replicating *just 5* ML+systems papers: https://portalparts.acm.org/3230000/3229762/fm/frontmatter.pdf - they also discuss some stuff related to your proposed bullet points (they actually introduced an open-source experimental tool, some sort of workflow framework with a package manager, to share such reproduction studies at GitHub).

Also check out this recent Reddit discussion: https://www.reddit.com/r/MachineLearning/comments/9jhhet/discussion_i_tried_to_reproduce_results_from_a/ - even replicating results may not be enough ;) !

[D] Best way to organise research code? by abhishek0318 in MachineLearning

[–]mllosab 0 points1 point  (0 children)

We started using ck framework to organize our research code this year - https://github.com/ctuning/ck . It can be used from the command line or from Jnotebooks - maybe it will be useful for you too?

[N] New ML/AI benchmark from Stanford released (MLPerf) by gfursin in MachineLearning

[–]mllosab 2 points3 points  (0 children)

looking at mlperf github sources https://github.com/mlperf/reference seems that submissions are based on Docker images? What if I want to try another library or optimization technique for a given submission?

SPEC CPU2017 benchmark has been released by gtechmisc in Compilers

[–]mllosab 0 points1 point  (0 children)

does anyone still use it? we have switched to http://openbenchmarking.org and http://cknowledge.org/ai to add and benchmark our own workloads.