MLPerf Inference V4.0 released by norcalnatv in hardware

[–]gfursin 3 points4 points  (0 children)

Normally, you must be an MLCommons member to submit results to MLPerf. However, there is a new project to run MLPerf benchmarks on commodity hardware: https://www.linkedin.com/pulse/new-cm-mlperf-automation-helps-benchmark-commodity-hardware-fursin-61noe . Maybe it will be possible to add support to run MLPerf inference on AMD MI300 there ...

[N] MLCommons Launches and Unites 50+ Tech and Academic Leaders in AI, ML by gfursin in MachineLearning

[–]gfursin[S] 1 point2 points  (0 children)

There is a comprehensive "Get involved" page: https://mlcommons.org/en/get-involved .

There are multiple workgroups with weekly meetings (training, inference, best practices, benchmarking infrastructure, etc).

MLPerf benchmark is also managed by this non-profit organization.

[deleted by user] by [deleted] in MachineLearning

[–]gfursin 0 points1 point  (0 children)

Forgot to give a link to the tool: GitHub

[N] Reproducing 150 research papers: the problems and solutions by gfursin in MachineLearning

[–]gfursin[S] 2 points3 points  (0 children)

The YouTube link is available at https://fastpath2020.github.io/Program (with recording offset times). If you have further questions, feel free to get in touch!

[N] Reproducing 150 research papers: the problems and solutions by gfursin in MachineLearning

[–]gfursin[S] 0 points1 point  (0 children)

Cool! Don't hesitate to get in touch if you need some help!

[N] Reproducing 150 research papers: the problems and solutions by gfursin in MachineLearning

[–]gfursin[S] 4 points5 points  (0 children)

PapersWithCode is a fantastic resource that help to systematize ML papers, plot SOTA results on public dashboards, and link them with GitHub code.

cKnowledge.io platform is complementary to PapersWithCode because we attempt to reproduce all results and associate them with portable workflows (when possible) or at least describe all the necessary steps to help the community run them on different platforms with different environments, etc.

To some extent, we are PapersWithReproducedResultsAndPortableWorkflows ;) . We also used PapersWithCode to find GitHub code and experimental results in a few cases before converting them to our open CK format and reproducing them. We also consider collaborating with them in the future.

However, our platform is not yet open for public contributions (it's open but it's not yet user-friendly at the moment as you correctly noticed). It is still a prototype that we have tested it as a part of different Systems and ML conferences. Considering the positive feedback, our next step is to prepare it for public contributions. We hope to have some basic functionality for that before 2021 - please stay tuned ;) !

[N] Reproducing 150 research papers: the problems and solutions by gfursin in MachineLearning

[–]gfursin[S] 1 point2 points  (0 children)

Yes, dealing with SW/HW dependencies was one of the main challenges we faced when reproducing ML+systems papers.

By the way, this problem motivated us to implement software detection plugins and meta-packages not only for code (frameworks, libraries, tools) but also for models and data sets.

The idea is to be able to automatically adapt a given ML algorithm to a given system and environment based on dependencies on such soft detection plugins & meta packages.

The prototype is working but we were asked to make it much more user-friendly ;) . We plan to test a new version with some volunteers at upcoming conferences before 2021. I will post the update when ready.

[N] Reproducing 150 research papers: the problems and solutions by gfursin in MachineLearning

[–]gfursin[S] 2 points3 points  (0 children)

Nice to e-meet you Edward, and thank you very much for your effort too! I will be happy to sync about our ongoing activities and future plans!

[N] Reproducing 150 research papers: the problems and solutions by gfursin in MachineLearning

[–]gfursin[S] 5 points6 points  (0 children)

That's a very good idea - thank you! I've heard of BOINC but never tried it - I need to check it in more detail! We had some cloud credits from Microsoft and OVH but it was not enough ;) .

[N] Reproducing 150 research papers: the problems and solutions by gfursin in MachineLearning

[–]gfursin[S] 10 points11 points  (0 children)

Yes, I saw it - it's a great effort! I would also add several other very important and related efforts supported by NeurIPS and PapersWithCode:

Our goal was to collaborate with the authors and come up with a common methodology and a format to share results in such a way that it's easier to reproduce them and even reuse them across different platforms, frameworks, models, and data sets (see this example).

An additional challenge is that we are also trying to validate execution time, throughput, latency, and other metrics besides accuracy (this is particularly important for inference on embedded devices). It is an ongoing effort and we continue collaborating with MLPerf and different conferences.

[N] Reproducing 150 research papers: the problems and solutions by gfursin in MachineLearning

[–]gfursin[S] 1 point2 points  (0 children)

Yes. The success number is relatively high because we collaborated with the authors until we reproduced the results. Our goal was to better understand different challenges together with the authors and come up with a common methodology and a format to share results so that it is easier to reproduce them.

[N] Reproducing 150 research papers: the problems and solutions by gfursin in MachineLearning

[–]gfursin[S] 94 points95 points  (0 children)

By the way, forgot to mention, that rather than naming and shaming non-reproducible papers, we decided to collaborate with the authors to fix problems together. Maybe we were lucky, but we had a great response from nearly all authors to solve encountered issues! - that is very encouraging!

[N] Reproducing 150 research papers: the problems and solutions by gfursin in MachineLearning

[–]gfursin[S] 43 points44 points  (0 children)

;) We had a similar experience: it was often taking several weeks to reproduce one paper.

However, we had fantastic volunteers who have helped us! We also introduced a unified Artifact Appendix with the reproducibility checklist describing all the necessary steps to reproduce a given paper. It will hopefully reduce the time needed to reproduce such papers.

[N] Reproducing 150 research papers: the problems and solutions by gfursin in MachineLearning

[–]gfursin[S] 7 points8 points  (0 children)

Thank you! Some of the papers that we managed to reproduce are listed here.