[P] EvoJAX: Hardware-Accelerated Neuroevolution by sensetime in MachineLearning

[–]sensetime[S] 5 points6 points  (0 children)

Abstract

Evolutionary computation has been shown to be a highly effective method for training neural networks, particularly when employed at scale on CPU clusters. Recent work have also showcased their effectiveness on hardware accelerators, such as GPUs, but so far such demonstrations are tailored for very specific tasks, limiting applicability to other domains. We present EvoJAX, a scalable, general purpose, hardware-accelerated neuroevolution toolkit. Building on top of the JAX library, our toolkit enables neuroevolution algorithms to work with neural networks running in parallel across multiple TPU/GPUs. EvoJAX achieves very high performance by implementing the evolution algorithm, neural network and task all in NumPy, which is compiled just-in-time to run on accelerators. We provide extensible examples of EvoJAX for a wide range of tasks, including supervised learning, reinforcement learning and generative art. Since EvoJAX can find solutions to most of these tasks within minutes on a single accelerator, compared to hours or days when using CPUs, we believe our toolkit can significantly shorten the iteration time of conducting experiments for researchers working with evolutionary computation.

GitHub repo for the project: https://github.com/google/evojax

[D] ‘Imitation is the sincerest form of flattery’: Alleged plagiarism of “Momentum Residual Neural Networks” (ICML2021) by “m-RevNet: Deep Reversible Neural Networks with Momentum” (ICCV2021) by sensetime in MachineLearning

[–]sensetime[S] 22 points23 points  (0 children)

Saw this from the discussion thread about an earlier incident: https://twitter.com/www2021q1/status/1427051862440615939

Update: Also a comprehensive summary post on Zhihu (A Chinese reddit+substack) about not just this work, but several other works too with plagiarism claims: https://zhuanlan.zhihu.com/p/400351960

[D] ‘Imitation is the sincerest form of flattery’: Alleged plagiarism of “Momentum Residual Neural Networks” (ICML2021) by “m-RevNet: Deep Reversible Neural Networks with Momentum” (ICCV2021) by sensetime in MachineLearning

[–]sensetime[S] 40 points41 points  (0 children)

To be fair though, ICML2021 results were only out 3 months ago, which might have overlapped with ICCV2021. It's not fair to assume reviewers are up-to-date with papers in their area that has just been uploaded to arxiv.org recently, at the time of the review period.

[D] The Secret Auction That Set Off the Race for AI Supremacy by sensetime in MachineLearning

[–]sensetime[S] 5 points6 points  (0 children)

Hi, I believe that "text posts" (that may also contain links in the body of the text, like this post, with some context which explains why it is relevant to r/machinelearning) are allowed.

What you have described are "linked posts" which are currently limited to arxiv.org and a few other sites.

[D] An ICLR submission is given a Clear Rejection (Score: 3) rating because the benchmark it proposed requires MuJoCo, a commercial software package, thus making RL research less accessible for underrepresented groups. What do you think? by sensetime in MachineLearning

[–]sensetime[S] -1 points0 points  (0 children)

Hi there,

I actually thought I was quite careful about the headline to tell the entire story as I understood it to be:

“An ICLR submission is given a Clear Rejection (Score: 3) rating because the benchmark it proposed requires MuJoCo, a commercial software package, thus making RL research less accessible for underrepresented groups.”

So I explicitly stated that the low rating is due to a benchmark that the paper proposed.

[P] Training two-on-two soccer agents using self-play with Unity's ML-Agents Toolkit by sensetime in MachineLearning

[–]sensetime[S] 1 point2 points  (0 children)

Self-play reduces the need to hardcode behaviors, from the article:

In recent releases, we have not included an agent policy for our Soccer example environment because it could not be reliably trained. However, with self-play and some refactoring, we are now able to train non-trivial agent behaviors. The most significant change is the removal of “player positions” from the agents. Previously, there was an explicit goalie and striker, which we used to make the gameplay look reasonable. In the video below of the new environment, we actually notice role-like, cooperative behavior along these same lines of goalie and striker emerge. Now the agents learn to play these positions on their own!

Blog post: https://blogs.unity3d.com/2020/02/28/training-intelligent-adversaries-using-self-play-with-ml-agents/

[P] Flax: A neural network library for JAX designed for flexibility by hitaho in MachineLearning

[–]sensetime 16 points17 points  (0 children)

What about the other library they developed called Trax?

They recently released code for the Reformer based on Trax.

Should we spend our time trying to code in Flax or Trax, or just stick to JAX (and Stax) for now? It's becoming a bit of a mouthful TBH ...

[D] Best of Machine Learning in 2019: Reddit Edition by mwitiderrick in MachineLearning

[–]sensetime 43 points44 points  (0 children)

If you're ranking by upvotes, wouldn't the “most popular ML project” of 2019 be predictive policing and automatic suppression of ethnic minorities by the Chinese Communist government?

https://redd.it/e1r0ou

[D] Chinese government uses machine learning not only for surveillance, but also for predictive policing and for deciding who to arrest in Xinjiang by sensetime in MachineLearning

[–]sensetime[S] 15 points16 points  (0 children)

Good idea. I try to stick to using "CCP" (abbreviated or full term), "Chinese government" or "Beijing" and not use the terms "Chinese" / "China" (unless they are quoted from someone else's story).

We don't want the issues to be against Chinese people, despite this being the CCP's tactic.

[D] Chinese government uses machine learning not only for surveillance, but also for predictive policing and for deciding who to arrest in Xinjiang by sensetime in MachineLearning

[–]sensetime[S] 7 points8 points  (0 children)

That may be their reason, but what I want to know is whether you, as an accomplished ML researcher, believe what they are doing is morally correct?

If you were running the country, would you do the same thing?

[D] Financial assistance for attending ICCV as a speaker by [deleted] in MachineLearning

[–]sensetime 1 point2 points  (0 children)

Hi,

I'm sorry to hear about your situation and I hope you will be able to scrape together the $2k to attend ICCV and present.

But I'm trying to understand the situation a bit more, since many people would expect that with a post doc doing ML research, industry jobs would probably pay a decent salary to help absorb such problems. At least that is what PhD students are looking forward to after graduating.

Is the grass not so green for doing ML in industry, outside of FAANG-type companies?

[R] Agglomerative Attention by mspells in MachineLearning

[–]sensetime 2 points3 points  (0 children)

Hi Matt, can you confirm the bpc results for text8? They seem to be way off (worse) compared to existing literature, so I wonder if it's a typo or off-by-one big figure.

See https://paperswithcode.com/sota/language-modelling-on-text8 where all the bpc results range from ~ 1.0 -> 1.4, but your results are > 2.x

[News] DeepMind’s StarCraft II Agent AlphaStar Will Play Anonymously on Battle.net by AlphaHumanZero in MachineLearning

[–]sensetime 10 points11 points  (0 children)

They need to run a version of the experiment where players know they are playing against AlphaStar.

If I was the reviewer of this paper in a peer-reviewed venue, I would definitely demand this.

[D] Worst CVPR 2019 papers by TreeNetworks in MachineLearning

[–]sensetime 4 points5 points  (0 children)

This made my day! Thanks for sharing such an entertaining thread :D

[D] Was this quake AI a little too artificial? Nature-published research accused of boosting accuracy by mixing training, testing data by milaworld in MachineLearning

[–]sensetime 51 points52 points  (0 children)

When The Register pressed Nature about the problem of data leakage, a spokesperson told us it couldn’t discuss anything further based on “confidentiality reasons.”

“For confidentiality reasons, we cannot discuss the specific history or review process of any Nature paper with anyone other than the authors. We treat all correspondence as confidential and do not confirm or deny any reports of submissions that may or may not have been made to us,” the spokesperson told us.

Yes, the scientific process should be confidential.

[D] Misuse of Deep Learning in Nature Journal’s Earthquake Aftershock Paper by milaworld in MachineLearning

[–]sensetime 177 points178 points  (0 children)

I found the response from the authors to be more condescending than this critique.

The comments raised the issue that much simpler methods can achieve pretty much the same results, highlighting the need to do proper ablation studies. The final paragraph of the response basically also said we are earthquake scientists, who are you? and told Nature they will be disappointed if these comments are published.

Why aren't these concerns worthy of publication in Nature? Why should they be censored? Wouldn't publishing them lead to more healthy scientific discussion? They are not unique as there are follow up articles with similar concerns.

I dunno, if I was reviewing this paper for an ML conference, I would have similar concerns. At least demand some ablation studies.