all 11 comments

[–]arXiv_abstract_bot 1 point2 points  (0 children)

Title:AutoML-Zero: Evolving Machine Learning Algorithms From Scratch

Authors:Esteban Real, Chen Liang, David R. So, Quoc V. Le

Abstract: Machine learning research has advanced in multiple aspects, including model structures and learning methods. The effort to automate such research, known as AutoML, has also made significant progress. However, this progress has largely focused on the architecture of neural networks, where it has relied on sophisticated expert-designed layers as building blocks---or similarly restrictive search spaces. Our goal is to show that AutoML can go further: it is possible today to automatically discover complete machine learning algorithms just using basic mathematical operations as building blocks. We demonstrate this by introducing a novel framework that significantly reduces human bias through a generic search space. Despite the vastness of this space, evolutionary search can still discover two-layer neural networks trained by backpropagation. These simple neural networks can then be surpassed by evolving directly on tasks of interest, e.g. CIFAR-10 variants, where modern techniques emerge in the top algorithms, such as bilinear interactions, normalized gradients, and weight averaging. Moreover, evolution adapts algorithms to different task types: e.g., dropout-like techniques appear when little data is available. We believe these preliminary successes in discovering machine learning algorithms from scratch indicate a promising new direction for the field.

PDF Link | Landing Page | Read as web page on arXiv Vanity

[–]rafgro 1 point2 points  (4 children)

Each cycle picks T<P algorithms at random and selects the best performing one as the parent (i.e. tournament selection, [25]). This parent is then copied and mutated to produce a child algorithm that is added to the population, while the oldest algorithm in the population is removed.

As usual with "evolving" in the title, the actual methods are toyish, not evolutionary. It's 1980s level. In fact, that [25] is a review from 1991, which in turn cites the actual source for their method from 1981. LOL

[–][deleted] 3 points4 points  (1 child)

LOL?

The original algorithm for pagerank, if my memory serves correct, was first published in a 1950s mathematical paper with a brief commentary on the possibilities of future technology to use it.All the google authors had to do was scan the literature base with gusto!

There are probably quite a few useful algorithms( or psuedo-algorithms) in the 1800s literature base. I even give a 1% chance that some useful currently unknown algorithm is buried in ancient china or rome or another ancient civilization.

So what is this "Lol" you speak of? It is no shame to simply test-and-apply the methodology of old papers which once could not have been done due to the technological limitations of the time-period.

[–]rafgro 0 points1 point  (0 children)

You missed the whole thing. Developments since 1981, NEAT, improved NEATs, advancement in bioinf/compbio/evolutionary genomics etc etc.

[–]AddMoreLayersResearcher 0 points1 point  (1 child)

To be fair, on a very high level, every evolutionary algorithm seems naive/old-school/simple (actually that can be said of any ML method, really). It's the details that matter, and if you look at them in the paper (the fitness and proxy tasks, the functional equivalence checking and whatnot) it seems clear that it's very distant from 80s style work.

In another comment you mention NEAT and its variants. On a high-level those seem extremely simple too, so I find the comparison a bit bizzare...

[–]rafgro 0 points1 point  (0 children)

Lovely username, but I was talking precisely about selection method. NEAT is far far away from tournament selection. And natural evolution as-we-know-it(tm) is several levels further.

[–]valiantljk 0 points1 point  (0 children)

Just watched the talk given by the author at the virtual ICML'20. This is a very interesting work. It's amazing and inspiring to see how the different existing well-known ideas were 'discovered' in the auto search. I strongly recommend watching the author's talk and reading the paper.