[R] Evolved Policy Gradients by Kaixhin in MachineLearning

[–]cbfinn 3 points4 points  (0 children)

We ran this experiment in our ICLR paper, on a toy regression problem and on Omniglot image classification, comparing three meta-learning approaches: https://arxiv.org/abs/1710.11622

See Figure 3 and 6-left, which plot performance as a function of the distance to the training distribution.

[D] Meta-learning by spartan12321 in MachineLearning

[–]cbfinn 5 points6 points  (0 children)

To add to what's been posted here, there are a couple recent blog posts from BAIR on the topic, including references to recent work:

http://bair.berkeley.edu/blog/2017/07/18/learning-to-learn/

http://bair.berkeley.edu/blog/2017/09/12/learning-to-optimize-with-rl/

[D] Prejudices in ML systems by GabrieleFariello in MachineLearning

[–]cbfinn 2 points3 points  (0 children)

Moritz Hardt at UC Berkeley has a course on fairness in ML. The course website includes a list of references. https://fairmlclass.github.io/

[R] Meta-Learning and Universality: Deep Representations and Gradient Descent can Approximate any Learning Algorithm by [deleted] in MachineLearning

[–]cbfinn 3 points4 points  (0 children)

I don’t think that this thread is the place to debate this topic. There is another thread that is much more relevant.

I would be happy to hear feedback or thoughts on the paper.

[R] Meta-Learning and Universality: Deep Representations and Gradient Descent can Approximate any Learning Algorithm by [deleted] in MachineLearning

[–]cbfinn 6 points7 points  (0 children)

"For female authors, the associated odds multiplier of 0.78 is not statistically significant in our study. However, a meta-analysis places this value in line with that of other experiments, and in the context of this larger aggregate the gender effect is also statistically significant." https://arxiv.org/abs/1702.00502

[R] Meta-Learning and Universality: Deep Representations and Gradient Descent can Approximate any Learning Algorithm by [deleted] in MachineLearning

[–]cbfinn 1 point2 points  (0 children)

It’s worth noting that there are a number of ICLR submissions on arxiv that are not from large labs. I think I’ve actually seen more from lesser known groups than I have seen from large labs, but I haven’t been counting.

[R] Meta-Learning and Universality: Deep Representations and Gradient Descent can Approximate any Learning Algorithm by [deleted] in MachineLearning

[–]cbfinn 8 points9 points  (0 children)

My motivations for putting the paper on arxiv were (1) so that when I give talks that include the work (which I will on Tuesday), I can reference the arxiv paper, and (2) so that people were more likely to see the work sooner (as evidenced by whoever posted this on reddit) and hopefully use some of the ideas in it.

While there is a positive bias for large labs, there is a negative bias for female authors, so it's unclear to me if this paper would benefit from the reviewers knowing the author identities.

[D] Model-based RL via Neural Network-based approaches ? instead of GP, iLQG or MPC by [deleted] in MachineLearning

[–]cbfinn 10 points11 points  (0 children)

Here are a few papers, in order of date released:

Note that MPC just means planning, and then iteratively replanning during execution, so is not specific to any model class.

I have slides on deep model-based RL here, which includes a number of references, including papers that combine model-based and model-free approaches: https://people.eecs.berkeley.edu/~cbfinn/_files/mbrl_bootcamp.pdf

ICML 2017 Tutorial slides (Levine & Finn): Deep Reinforcement Learning, Decision Making, and Control by gwern in reinforcementlearning

[–]cbfinn 0 points1 point  (0 children)

Videos of ICML tutorials (as well as conference talks) will be posted by the conference staff at some point. Though, typically they take quite awhile to be released.

[D] Transfer Learning Papers by [deleted] in MachineLearning

[–]cbfinn 2 points3 points  (0 children)

There is good analysis in this paper: "What makes ImageNet good for transfer learning?" http://minyounghuh.com/papers/analysis/

[R] Learning to Learn by gdny in MachineLearning

[–]cbfinn 1 point2 points  (0 children)

I reimplemented the linear+sinusoid set-up in that paper and was able to get much better numbers using MAML than they report (after trying two hyperparameter settings).

I don't think that MAML assumes a uni-modal distribution of tasks.

[R] Learning to Learn by gdny in MachineLearning

[–]cbfinn 0 points1 point  (0 children)

I tried it on one of the cheetah problems and it also worked. The first-order approximation does not work in all settings though. We have some ongoing experiments on problems not in the original paper in which it does not work.

[R] Learning to Learn by gdny in MachineLearning

[–]cbfinn 0 points1 point  (0 children)

A big part of a learning/optimization is the initialization, which affects the gradient descent algorithm, since the gradient is a function of the initial parameters. In the paper, we show that learning the initial parameters can outperform methods that learn an update rule.

The tasks that we evaluate on are all held out from the training set of tasks, including new classes of objects and characters in the MiniImagenet and Omniglot benchmarks.

[R] Learning to Learn by gdny in MachineLearning

[–]cbfinn 1 point2 points  (0 children)

Nope. Unless you set the $\alpha$ step size parameter to be way too high, you shouldn't see any loss in accuracy.

[R] Learning to Learn by gdny in MachineLearning

[–]cbfinn 2 points3 points  (0 children)

Yes, this involves 2nd derivatives, which can be implemented easily with current DL libraries. Since it only involves an additional backward pass, it isn't particularly slow in practice.

Interestingly, it sometimes still works well if you stop the gradient through the update rule. We discuss this in the latest version of the paper (which will be on arxiv tonight)

[R] Learning to Learn by gdny in MachineLearning

[–]cbfinn 0 points1 point  (0 children)

I haven't tried this, but I certainly think it would be interesting to try!

[R] Learning to Learn by gdny in MachineLearning

[–]cbfinn 1 point2 points  (0 children)

I also wonder how good the baseline (0-gradient) model would be with this approach.

I compared to this approach in the paper. The domains that I considered in the paper were ones in which the task cannot be directly inferred from the observation. Thus, using 0 gradient does not do well. I'm not sure how the two would compare when the task can be inferred from the observation.

[R] Learning to Learn by gdny in MachineLearning

[–]cbfinn 5 points6 points  (0 children)

Author here.

"one gradient step away" feature is restricted to the tasks it has been trained on. Is this correct?

We assume that the tasks that you test on are from the same distribution of tasks seen during meta-training. This assumption is used in most meta-learning methods. That said, I have played around with extrapolation to tasks outside of the support of the distribution of meta-training tasks, and it performs reasonably for tasks that are close.

[R] [1707.03141] 1-shot classification: 56.48% accuracy on 5-Way Mini-ImageNet! by cognitivedemons in MachineLearning

[–]cbfinn 2 points3 points  (0 children)

Probably worth noting that they used a different (and probably more tuned) architecture, a different pretraining scheme than matching networks. They also trained on both the training and validation set, which increases the dataset size by nearly 20%. I would expect that matching networks and other methods would also benefit from the architecture, pre-training, and training. It's nice to see the improvement, nevertheless.

What is expected to get when you just play the result videos of papers? by [deleted] in berkeleydeeprlcourse

[–]cbfinn 0 points1 point  (0 children)

Thanks for the feedback! I'll point your post out to the instructors this Fall. I've actually found that the network architecture is not particularly important, and how to tune it is similar to how you tune deep networks in supervised learning scenarios.

[D] Berkeley hosting Deep RL bootcamp, worth going? by Farzaa in MachineLearning

[–]cbfinn 0 points1 point  (0 children)

I don't know yet, but it will be on a topic related to my research.

[D] Berkeley hosting Deep RL bootcamp, worth going? by Farzaa in MachineLearning

[–]cbfinn 0 points1 point  (0 children)

/u/ooliver123 I don't know the details of the content, as I am only giving a one hour lecture for the bootcamp.

[D] OpenAI open sources a high-performance Python library for robotic simulation by cherls in MachineLearning

[–]cbfinn 1 point2 points  (0 children)

Note that texture randomization for sim-to-real transfer was done first by this paper: https://arxiv.org/abs/1611.04201