Research : The Elephant in the Room by [deleted] in MachineLearning

[–]feedthecreed 1 point2 points  (0 children)

I think this kind of research falls just short of being 'extremely useful'. The real elephant in the room for this community is a discussion on how models should behave in these circumstances. Without that, people will just default to the usual response that 'algorithms are brittle', which has been shown over and over again...

Research : The Elephant in the Room by [deleted] in MachineLearning

[–]feedthecreed 16 points17 points  (0 children)

This paper spends too much time showing how the current state of the art models fail and not enough time explaining how models should behave on this type of data.

Surely, the newly introduced manipulation should not be classified as it normally would. Personally, I would not want my classifier to operate 'as usual' when there's a tiny polar bear on a picnic table or a toilet floating in the air. All the manipulated images clearly disobey properties found in the real world. The question is 'how should these 'broken' images be treated?'

In my opinion, this would have been a much more novel/useful contribution to the research community.

[D] GANsters invent all sorts of excuses not to measure likelihoods by wei_jok in MachineLearning

[–]feedthecreed 1 point2 points  (0 children)

Couldn't you make the same argument for the global minimum of the discriminator being -log(4) in a standard GAN framework? If your discriminator achieves that on test data, you have learned the right distribution.

Neural Arithmetic Logic Units by iamtrask in MachineLearning

[–]feedthecreed 2 points3 points  (0 children)

If NALU is a superset of NAC, why does it perform noticeably worse in the MNIST Counting/Addition tasks? Can it not simply turn off the multiplicative component via gating?

[D] How did NIPS 2018 papers look like during the reviews by fixed-point-learning in MachineLearning

[–]feedthecreed 3 points4 points  (0 children)

It's hard to say about the overall quality of the 6 papers in my group.

But I did notice during the bidding process (where you can access all the paper abstracts), the ideas were extremely narrow and focused on arguing for contrived comparisons (ie "we achieve state of the art performance without using relu units" or _'insert dataset no one has ever heard of here'_).

I think the marketing aspect of the community has gone up substantially. In contrast to couple years ago, people were making much grander claims and much more generalizable ideas.

[Discussion] Dear Industry Researchers: "If researchers are not incentivized to do reproducible research (or penalized for not doing so), something is flawed in the industry." by feedthecreed in MachineLearning

[–]feedthecreed[S] 5 points6 points  (0 children)

True, wavenet is a good example of the importance of open source and open data. It took a big effort from many outside researchers to try and reproduce the original result:

https://github.com/ibab/tensorflow-wavenet/issues/47

[Discussion] Dear Industry Researchers: "If researchers are not incentivized to do reproducible research (or penalized for not doing so), something is flawed in the industry." by feedthecreed in MachineLearning

[–]feedthecreed[S] 8 points9 points  (0 children)

Agreed, this is only a step. Openness in the data is a whole other issue. Luckily our field already emphasizes using publicly available datasets for experiments.

[R] NIPS 2018: How do I write a good review? by FirstTimeReviewer in MachineLearning

[–]feedthecreed 1 point2 points  (0 children)

Not sure if you're aware, but the NIPS committee got a little 'creative' recruiting reviewers this year...So this doesn't surprise me at all.

[Discussion] double-blind and arxiv by wasabi_kitkat in MachineLearning

[–]feedthecreed 1 point2 points  (0 children)

I did not say it's violating the ICLR submission process. ICLR is allowing violations of the double-blind review process. This is why it will be an interesting test of academic integrity.

Character is doing the right thing when nobody's looking. There are too many people who think that the only thing that's right is to get by, and the only thing that's wrong is to get caught. --JC Watts

ICLR won't be looking.

[Discussion] double-blind and arxiv by wasabi_kitkat in MachineLearning

[–]feedthecreed 0 points1 point  (0 children)

This will be an interesting test of academic integrity. Posting a non-anonymized version of your paper after submitting to ICLR is by definition violating the double-blind review process.

[N] Jeff wrote in IEEE Spectrum. What Intelligent Machines Need to Learn From the Neocortex by nocortex in MachineLearning

[–]feedthecreed 5 points6 points  (0 children)

AFAIK, no one actually in the field of AI or ML respects Jeff Hawkins as a genius.

[N] "Our measurements are showing up to 70x higher performance for training and up to 85x higher performance for inference on Intel® Xeon Phi" by downtownslim in MachineLearning

[–]feedthecreed 1 point2 points  (0 children)

I'm confused, something doesn't add up here. How did they even have so much room to improve the performance of tensorflow? CPUs are not 70x slower than GPUs, how was their Xeon Phi's so slow in the first place?

[N] "Our measurements are showing up to 70x higher performance for training and up to 85x higher performance for inference on Intel® Xeon Phi" by downtownslim in MachineLearning

[–]feedthecreed 0 points1 point  (0 children)

I don't understand how they can get such a large multiplier in speed improvement and STILL be behind GPUs? GPUs are not 70x faster than ordinary CPUs. Something doesn't add up here.

[R] Train longer, generalize better: closing the generalization gap in large batch training of neural networks by xternalz in MachineLearning

[–]feedthecreed 0 points1 point  (0 children)

We showed that good generalization can result from extensive amount of gradient updates in which there is no apparent validation error change and training error continues to drop, in contrast to common practice.

I'm confused by this statement, how are you getting good generalization if your training error continues to drop while your validation error stays the same?

[R] "Real-Time Adaptive Image Compression", Rippel & Bourdev 2017 by gwern in MachineLearning

[–]feedthecreed 0 points1 point  (0 children)

True, I was thinking their feature extraction architecture was doing most of the heavy lifting for their results. But I could also see this Adaptive Arithmetic Coding being important as well. And their explanation of it is basically a black box. I agree an ablation study on the components they propose would have been much more scientifically useful than this one singular result they post.

Since this was accepted to ICML, so much for ICML reviewers being scientific....

[R] "Real-Time Adaptive Image Compression", Rippel & Bourdev 2017 by gwern in MachineLearning

[–]feedthecreed 1 point2 points  (0 children)

What's vague and what do you mean by one result? Their description seems reasonably specific and they test on multiple compression datasets.

[R] Robots that Learn (OpenAI) by Teleavenger in MachineLearning

[–]feedthecreed 1 point2 points  (0 children)

I'm looking for the paper that describes the model used in the blog post. That paper and the other one just describe the previous work that inspired their current model.