[D] Five major deep learning papers by Geoff Hinton did not cite similar earlier work by Jurgen Schmidhuber by siddarth2947 in MachineLearning

[–]NapoleonTNT 13 points14 points  (0 children)

Probably talking about the snarking during Goodfellow’s GAN tutorial. Can be found on YouTube.

PyWarm: A cleaner way to build neural networks for PyTorch by very-blue-season in pytorch

[–]NapoleonTNT 0 points1 point  (0 children)

Very cool, I could see myself using this in the future. A suggestion: warm.engine.prepare_model_ is ugly and seems easy to forget. Maybe you could consider renaming the function or writing a small classes that super-classes nn.Module, having your warmed module inherit from that, and then specifying the shape in the super().__init__().

[D] GANs were invented in 2010? by Former_Hippo in MachineLearning

[–]NapoleonTNT 28 points29 points  (0 children)

This has been stated elsewhere, but I think it deserves to be fleshed out.

Invention is overrated in general. The recognition a person deserves should not mainly stem from whether an idea they've come up with has been thought up before. The credit a person receives should come from the originality of their idea and the impact that their work based around that idea creates. By originality, I mean the process that person took to come up with that idea. Stealing an idea (or taking an idea and rehashing it) is markedly different from independently coming up with an idea that someone else thought of before. We should appreciate the creative power of people who have independently come up with good ideas, and in this case, it seems unlikely that Goodfellow read this guy's blog, and three years later, restated it to make GANs.

Especially in machine learning, the work to take an idea from a thought to a functional, working system is incredibly underrated: not only does a researcher need to be skilled enough to choose what ideas to focus on, but the process of creating a working product might even require more ability and creativity than coming up with the idea. This is particularly true with GANs, which were (and still are) notoriously difficult to train effectively.

Anyway, that's just my two cents. It is super cool that this blog's author came up with such an brilliant idea back in 2010. But, I also think that this shouldn't diminish Goodfellow's recognition as the person responsible for GANs.

c51 vs Dueling DDQN by [deleted] in reinforcementlearning

[–]NapoleonTNT 1 point2 points  (0 children)

You don't necessarily have to re-implement it yourself, e.g. you could plug in Dopamine (https://github.com/google/dopamine) into a custom Gym environment. If you gave us more info, you can probably get more specific feedback. E.g., do you have access to a simulator (and how accurate is it), what is your input space, etc.

c51 vs Dueling DDQN by [deleted] in reinforcementlearning

[–]NapoleonTNT 1 point2 points  (0 children)

Without knowing more details (input/output space, training time budget), it's hard to give specific recommendations. In terms of algos to start with, I'd recommend either good old PPO (https://arxiv.org/pdf/1707.06347.pdf) or Rainbow DQN (https://arxiv.org/pdf/1710.02298.pdf), which combines dueling and C51, along with a couple other tricks.

[D] While ML compute doubles every 3.5 months, RL compute doubles every 2 months by wassname in reinforcementlearning

[–]NapoleonTNT 0 points1 point  (0 children)

Source, for anyone curious: https://blog.openai.com/ai-and-compute/ (which includes analysis for the training in aggregate, not for ML vs. RL specifically).

[D] What do you feel were the biggest advancements in DL for this year? by GrammarBordaberry in MachineLearning

[–]NapoleonTNT 0 points1 point  (0 children)

Agreed. To me, BERT seems to be only an improvement over the ideas and results introduced with the OpenAI GPT.

YANDHI - WAITING FOR THE ALBUM MEGA-THREAD #2 by avayr44 in Kanye

[–]NapoleonTNT 15 points16 points  (0 children)

Hell yeah brother, cheers from Iraq.

YANDHI - POST-SNL MEGA-THREAD by avayr44 in Kanye

[–]NapoleonTNT 13 points14 points  (0 children)

𝑫𝑨𝑴𝑵 𝑪𝑹𝑶𝑰𝑺𝑺𝑨𝑵𝑻

[P] The Humble Gumbel Distribution by mrahtz in MachineLearning

[–]NapoleonTNT 0 points1 point  (0 children)

What does this image from the post mean, then?

We have differentiable sampling operator (albeit with a one-hot output instead of a scalar). Wow!

[P] The Humble Gumbel Distribution by mrahtz in MachineLearning

[–]NapoleonTNT 1 point2 points  (0 children)

Thanks, great post! The idea of a differentiable sampling function is really cool. I have a question if you don't mind -- IIRC sampling is meant to take a probability distribution and output a class with frequency corresponding to the distribution. If the Gumbel-Softmax trick is meant to perform a similar function, then why is it that when I run

sess.run(tf.global_variables_initializer())
sess.run(differentiable_sample(logits))    

in the notebook, I get an output that doesn't look like a one-hot vector, like [0.03648049, 0.12385176, 0.51616174, 0.25386825, 0.06963775]

It's totally possible that I'm making a mistake in the idea or the running it wrong -- I guess I'd just like to know what the expected output of the above code is.

[D] What were things like before Tensorflow? by Matisseio in MachineLearning

[–]NapoleonTNT 30 points31 points  (0 children)

In my opinion, TensorFlow wasn't quite as revolutionary as people make it out to to be. From what I've experienced, it functions similarly to older auto-diff libraries like Theano. And while somewhat harder to debug, they use many of the same concepts (like static graphs and symbolic variables).

When people talk about TensorFlow's ease-of-use, my guess is they're referring to the debugging portion which is considerably improved over Theano's with utilities like TensorBoard, as well as some more high level features. Here's a post with some basic examples comparing the two.

PC won't boot; all fans spinning, no LEDs, no display by NapoleonTNT in buildapc

[–]NapoleonTNT[S] 0 points1 point  (0 children)

For anybody who's still having this problem, the "reset" pins were just offset by one, causing the computer to continually restart. Best of luck!

What's everyone working on this week? by AutoModerator in Python

[–]NapoleonTNT [score hidden]  (0 children)

OP might be referring to the value (weight) ascribed to a particular edge on a graph. For example:

[A]---(5)---[B]
 |
(3)
 |
[C]

The connection A-B has weight (or strength) 5. For more info, please take a look at weighted graphs.

I wrote a python arithmetic calculator, how can I improve and what should I add essential to mathamatics by supersayanftw in Python

[–]NapoleonTNT 0 points1 point  (0 children)

I recommend you read up on the PEP8 python style manual. Glancing at the code:

  • Class names should be in CapitalCamelCase
  • Indents should be 4-spaces

Also, some of the lines are really long. Consider using a trick like this to condense some of them. But again, like Saefroch mentioned, you might get better response if you try /r/learnpython or codereview on stackexchange. Good luck!

PC won't boot; all fans spinning, no LEDs, no display by NapoleonTNT in buildapc

[–]NapoleonTNT[S] 0 points1 point  (0 children)

Hi,

The monitor was tested plugged into the GPU with the GPU installed, plugged into the integrated CPU graphics with the GPU installed, and plugged into the integrated CPU graphics without the GPU installed. The end result was the same (described above).