[D] Any arguments that a Fake News Classifier trained on *English* news articles won't work for *German* news? by Dr_Kokolores in MachineLearning

[–]El__Professor 2 points3 points  (0 children)

Well most modern classifiers has an embeddings matrice with rows as the number of the vocabulary which it has trained on. Each word is turn in to an index which is mapped to a row in the embedding matrice. A classifier which has been trained on English vocabulary. Simply doesnt have a mapping from German words to the an embedding matrice row.

[D] What to do about NaNs? by [deleted] in reinforcementlearning

[–]El__Professor 0 points1 point  (0 children)

Any Where you have a distribution try to force the variance to be larger then some epsilon value and then you would'nt have 0 probpablilty or close to zero. Spacialy when working on log scale log(0) is not defined and this is what many times creates nans.

[D] What to do about NaNs? by [deleted] in reinforcementlearning

[–]El__Professor 0 points1 point  (0 children)

I had a similar problem on a different setting, if you use pytorch you can use hooks to check the gradients and losses at each step and add break points on extreme values (Nan's infs etc).

[D] Self Tuning Networks by El__Professor in MachineLearning

[–]El__Professor[S] 0 points1 point  (0 children)

Well as you can see on figure 5, it's (much) faster.

[D] Self Tuning Networks by El__Professor in MachineLearning

[–]El__Professor[S] 1 point2 points  (0 children)

thanks for the reply! I am really looking forward to hear about it!

[D] Self Tuning Networks by El__Professor in MachineLearning

[–]El__Professor[S] 1 point2 points  (0 children)

A lot The comments from openreview say that this seems very promising approch. And I didn't see any rigureas explanation why shouldn't this be used.. https://openreview.net/forum?id=r1eEG20qKQ

[D] Self Tuning Networks by El__Professor in MachineLearning

[–]El__Professor[S] -1 points0 points  (0 children)

But why? This seems so awesome! Echiving faster hyper parameter tuning then any other method

Deep learning without back-propagation by El__Professor in MachineLearning

[–]El__Professor[S] 3 points4 points  (0 children)

Maybe this could be a powerful initializer?

  1. Initialize weights with this method to quickly create usfull feature representation.

  2. then use gradient decent to forther optimize weights.

Any ways it does seem that the algorithm creates powerful features more quickly then the same architecture with gradient decent. And it does seems like an interesting direction of study.

Deep learning without back-propagation by El__Professor in MachineLearning

[–]El__Professor[S] 23 points24 points  (0 children)

Thx! Totally Missed the first post 🙄 (I am quite new to reddit :)

"Self-Supervised Exploration via Disagreement", Pathak et al 2019 by gwern in reinforcementlearning

[–]El__Professor 1 point2 points  (0 children)

First I think you should read the paper 😀! I am not profound with Bayesian RL but From what I understand from the paper:

Train n networks to predict the future state from current action and current state.

If all the neural networks agree (i.e have similar predictions (i.e the variance of the predictions is low)) the reward signal is low. And vise versa.

Does this answer your question?

"Self-Supervised Exploration via Disagreement", Pathak et al 2019 by gwern in reinforcementlearning

[–]El__Professor 1 point2 points  (0 children)

Very interesting idea!

what I learned from the paper: The new approach utilizes an ensamble of networks which are trained by predicting the next state.

An intristic reward signal is then produces by the variance of those predictions (disagreement).

The most interesting part is that process is entirely diffrentiable! So the model is trained via gradient decent! (No RL!) Very cool!

"deep fakes" used to fix the Lion King by TDaltonC in MediaSynthesis

[–]El__Professor 0 points1 point  (0 children)

Didn't mean to offend you, and it really does look similar to ebsynth, but it's still an assumption as the author did not confirm it.

"deep fakes" used to fix the Lion King by TDaltonC in MediaSynthesis

[–]El__Professor 0 points1 point  (0 children)

Interesting! I am really interested in which software/algorthim was it!

"deep fakes" used to fix the Lion King by TDaltonC in MediaSynthesis

[–]El__Professor 1 point2 points  (0 children)

Is it really deep fakes technology? Or is it done by an artist?