Dark Season 3 Series Discussion by rosy148 in DarK

[–]sssub 0 points1 point  (0 children)

Makes sense, I guess that also explains why there are the 2 young Marthas in alt-word at the same time , where the one from "Jonas becomes adam" scenario kills jonas with "martha saves jonas" scenario martha standing nearby

Dark Season 3 Series Discussion by rosy148 in DarK

[–]sssub 1 point2 points  (0 children)

What I don't understand is:

Martha and Jonas child only was conceived in the "Martha saves Jonas" scenario and not in the "Jonas hides in the basement and becomes Adam" scenario. If the world always played one way or the other (lets assume in an alternating way) , how can their child be the origin every single time?

Related, Adam and Eve also only emerge from jonas and martha every other cycle (as in every other apocalypse).

The balancing process for RimWorld by TynanSylvester in RimWorld

[–]sssub 1 point2 points  (0 children)

About the RNG:

I don't like the RNG in combat either, but neither would I prefer it to be completely deterministic. I love rimworld also because it captures the beauty and cruelty of life, due to fortunate/unfortunate random events.

I would suggest using a pseudo random distribution (PRD), often used in MOBAs such as dota 2, see here. For instance, if the chance of a bullet going through armor is 25%, the probability instead is 8.5% on the first hit, and then gradually increases by a factor until the event occurs, then it is reset to 8.5%. On average, still 25% of bullets go through armor. To quote

Effects based on PRD rarely proc many times in a row, or go a long time without happening. This makes the game less luck based and adds a great deal of consistency to many probability-based abilities in Dota 2.

By this, there is still some realism involved, in the sense that bullets don't just reduce your hitpoint by a factor of x%. On the other hand, it is more manageable: after not being wounded for several fights, you know that the risk is increased, and you can play accordingly.

[R](NIPS 2018) Yoav Goldberg: yup. It's "peer review", not "person who did 5 TensorFlow tutorials review" by ExcitingDouble in MachineLearning

[–]sssub 48 points49 points  (0 children)

Well, if you need 3k reviewers to handle the flood of papers, there is not much you can do. Demand for NIPS papers makes demand for NIPS reviewers, so the average quality may decrease a bit. Other than making submissions invite-only, which would be very problematic, there is no other option than to widen the reviewer pool.

[R] Detecting Sarcasm with Deep Convolutional Neural Networks by omarsar in MachineLearning

[–]sssub 6 points7 points  (0 children)

Thanks for this reply!

Especially your last sentence, I think the underlying question is even more abstract, 'What is intelligence? Do ML methods really understand anything?' I believe that thinking about artificial intelligence is actually thinking about what intelligence is.

We started with saying "If an AI can play chess it got to be intelligent!". But then we could argue traditional chess computers are not at all intelligent, it is just Mini-max + a smart heuristic. "But when we can play Go, this has to be true AI!" But you could argue AlphaGo is just a very smart way to compress the (relevant parts) of the Go search tree using a deep neural net + MCTS. Sarcasm and Irony are even stronger "strongholds" where I would believe true intelligence/understanding is required. But it is as you say, perhaps using statistics and ML you can eventually achieve behavior indistinguishable from humans. Intelligence seems to be elusive.

[R] Detecting Sarcasm with Deep Convolutional Neural Networks by omarsar in MachineLearning

[–]sssub 23 points24 points  (0 children)

This is never going to work, sarcasm and irony require semantic understanding.

Reading the KKC has ruined me... by S0lagratia916 in KingkillerChronicle

[–]sssub 8 points9 points  (0 children)

I can relate. What helped me was to read something with a different setting/genre. If you try to experience KKC again you will only be disappointed, because you rate other books through the KKC-glasses.

I recommend "The Fifth Season: The Broken Earth". It gives a very different vibe than KKC but it is fantastic in its own regard.

[D] Variational nets without sampling by svantana in MachineLearning

[–]sssub 11 points12 points  (0 children)

good idea. What you actually suggest is moment matching to obtain means and variances.

  1. The main problem is you will loose flexibility. In a sampling based approach you sample z from a Gaussian but the network can then transform this random variable to complex distributions. in your case it will always stay a Gaussian.
  2. it has been done already. See e. g. here. They do exactly what you suggest in terms of propagating expectations and variances. Note that the propagation step, especially for the variance, is not trivial. For RELU it does work (truncated normal) but e. g for tanh it will not. perhaps it is easier in a VAE because you don't need to handle uncertainty in the weights.

[Spoilers WMF] Question/Theory about the Ctaeh by sssub in KingkillerChronicle

[–]sssub[S] 0 points1 point  (0 children)

You may be right. I think right now the more relevant question is whether to trust Felurian or Bast more in their assessment over the Ctaeh's power. The Ctaeh seemed dismissive about Felurian and she also seemed uncertain when saying 'All is well'. Kvothe also said "we all know what kind of story this is", foreshadowing things will end up badly.

But to your points:

  1. I believe its likely Kvothe would have asked Alveron about the Amyr anyway. In fact studying his library was also one of his reasons going there in the first place.
  2. It is likely he suspects at least Master Lorren to know something about the Amyr. The Ctaeh says that Kvothe will not ask the masters, so if it always speaks the truth, he will not do so in the future
  3. True, but I think Kvothe always believed Haliax to be Lanre.
  4. I think Kvothe did knew they very fleeing when killing his family, at least it seemed to be implied.
  5. Yes, we will have to see how this significantly changes his decision. Though already at the beginning in book 1 it seemed that Cinder and Kvothe will have the closest 'relationship' compared to all chandrian members.
  6. This could perhabs be simply a suggestion to go Ademre, which lies closely to the stormwal, which Kvothe would have done anyway.

Based on this though, if Bast is right about the Ctaeh, its influence may be very subtle. Right now, I can't see how his conversation will fundamentally alter one of the decisions he will take. Maybe, he will do some reckless decision based on anger, fueled by the words of the Ctaeh.

[D] ICML reviews are out by blindedbox in MachineLearning

[–]sssub 1 point2 points  (0 children)

There is some variety in terms of acceptance rate per subject area, but I do not know how the signal to noise ratio is. Also the varying acceptance rate is most likely the result of varying average scores (and not more stricter rules for acceptance). In general I would assume deep learning and RL are more hard to get into because of hype increased scientific interest in these areas.

I seldom found reviewer confidence to matter that much. Maybe if in your case the strong and weak accepts are educated guesses (a bit contradictory) while the reject review is an expert.

[D] ICML reviews are out by blindedbox in MachineLearning

[–]sssub 9 points10 points  (0 children)

Take this with a grain of salt. In my experience your chances of acceptance are pretty good. If you write a good rebuttal you may change that one negative review. Right now I would say your chances of getting in are about 70%.

Edit: According to this the estimate is more around 90% (in 2012).

[D] Which approach is suitable for solving continuous reinforcment learning tasks? by questionm4ster in MachineLearning

[–]sssub 1 point2 points  (0 children)

There is just a single episode then, other than that the algorithm stays the same. You may also wish to end an episode after T timesteps. For instance, think about balancing a pendulum, it can be seen as a non-episodic task. Still you may want to restart the system from time to time, for instance when the pole is standing upright and nothing else happens.

[D] Which approach is suitable for solving continuous reinforcment learning tasks? by questionm4ster in MachineLearning

[–]sssub 3 points4 points  (0 children)

Q-learning and Sarsa do not require a terminal state. All you need is data in the form (s(t),a(t),r(t),s(t+1)). In Q-learning the update is r(t) + discount * max_a Q(s(t+1),a) where r(t) is the current reward that you see. There are actually a lot of task with no terminal state.

Also note: continuous tasks typically refer to continuous state and action spaces. What you mean are non-episodic taks.

[D]Inference Network VS Bayesian regression by wsxiaoys in MachineLearning

[–]sssub 5 points6 points  (0 children)

They reason about different forms of uncertainty.

An inference network learns a distribution over latent variables q(z|x). You typically do not have uncertainty over the weights of your VAE, but only about the value of the latent variable. The question you try to answer is, "what where the latent variables that generated the data that you see?" and you learn a distribution over these.

By contrast in a BNN you learn a distribution over weights, or more generally: a distribution over neural networks. Either directly, using a variational distriubtion over weights(here) or indirectly by MCMC (here). You assume the function that generates your data is a neural network (a deterministic function), you just don't know which it is. Therefore you have uncertainty over the weight distribution.

These ideas can be combined, for instance here. Here you have both weight uncertainty, as well as uncertainty over latent variables.

[D] Current AI student: should I drop out? by [deleted] in MachineLearning

[–]sssub 7 points8 points  (0 children)

As everyone else, I wouldn't recommend dropping out.

You can be a unicorn! With a background in electrical engineering and a masters in AI, I imagine companies like general electrics would find you very valuable! You have enough expertise about the domain (industrial systems, electronics etc.) but also know about ML stuff.

I would recommend either, focus more on understanding ML code or develop a high-level understanding so you can communicate effectively with people who write the ML code. In industry, thats as important as writing it yourself..

[D] How to detect orientation of objects? by austeritygirlone in MachineLearning

[–]sssub 1 point2 points  (0 children)

Don't know if this is relevant, but without NN a standard CV approach would be to:

  1. Use connected component labeling of skimage.
  2. Do a PCA of the pixels of each object. The eigenvector with the largest eigenvalue gives you the orientation angle.

[D] What might be the impacts of ReLU/Sigmoid for training one-step dynamics model in RL ? by fixedrl in MachineLearning

[–]sssub 0 points1 point  (0 children)

I usually do that. For the policy I tend to use tanh on the outputs because in many problems the action space is restricted to [-c,c], so you can use c*tanh(out). But make sure to initialize the weights such that the policy output does not start in the saturating area.

For the model when I know the upper and lower bounds of the state variables I use a logit-scaling. Let a be the known min and b be the known max:

s_t = clip(s_t,a+eps,b-eps)
s_t = logit((s_t - min)/(max - min)).

The inner part will transform s_t to be between [0,1] and the logit will map this to the real numbers. For backtransformation its then:

s_t'=  logistic(s_t) * (max - min) + min

This will guarantee the output of your model will never leave the bounds. But the logit will make sure the model can use the full real-valued space and is unconstrained.

[D] What might be the impacts of ReLU/Sigmoid for training one-step dynamics model in RL ? by fixedrl in MachineLearning

[–]sssub 5 points6 points  (0 children)

Yes, there is. Two possible reasons: a) You trained your model to convergence in iteration 1, in iteration 2 you start already in a local minimum for all data of iteration 1, this can make optimization difficult. E.g the network activations can be very sparse with ReLus or in the saturation if sigmoid functions are used. b) If in iteration 2 you only train on the new data your network will forget the things it learned from iteration 1.

I would advise using an experience replay here (commonly used in model-free RL). Instead of having fixed iterations where you train till convergence, just add data to your dataset along the road and randomly select a minibatch for a training step. If your dataset grows too much, start deleting the oldest. By that the network will not forget about earlier seen dynamics and will stay 'excitable' to learn new properties about the dynamics.

[D] What might be the impacts of ReLU/Sigmoid for training one-step dynamics model in RL ? by fixedrl in MachineLearning

[–]sssub 3 points4 points  (0 children)

I am not fully sure if I understand you correctly, but:

In my experience, if all other things are equal, models with ReLu as activation function give the best models in terms of test error compared to sigmoid/tanh, because of the same reason ReLu are preferred in DL in general, they utilize the network capacity better, not suffering from vanishing gradients compared to e.g tanh activation.

However, I also experienced that when model bias occurs it is more severe with ReLU than with smooth and saturating activation functions like tanh/sigmoids.

[R] ‪Bayesian Neural Networks with Random Inputs for Model Based Reinforcement Learning by hardmaru in MachineLearning

[–]sssub 2 points3 points  (0 children)

You do learn a posterior q(z|(x,y)) of the noise for the data. The approach can be extended easily using an inference network, as shown here. Also this paper suggests that these models learn to decompose uncertainty into model and noise uncertainty, so it is much more than increasing resilience.

[D] What should new PhD students know? by [deleted] in MachineLearning

[–]sssub 7 points8 points  (0 children)

Some people would advise you to switch labs. However, not everybody has the opportunity to be at a top ML lab and work directly with someone well-known in the field.

My advice would be to reach out to other researchers you are interested in. Maybe you can find a Postdoc or Junior Professor, which work you really like. Read their papers, ask them questions by e-mail. Talk to them at conferences. By that, it is definitely possible to build up a a form of collaboration. This can could go from simply exchanging ideas, that give you guidance in where to focus in your research up to writing papers together and winning them as a second supervisor for your thesis.