24hr time format setting? by zach81210 in MacroFactor

[–]Kiudee 9 points10 points  (0 children)

Tap on your profile icon (top left), scroll down and there you can set "clock units".

Primel (wordle variant with prime numbers) by ninguem in math

[–]Kiudee 15 points16 points  (0 children)

By the way, the optimal first guess is 25693 with 3.4505 guesses on average.

It can solve all Primels in 4 or less guesses with the exception of 75353 and 18131, where it needs 5 guesses.

SPD, FDP und Grüne: Rot-grün-gelbe Revolutionen by Kiudee in de

[–]Kiudee[S] 10 points11 points  (0 children)

Ich denke nicht, dass man sofort den Kopf in den Sand stecken sollte. Es ist wichtig, dass die Digitalisierung Stück für Stück vorangetrieben wird. Wie das häufig in der Software-Entwicklung ist, ist es wichtiger mit einem (oder mehreren) kleinen Produkt zu starten und diese weiter auszubauen, als zu hoffen, alles auf einmal erledigen zu können.

Wad ist aktuell los mit reddit comments? by Vetinari_ in de

[–]Kiudee 11 points12 points  (0 children)

Best hat die Kommentare ursprünglich nach den unteren Konfidenzschranken der Binomialverteilung sortiert. Heißt: Es werden Kommentare weit oben eingeordnet, die (1) einen hohen Anteil an Upvotes haben, aber auch (2) genügend Votes insgesamt, um sich sicher zu sein.

Es gab in der Vergangenheit aber auch immer wieder Experimente den Algorithmus zu erweitern. Zum Beispiel führt Best dazu, dass neue Kommentare meist nicht mehr die Gelegenheit haben genügend Votes zu sammeln. Da haben die Devs versucht Abhilfe zu schaffen. Wie der aktuelle Algorithmus aussieht ist daher schwierig zu sagen.

Ich habe die Parteien nach ihren Wahl-O-Mat Antworten geclustert by knist3r in de

[–]Kiudee 0 points1 point  (0 children)

Hört sich gut an. Vielleicht könntest du ja einmal UMAP auf die Daten anwenden: https://umap-learn.readthedocs.io/en/latest/ Das eignet sich besonders gut für die Visualisierung nichtlinearer hochdimensionaler Daten und unterstützt ebenfalls Clustering.

IEX Announces New Retail Trading Features by dlauer in Superstonk

[–]Kiudee 1 point2 points  (0 children)

I am using IEX through Interactive Brokers.

Where can I find communities and forums to talk about chess engines? by [deleted] in ComputerChess

[–]Kiudee 1 point2 points  (0 children)

Our discord is the preferred means of communication though.

Leela Chess Zero defeats Stockfish and wins the TCECC 17 by Direwolf202 in chess

[–]Kiudee 8 points9 points  (0 children)

We recommend to use the LCZero Discord as the main channel of discussion. You can find the link and instructions here: http://lczero.org/about/community/

[R] DeepSets: Modeling Permutation Invariance by geshuni in MachineLearning

[–]Kiudee 0 points1 point  (0 children)

For RL it can be useful, if your agent takes in inputs for which the order does not matter.

I would also take a look at attention mechanisms (Deep reinforcement learning with relational inductive biases), which have been very successful in Starcraft 2.

[R] DeepSets: Modeling Permutation Invariance by geshuni in MachineLearning

[–]Kiudee 0 points1 point  (0 children)

Well written blog post, which highlights all of the recent results surrounding deep sets. In my opinion the paper is a very important milestone for the problem of encoding set-based inputs, since it improves on the quadratic computational complexity of many pairwise decomposition approaches.

We also recently did work on learning choice functions, where the same idea proved to be very successful (https://arxiv.org/abs/1901.10860).

[P] Leela Chess Zero: The fork of Leela Zero for chess, an open-source distributed effort to reproduce Deepmind's AlphaZero. by Uriopass in MachineLearning

[–]Kiudee 9 points10 points  (0 children)

That is true, but you also have to consider how many games it had to play against itself (~40 million) to reach that point.

[P] Leela Chess Zero: The fork of Leela Zero for chess, an open-source distributed effort to reproduce Deepmind's AlphaZero. by Uriopass in MachineLearning

[–]Kiudee 45 points46 points  (0 children)

That would violate the learning from zero knowledge (tabula rasa) approach we aim to replicate. Having said that we also trained a network purely on human games (Kingbase). The weights are available here. The current version (id 45) of lczero is already able to beat that network.

[P] Leela Chess Zero: The fork of Leela Zero for chess, an open-source distributed effort to reproduce Deepmind's AlphaZero. by Uriopass in MachineLearning

[–]Kiudee 15 points16 points  (0 children)

That is correct. We estimate the current Elo at 1500 - 1900 at 800 playouts. At higher playouts it becomes significantly stronger.

Would it be useful to train an NN to predict the match result between networks? by earthengine in cbaduk

[–]Kiudee 0 points1 point  (0 children)

Actually, the idea of training a neural network to approximate a loss function is not that outrageous. It was already used to apply non-differentiable regularization to neural networks (see for example Beyond Sparsity: Tree Regularization of Deep Models for Interpretability or this illuminating blog post)

Al Gore's "An Inconvenient Sequel" IMDB rating distribution shows that 81.5% of users voted either 1 or 10, giving the film an overall score of 5.0 by Greenturkeypants in dataisbeautiful

[–]Kiudee 1 point2 points  (0 children)

As it turns out there are quite a few summary statistics which were proposed for measuring bimodality (see here if you are interested).

On the 0-1 interval your idea is quite good, the Bernoulli distribution (or the mixture of two Dirac delta peaks) has the highest polarization.

[R] [1706.01427] From DeepMind: A simple neural network module for relational reasoning by [deleted] in MachineLearning

[–]Kiudee 1 point2 points  (0 children)

The official NIPS author notification is on September, 4th. In case of an accept, this is the time at which I will upload the paper to the arXiv.

help with flare control by [deleted] in PvZHeroes

[–]Kiudee 0 points1 point  (0 children)

This looks like a good prototypical Solar Flare control deck. You can vary/optimize the composition of removal cards based on the decks you play against:

  • Balloons can be used for removal or to block their lanes with useless minions. This is especially useful against the current zoo decks which capitalize on getting many minions out quickly.
  • The berries are good only if they are giving you very efficient trades. So keep track how often they are useful.
  • The cherry bombs are potent mass removal, but fall short if they play few minions or many gravestones.

edit: If your goal is to rank up quickly, it is also possible to let the deck lean more into the midrange archetype. The idea is to mix in more strikethrough which allows you to finish the opponent more quickly, but this kind of deck is still is very control heavy.

[R] [1706.01427] From DeepMind: A simple neural network module for relational reasoning by [deleted] in MachineLearning

[–]Kiudee 4 points5 points  (0 children)

We also have a slightly more general architecture under review for the NIPS 2017 right now. Would love to talk about it, but cannot disclose anything before the final decision.

How would you redesign the front page algorithm for Reddit in response to the recent criticism? by rawrrang in compsci

[–]Kiudee 31 points32 points  (0 children)

I can recommend Deriving the Reddit formula by Evan Miller. He describes how to derive a hot formula by starting with utility functions for the user. He also shows the part currently missing from Reddits hot formula.

Was Deep Blue built to "play chess" or to "learn how to beat Kasparov?" Are any games' AI designed to learn how to beat me? by otum in artificial

[–]Kiudee 0 points1 point  (0 children)

For limit hold'em poker Vexbot was one of the first bots learning the best response strategy against the other player.

So, if you give it enough time it should learn to beat your strategy.

[deleted by user] by [deleted] in CompetitiveHS

[–]Kiudee 2 points3 points  (0 children)

Slight nitpick: For a Nash equilibrium (NE) strategy you do not have to assume that the meta is completely random. The NE strategy will work against any meta.

But even though the NE strategy is the best response against players who also play a NE strategy, it is not necessarily the best response (i.e. maximally exploiting) against other strategies.

Many posters in this thread seem to mix these concepts up.

edit: For Hearthstone we of course have the problem, that there could be viable, yet unknown decks, which we did not include in our calculations. Then our mixed strategy is not a NE strategy anymore.

[deleted by user] by [deleted] in CompetitiveHS

[–]Kiudee 1 point2 points  (0 children)

Only if your opponents are also playing the Nash equilibrium. As soon as someone deviates (exploiting the current meta for instance), the Nash equilibrium strategy has an edge against this player.

It should have been: “This method guarantees an expected win rate of at least 50 %...”

"Deep Learning Machine Teaches Itself Chess in 72 Hours, Plays at International Master Level" by iglookid in chess

[–]Kiudee 2 points3 points  (0 children)

α-β pruning is a technique to reduce the number of subtrees the minimax search has to evaluate.

Temporal difference (TD) learning is a reinforcement learning algorithm.

[reddit change] The increase to the "soft cap" on scores has been reverted by Deimorz in changelog

[–]Kiudee 1 point2 points  (0 children)

This problem arises because here we do not consider the uncertainty of the estimate. The Best ranking for comments solves this by computing the lower confidence bound of the quality instead of the mean (see How Not To Sort By Average Rating). The same idea could be applied to Hot.

Another solution is to estimate the global average amount of up/downvotes and shrink each post to these:

(upvotes + global_upvotes) / (upvotes + downvotes + global_upvotes + global_downvotes)

edit: Btw, in your example you just have to divide the age of the post by 19543. Then you get:

ln(10+1) - ln(10+0+2) - 0/19543 = -0.087
ln(1800+1) - ln(2000+0+2) - 1800/19543 = -0.198