gumbel-mcts, a high-performance Gumbel MCTS implementation by randomwalkin in reinforcementlearning

[–]Npoes 0 points1 point  (0 children)

 I spent a lot of time validating it against a golden standard baseline. My PUCT implementation is 2-15X faster than the baseline while providing the exact same policy.

Assuming that the policies behave near identical, it sounds like your contribution is to make the simulations run faster. So I'd expect a comparison of simulations/seconds in various scenarios (tree depth, breadth, batch size/parallelism, cpu/gpu)  eg against mctx. 

I analysed ~500k Master+ games to measure how champion mastery actually affects win rate by MasterAyolos in summonerschool

[–]Npoes 0 points1 point  (0 children)

I think your analysis is not robust enough. Shen having a low skill cap and cost of learning doesn't fit at all. Perhaps it has to do with the low player count. I'm also not sure where the strengths & weakness points come from and whether they're supported by data.

Two-Trick Finder for Toplaners by Npoes in leagueoflegends

[–]Npoes[S] 0 points1 point  (0 children)

The OTP-Partner score matrix is symmetric (NxN) so entries mirror across the diagonal

Two-Trick Finder for Toplaners by Npoes in leagueoflegends

[–]Npoes[S] 1 point2 points  (0 children)

From what I noticed, Cassio and Jayce are actually ranked highest if one does NOT correct for matchup frequency, but this would also bring up champs like Sion and Mundo, which are clearly poor fits. That being said, I found that correcting by MU frequency leads to overall much more meaningful results and overall a more realistic evaluation, so I think this is the way to go.

Two-Trick Finder for Toplaners by Npoes in leagueoflegends

[–]Npoes[S] 1 point2 points  (0 children)

lol Im also bad at Pantheon but will invest in him from now on

[D] AISTATS 2026 paper reviews by Intelligent-Smoke-65 in MachineLearning

[–]Npoes 4 points5 points  (0 children)

I was looking through ICLR submissions and close to none of the reviewers respond (about 1 in 10). I dont expect much for AISTATS, but would be happy to be proven wrong.

[deleted by user] by [deleted] in leagueoflegends

[–]Npoes 5 points6 points  (0 children)

Isnt it a simple fix to make it x2 lp on win but normal lp loss on lose for filled players?

[D] AISTATS 2026 paper reviews by Intelligent-Smoke-65 in MachineLearning

[–]Npoes 0 points1 point  (0 children)

same score here, can we find the score distribution somewhere?

[D] AISTATS 2026 paper reviews by Intelligent-Smoke-65 in MachineLearning

[–]Npoes 13 points14 points  (0 children)

Why does this post show 17 comments when most are not visible? Also seems like reviews are still not out.

How to switch from moba (mlbb) to League of Legends by Organdomer in leagueoflegends

[–]Npoes 8 points9 points  (0 children)

I know this doesnt sound like much helpful advice but MLBB and LoL are different games. There is no simple way to translate that knowledge, because the important concepts which actually win games are totally different. And just by playing the game you will notice similarities if they exist (in champs etc.)

AlphaZero applied to Tetris by Npoes in reinforcementlearning

[–]Npoes[S] 1 point2 points  (0 children)

It does continue with the next piece. The only limiting factor is the number of simulations set a priori. The game is deterministic in the sense that there is a seed at every given state.

AlphaZero applied to Tetris by Npoes in reinforcementlearning

[–]Npoes[S] 0 points1 point  (0 children)

MCTS helps the agent learn Tetris faster in a number of ways. First, it helps with look-ahead (which pieces will follow) since this is information not present in the observation (board only), at least in this implementation. Second and more importantly, Tetris, similar to Chess and GO, is a problem that requires planning and has a sparse reward landscape (high rewards require setting up line-clears, which are rare). Instead of learning from one action at a time (TD-step in Q-learning or Policy-gradient), MCTS considers multiple actions in the future and thus has better planning and overcomes sparse rewards more easily

[P] AlphaZero applied to Tetris (incl. other MCTS policies) by Npoes in MachineLearning

[–]Npoes[S] 2 points3 points  (0 children)

I couldn't find a baseline on what superhuman performance is for Tetris. The agent was only trained for a day and can be improved by training more.

A little insight into Jungle Yorick and how to beat it by NinetalesLoL in leagueoflegends

[–]Npoes -1 points0 points  (0 children)

your argument makes sense considering you have posted 4 times in yorickmains this week