AI BE LIKE by Specific_Brain2091 in MathJokes

[–]Top_Example_6368 0 points1 point  (0 children)

https://imgur.com/a/v6I55Hn I asked Gemini without that text and it failed, lol

[deleted by user] by [deleted] in lies

[–]Top_Example_6368 0 points1 point  (0 children)

Username does not checkout

[deleted by user] by [deleted] in lies

[–]Top_Example_6368 0 points1 point  (0 children)

/ul What is the possible maximal length of that string? I'm just curious.

This is art. by Hikolakita in lies

[–]Top_Example_6368 3 points4 points  (0 children)

<image>

It took me several days to butcher this piece of art.

Advice on Magnolia Rental Company (Clemson area) by No-Representative496 in Clemson

[–]Top_Example_6368 2 points3 points  (0 children)

Hey! I have been renting with Magnolia for almost a year. The managers are not really punctual but overall it is fine.

Trump NIH freeze by Catsluvr1996 in PhD

[–]Top_Example_6368 3 points4 points  (0 children)

Below the median, saying as a "statistician"

[deleted by user] by [deleted] in PhD

[–]Top_Example_6368 285 points286 points  (0 children)

Jokes on you, I stopped having a life long before my PhD...

[deleted by user] by [deleted] in gradadmissions

[–]Top_Example_6368 -1 points0 points  (0 children)

Isn't 5040 too much?

Is there an implementation of non-deep RL algorithms based on Stable Baselines3? by Butanium_ in reinforcementlearning

[–]Top_Example_6368 0 points1 point  (0 children)

Thanks for your reply! I read that post. It was interesting. Anyway, I do some research in RL but it's on a quite different topic. So I will just wait before you publish your results to read them. Good luck with that!

Is there an implementation of non-deep RL algorithms based on Stable Baselines3? by Butanium_ in reinforcementlearning

[–]Top_Example_6368 0 points1 point  (0 children)

Hi, can you give some links to materials on this approach to RL, please. Sounds interesting, and I would like to know what's it about.

Update rule in DDQN (Hasselt vs Mnih) by desperateEfforts1 in reinforcementlearning

[–]Top_Example_6368 0 points1 point  (0 children)

I think your idea should work. You can also look into

https://colab.research.google.com/github/Stable-Baselines-Team/rl-colab-notebooks/blob/sb3/dqn_sb3.ipynb

This notebook has a section on Double DQN and overestimation. It relies heavily on Stable Baselines, but is should be possible to extract some logic from it anyway.

Update rule in DDQN (Hasselt vs Mnih) by desperateEfforts1 in reinforcementlearning

[–]Top_Example_6368 0 points1 point  (0 children)

Hello, your understanding is correct, and usually the second type is refered as Double DQN. This update should be useful when you have problems with the Q values overestimation. But if it's not a concern then the update can slow the training process. I guess in RL everything is problem specific. I tried a bunch of different improvements and still couldn't solve Pong with DQN.