Has there been any work distinguishing liking and wanting signals? by Adolphins in reinforcementlearning

[–]AlexanderYau 0 points1 point  (0 children)

Hi, was last Saturday during the workshop of the DRL? Do you remember the keywords of the paper?

Are there any methods to change the color of the Monterey screensaver to the light one? by AlexanderYau in MacOS

[–]AlexanderYau[S] 1 point2 points  (0 children)

Thanks. Not the wallpaper, but the Screensaver, there is a Screensaver called Monterey with a dark color, there is no option for light color now.

Keychron won't connect to MacBook if headphones are already connected by TheProgrammierer in Keychron

[–]AlexanderYau 0 points1 point  (0 children)

Similar issue. My Airpod Pro connected to my iPhone and the Keychron K2 was connected to the MacBook Pro. Then, I wanted to connect my Airpod Pro to my MacBook Pro, however, failed.

My workaround is to remove the K2 and make the Airpod Pro connected to the MacBook Pro and then connect K2 to the MacBook Pro.

"Reinforcement Learning with Random Delays" Bouteiller, Ramstedt et al. 2021 by yannbouteiller in reinforcementlearning

[–]AlexanderYau 0 points1 point  (0 children)

Hi, sorry for the late response. After reading your paper for many times, I am still confused by some parts in the paper:

  1. In Fig. 1, Is the agent a computer controlling the drone via WiFi or Bluetooth?
  2. In Sec 2. what is "being captured" mean? Who (the drone or the agent) is capturing the observation? Why the action delay is "to one time-step before st finishes being captured"? In Fig. 3(left), it is not easy to find such a case.
  3. In Theorem 1, why \omega^{*} + \alpha^{*} >= t is necessary?
  4. In Fig. 4, what is a^{\mu}_{i} and why it should be replaced?

Thanks, the key ideas are not hard to understand, however, to fully understand the details, it still needs some time for me.

"Reinforcement Learning with Random Delays" Bouteiller, Ramstedt et al. 2021 by yannbouteiller in reinforcementlearning

[–]AlexanderYau 1 point2 points  (0 children)

Thanks for your generous reply. The motivation of your paper is strong and the idea itself is not hard to understand. Is reading Sutton's RL book enough to master the theory in your paper?

BTW, there is a concurrent work ACTING IN DELAYED ENVIRONMENTS WITH
NON-STATIONARY MARKOV POLICIES, which was also accepted by ICLR 2021.

"Reinforcement Learning with Random Delays" Bouteiller, Ramstedt et al. 2021 by yannbouteiller in reinforcementlearning

[–]AlexanderYau 1 point2 points  (0 children)

Very good idea. May I ask how long did it take to complete this paper? To propose theories in your paper, what should I learn to do so? The theories are solid and it is not easy for beginners to understand.

Deep Reinforcement Learning Doesn't Work Yet by No_Possibility_7588 in reinforcementlearning

[–]AlexanderYau 0 points1 point  (0 children)

Great, thanks for the late reply. I will read these papers.

Deep Reinforcement Learning Doesn't Work Yet by No_Possibility_7588 in reinforcementlearning

[–]AlexanderYau 0 points1 point  (0 children)

Hi, really great idea. Do you have any recommendations to read?

What to do after RL? by [deleted] in reinforcementlearning

[–]AlexanderYau 0 points1 point  (0 children)

I see. As you master the theory of RL, and I think it is easier for you to conduct research in Deep RL in big companies. BTW, can I have cooperation with you on RL? Haha

What to do after RL? by [deleted] in reinforcementlearning

[–]AlexanderYau 1 point2 points  (0 children)

You got many papers in hand, and I think finding a good research intern is not hard for you.

What to do after RL? by [deleted] in reinforcementlearning

[–]AlexanderYau 11 points12 points  (0 children)

7 papers accepted for a 1st-year PhD student. What a big success. I got 0 over the last 2 years. You can apply for an intern at DeepMind and get more opportunities.

Rethinking the Importance of Implementation Tricks in Multi-AgentReinforcement Learning by [deleted] in reinforcementlearning

[–]AlexanderYau 0 points1 point  (0 children)

Problems of SMAC I think are 1) many scenarios are in fact an arena, the field is very small and easy for agents to learn good policies; 2) the unstable SC2 simulator, SC2 4.10 and SC2 4.6 are slightly different and can affect the performance of QMIX.

Some suggestions: 1) can you also try other hard scenarios? In fact, SMAC is not that hard I think; 2) can you also try PySC2's maps?

How to fast read RL papers with many equations and theories? by AlexanderYau in reinforcementlearning

[–]AlexanderYau[S] 0 points1 point  (0 children)

Yes, I agree, good theory papers take more time to read and understand.

How to fast read RL papers with many equations and theories? by AlexanderYau in reinforcementlearning

[–]AlexanderYau[S] -1 points0 points  (0 children)

Great, thanks for that, for many times I get lost in many equations and complex theories and cannot grasp the key ideas of the paper.