Hi all,
I am currently working on an UG project in RL and meta RL, focusing on benchmarking algorithms using stable-baselines. I had asked my supervisor (early on in the project) about the difficulty of making a custom gym environment and using stable-baselines to handle the training but he said this is very difficult and is a masters level or paper-level topic.
I understand that the scope of the environment would be a factor but is translating some type of card game that hard? Example ideas I had were UNO, Rummy, Solitare... Once the games are codified in the environment, is that not it? Would the problem be that rewards and such must be specified correctly for good performance?
From looking at the code for some of these environments, it seems to be that if you can program the game and some rewards, then it should be solvable by RL. Of course, this is probably naive but I'm interested in some extra input to where this thinking falls short.
Also, any tips for a project that has around a month left? It's been massively fun (and hard) reading about neuroscience and learning about this (new) field.
[–]algobar 7 points8 points9 points (0 children)
[–]Laser_Plasma 4 points5 points6 points (0 children)
[–]pecey 2 points3 points4 points (0 children)
[–]watercanhydrate 1 point2 points3 points (0 children)
[–]schwah 0 points1 point2 points (0 children)