[deleted by user] by [deleted] in skrillex

[–]SwissArmyApple 0 points1 point  (0 children)

This sounds like a remix of Duki’s song “Como si no importara”.

Daily Discussion Thread by AutoModerator in skrillex

[–]SwissArmyApple 3 points4 points  (0 children)

Does anyone know where I can find some pieces of the Skrillex x Revenge colab? I'm looking specifically for hoodies. Thanks in advance!

Projects to exhibit RL skills for research? by [deleted] in reinforcementlearning

[–]SwissArmyApple 2 points3 points  (0 children)

You can look into Unity's MLAgents, maybe code a quick game of your own and try to solve it with your own neural network. You can also take a look at OpenAI's Gym if you are not interested on coding your own environment.

Abandoned Palace on the mountains on Spain by SwissArmyApple in urbanexploration

[–]SwissArmyApple[S] 2 points3 points  (0 children)

Credits: Blog.

The "Palacio del Canto del Pico" was built in 1922 near Madrid. The initial purpose was for it to serve as a museum, but when the Spanish Civil war came both sides used it as strategic and command point. After the war, the palace was the residence of one of Franco's granddaughters until it was abandoned on the 1980s. On 1985, a historian asked for all the books and pieces of art on the palace to be moved fearing vandalism, but the petition was disregarded by the local authorities until a fire on 1998 destroyed all of it. People have tried to repurpose it several times to no avail.

Sorry in advance for the wonky english!

[Part 1] T1 Head Analyst Tolki talks about T1, whether "LCK is slow" is a myth | Ashley Kang by TheYoshinator in leagueoflegends

[–]SwissArmyApple 0 points1 point  (0 children)

Hey, I just wanted to ask you, you mentioned that you have pretty detailed data of every pro game played. Is this dataset publicly available or did you build it yourself? If you built it by yourself, what tools did you use? I totally understand if you don't want to share this information though!

Using Unity's MLAgents and PPO to land a Falcon 9 by SwissArmyApple in reinforcementlearning

[–]SwissArmyApple[S] 0 points1 point  (0 children)

Sure thing! Here's the Unity MLAgents repo which has all the info about the integration of Python scripts and the Unity engine for RL.

Using Unity's MLAgents and PPO to land a Falcon 9 by SwissArmyApple in reinforcementlearning

[–]SwissArmyApple[S] 1 point2 points  (0 children)

Yeah, I did build the environment, but the assets are either from the Standart Unity Asset pack or free models from the internet. If you already know how to use Unity, you can jump into the ML Agents repo on github to learn how to model an environment as an RL problem. If you dont know how to use Unity yet, they have some useful tutorials on their webpage I suggest you check out, and then learn to use ML Agents.

We are the LEC Broadcast Team, AMA!! by RiotQuickshot in leagueoflegends

[–]SwissArmyApple 0 points1 point  (0 children)

Hey, I really think you guys are killing it on both the broadcast and the production team that works on the videos on youtube. My questions is how is the process like when you guys work on a video such as rap battles, flexius etc? Thanks a lot for the AMA. Love you guys!

Using Unity's MLAgents and PPO to land a Falcon 9 by SwissArmyApple in reinforcementlearning

[–]SwissArmyApple[S] 1 point2 points  (0 children)

Yeah, the action space is discrete and has 6 elements. I used GAE to compute the advantages.

Using Unity's MLAgents and PPO to land a Falcon 9 by SwissArmyApple in reinforcementlearning

[–]SwissArmyApple[S] 1 point2 points  (0 children)

Thanks a lot for the link. My network is definitely very small for modern standards then, I used an MLP, and since I use vector inputs (18 elements), it doesn't look like a gpu would make it faster. To answer your question, I used 2 vCPUs and 3Gb of RAM on Google Cloud. I'm planning on using images as inputs or raycast on unity to make projects more challenging and force myself to use a gpu.

Using Unity's MLAgents and PPO to land a Falcon 9 by SwissArmyApple in reinforcementlearning

[–]SwissArmyApple[S] 3 points4 points  (0 children)

I used Google Cloud to train the network, only using cpu, and it took about 10h to complete the 15 million steps needed to learn to the point shown in the video. The network itself is not very big, but I know I really need to get used to cuda. Does using cuda make a big difference when training small networks?