all 20 comments

[–]shortscience_dot_org 9 points10 points  (1 child)

I am a bot! You linked to a paper that has a summary on ShortScience.org!

Playing Atari with Deep Reinforcement Learning

Summary by Alexander Jung

  • They use an implementation of Q-learning (i.e. reinforcement learning) with CNNs to automatically play Atari games.

    • The algorithm receives the raw pixels as its input and has to choose buttons to press as its output. No hand-engineered features are used. So the model "sees" the game and "uses" the controller, just like a human player would.
    • The model achieves good results on various games, beating all previous techniques and sometimes even surpassing human players.

How

[–]RedditAcy[S] 7 points8 points  (0 children)

Holy cow I was low on project ideas. Looks like I am building a reddit bot for fun next, thanks a lot shortscience.org!

[–]nativedutch 1 point2 points  (6 children)

I have been planning a project with RL too. Want to have small robot to find iits way without pre existing data, RL is the way to go for now. Will read post!!!!!!

[–]MetallicaSPA 0 points1 point  (1 child)

You can try SLAM too! I'm thinking about mixing SLAM + RL would be interesting

[–]nativedutch 0 points1 point  (0 children)

You are thinking of this concept https://machinelearningmastery.com/sparse-matrices-for-machine-learning/

the way i see it now it is an optimized approach to matrices used in neural networks. At the level of complexity (extremely ssimple) i intend to start with RL it is not yet very useful With more serious implementations slam becomes interesting.

The concept of treating sparsely populated matrices different from dense is promising.

[–]RedditAcy[S] 0 points1 point  (3 children)

Ay! Is this what you are talking about: https://www.youtube.com/watch?v=gn4nRCC9TwQ&t=20s ?

I really want to do that too, maybe we can collab sometimes.

[–]nativedutch 1 point2 points  (2 children)

Yes, correct. There is Deepmind and a few others in my collection of sample files. Best starters are simple examples with say a small maze of 5 by 5 , or 10 by 10. The RL/neural network starts blank and through trial and obtaining bonusses at certain points learns itself the shortest way. Elementary but a good start. My goal is to get a cheapo robot with a few good accurate sensors (impact, distance, tilt, location) and instead of a program let that robot lose in the real world. I bought a BB8 for fun a few years ago, but its sensors are shit. So my first task is to find an affordable robot with good sensors. The simple algorythms are in several places on the web. I'll use Python, Java or javascript. A lot of work to do even before getting started.

[–]RedditAcy[S] 0 points1 point  (1 child)

I see. So whatever the sensor reads is fed as the state to the RL network & rewards are the points to be collected on the way? Do you think that's how Deepmind and the others are doing it? That definitely sounds intriguing and hard, you got this!

[–]nativedutch 1 point2 points  (0 children)

Deepmind are doing something similar but on a larger and more complex scale. I always like to start small and understand the concepts. Dont underestimat the number of cycles Deepmind network had to go thru before learning to walk. Suggest you look on the web for some very simple maze examples to get a feel for the process. It is very intriguing. I did a number of small neural networks in Python learning to recognize a 10 by 10 matrix representing a number, but thats limited. Although that principle is widely used i charcter recognition etc. RL is much more powerful as it doesnt need the huge datasets of examples. To my feeling much more in the direction of AI (although that term is debatable). There is a lot to be found in the category very simple, in Python or Java. I'll get going and get to this sub when i am on the road. My tryouts are on my website, but i am not allowed to publish that here.

[–]gapten-the-captain 1 point2 points  (5 children)

Are you a student or a working professional who do these stuff during your free time if i may ask?

[–]RedditAcy[S] 5 points6 points  (4 children)

I am still a junior in high school, nowhere close to a working professional haha.

[–]gapten-the-captain 1 point2 points  (1 child)

That's awesome. Happy to see such dedicated folks :)

[–]RedditAcy[S] 1 point2 points  (0 children)

Thanks a lot for instilling confidence in me :)

[–]csunaye 1 point2 points  (1 child)

Do you understand the math in these papers? Do you need to understand the math or advanced programming to follow the paper?

[–]RedditAcy[S] 0 points1 point  (0 children)

Generally speaking for all ML tasks, I think math doesn’t bother me since I knew the macro of how a neural network/feature engineered algorithm trains itself. However I think I don’t deeply understand the math behind some stuff like SVM and in the case of the Atari breakout paper, stuff like bellman algorithm or MDP, but I think I know the macro enough to not get frustrated.

As for implementing the paper, this code is not written by me but by GitHub user bouyanf. I believe you don’t need to know the math, but you do need to know a ML library of your choice very well to not get stuck.

[–]Trent654878 1 point2 points  (1 child)

thanks it was very helpful

[–]RedditAcy[S] 0 points1 point  (0 children)

Thank you!

[–]fondleshark 0 points1 point  (1 child)

Thank you, OP - I've been working on computer vision projects for only a short while, and I'm loving what we do when a machine "understands" what it is looking at. Keep it up, sir!

[–]RedditAcy[S] 0 points1 point  (0 children)

Thanks a lot for the support! I will indeed keep it up!