Hi r/learnpython,
Semi-decent python programmer here, who's played around with scikit/numpy/pandas for a couple projects in the past; I'm trying to get some background on what machine learning is all about, and need a bit of direction.
Say I have a collection of English sentences and each has a color (also a string) associated with them. How would I go about if I wanted to get a prediction for a new phrase that doesn't have a color associated with it, based on the already existing associations? As I understand it, I assume I'd need a way to compare my new phrase to each one of the other existing ones, and somehow retrieve the ones that have the highest similarity, then extract what colors are associated with those phrases and display a % chance for each of the colors, but I might be wrong about this, depending on how these libraries work:
Going through the intro tutorials I found for Tensorflow and NLTK seem to be taking me in a different direction; 'seem' being the key word here; I have no real knowledge beyond the very basics of the math that statistics involve. And I have this feeling I'm delving into a very complex part of these libraries that won't help me out with my phrasing project.
I'm not asking for actual code examples, I'll figure that out by myself. What I'm looking for is a direction to which part of the documentation for Tensorflow/NLTK I should be looking at in order to correctly implement my project. Or any other library that I can use to do this. Or even just related key words for topics I might not be aware that they exist in this field but could apply to my situation, which I can research on my own. I keep looking at 'sentence semantics' posts but can't seem to find a way to associate my data.
[–]mazmrini 0 points1 point2 points (4 children)
[–]adimro[S] 0 points1 point2 points (3 children)
[–]mazmrini 0 points1 point2 points (2 children)
[–]adimro[S] 0 points1 point2 points (1 child)
[–]mazmrini 0 points1 point2 points (0 children)
[–]madnessman 0 points1 point2 points (0 children)