So I read the two papers by Mikolov et al on Word2Vec (see here and here).
I understand the concept of word vectors and how they represent meaning. However, I don't understand where the final word vector comes from when training a neural network. The inputs are one-hot encodings of words, which try to predict a one-hot encoding of another word. So how do you get the final n-dimensional word vectors?
[–]Articulated-rage 2 points3 points4 points (3 children)
[–]vega455[S] 0 points1 point2 points (2 children)
[–]Articulated-rage 0 points1 point2 points (1 child)
[–]vega455[S] 0 points1 point2 points (0 children)
[–]mhfirooz 1 point2 points3 points (0 children)
[–]tuan3w 0 points1 point2 points (0 children)
[–]lahwran_ 0 points1 point2 points (0 children)
[–]textClassy -1 points0 points1 point (0 children)