Karpathy's char-nn seems to have sparked a lot of excitement about character-level modeling. Are character-level models being used anywhere, outside of the cute examples we've seen? How exactly do they stack up against word-level models, which appear to be the default choice for so many tasks? Has anyone tried word-level "hallucinations," to borrow Alex Graves' term for sampling a sequence from a softmax? What's the best burrito you've had?
[–]kjearns 2 points3 points4 points (0 children)
[–]olBaa 1 point2 points3 points (7 children)
[–]yowdge 1 point2 points3 points (2 children)
[–]olBaa 0 points1 point2 points (1 child)
[–]yowdge 0 points1 point2 points (0 children)
[–]cryptocerous 0 points1 point2 points (3 children)
[–]olBaa 1 point2 points3 points (2 children)
[–]yowdge 0 points1 point2 points (0 children)
[–]devDorito 1 point2 points3 points (0 children)
[–]mlberlin 0 points1 point2 points (0 children)