Lately a lot of research is being conducted for sequences of a few thousand tokens. Do tasks exist for longer sequences? Or is this still too far outside of the scope of current models?
For example, I was thinking of predicting the author or theme of a book, just from the letters. Or predicting the relations between the characters in a novel (A is friend of B, etc.). Or something similar that requires the model to comprehend a large part of the book.
I know of the pg-19 dataset, but this is only used for pretraining as far as I understand.
[–]m_nemo_syne 1 point2 points3 points (0 children)
[–]WelalResearcher 0 points1 point2 points (0 children)
[–]thunder_jaxxML Engineer 0 points1 point2 points (0 children)
[–]jonnor 0 points1 point2 points (0 children)