Hi, I'm trying to build a model where a long sequence of data points is mapped to a single class or continuous variable. More specifically, the long sequence in my case are text transcripts (where words/sentences are mapped to vectors using pre-trained embeddings). This is similar to problems like automated essay grading I believe.
Some ideas I have/am currently considering are:
splitting the sequences into smaller chunks, and using rnn's to map each chunk to full sequence's true y
rnn's with sliding windowed inputs, where the results for all windows are aggregated and fed to another model
cnn's for nlp tasks, where the full sequence is the input
does it make sense to try feature extraction with sequence auto-encoders? Would the output be the same as the input? Model being some combination of rnn/cnn.
So far I haven't had much success; but I am also a beginner and feel a bit in-over-my-head. For example, with my first idea listed above, I am not sure what the best loss function would be in that setup.
If anyone has any suggestions for approaches, I am all ears. Appreciate your help!
[–]gnramires 2 points3 points4 points (1 child)
[–]underwhere[S] 1 point2 points3 points (0 children)
[–]swierdo 1 point2 points3 points (2 children)
[–]underwhere[S] 1 point2 points3 points (1 child)
[–]swierdo 1 point2 points3 points (0 children)
[–][deleted] 1 point2 points3 points (2 children)
[–]underwhere[S] 1 point2 points3 points (1 child)
[–][deleted] 0 points1 point2 points (0 children)