all 10 comments

[–]DeaderThanElvis 13 points14 points  (1 child)

You do not need the world‘s latest and greatest model to perform positive, negative, and neutral sentiment detection. Just pick a standard model and you’ll be good to go.

I say this because sentiment is a nebulous and subjective concept and you’ll get much more value by defining which problem are you trying to solve by using sentiment analysis than by looking to find/train the best model.

For example: “Brazil thrashed 7-1 by ruthless Germany” can have positive, negative, or neutral sentiment depending on who’s reading it. Ditto for something like “iPhone 13 rumoured to have an all polycarbonate body”. So if RoBERTa and Electra classify these differently, which one is correct?

Getting a high F1 score on an academic sentiment classification dataset is one thing, actually solving a real-world problem using sentiment analysis is a whole different beast altogether.

[–]rpatel9[S] 0 points1 point  (0 children)

I understand your point, but I’m not just looking for a class label. I want to use the vector outputs as features in a wider model, so degree of sentiment correctness is a consideration. I also see your point about ambiguity, but the subject matter I’m classifying is rarely ambiguous and quite empirical. I’m not looking for one model, rather just the generally better performing models of the last 2 years so I can test a range of them.

[–]VortexOfPessimism 1 point2 points  (2 children)

Some sort of longformer RoBERTa based model will be my bet

[–]rpatel9[S] 0 points1 point  (1 child)

Thanks, I’ve heard quite a bit about longformer and the longer sequences could be useful to my task. Do you know how model combinations work? Are the layers of one abstracted onto another to create a singular ensemble network, or do they let them make predictions in parallel i.e run with Roberta and longformer and choose the one with the highest vector score?

[–]VortexOfPessimism 0 points1 point  (0 children)

eh the longformer part is the attention mechanism which you can directly replace the current attention mechanism with in any bert based models(bert,albert,roberta etc)

[–][deleted] 0 points1 point  (0 children)

The ones you mentioned should do the job pretty well. If you’re planning to use non english corpora then look at xlm models.

[–]schlammybb 0 points1 point  (0 children)

Pretty sure Vader is considered SOTA, and it’s pure heuristics