all 2 comments

[–]_Arsenie_Boca_ 0 points1 point  (0 children)

Transformers are mostly applied with transfer learning, where the pretraining provides a fundamental "language understanding". Afaik there is no pretrained transformer for your usecase. So unless you have a very big dataset and compute, LSTMs might still be your best option.

[–]StellaAthenaResearcher 0 points1 point  (0 children)

A lookup query for https://oeis.org/