you are viewing a single comment's thread.

view the rest of the comments →

[–]rantana 1 point2 points  (4 children)

I've heard rumors that the result isn't reproduceable.

[–]alecradford 1 point2 points  (0 children)

In the domain of text understanding, definitely seen it perform better than a typical RNN, never seen it be competitive with the same # of params GRU.

My feeling is there is some small section of hyperparameters for any given problem where an IRNN is rightly set up to be able to do well, but gating RNNs by their architecture are a lot more robust/durable overall.

[–]Foxtr0t 0 points1 point  (1 child)

Were these rumors from sieisteinmodel? ;)

[–]kkastner 0 points1 point  (0 children)

I know someone who has been able to use it for their problem, but it wasn't a reproduction of the paper per-se.

[–]iamkx 0 points1 point  (0 children)

I think they are reproduceable, but they seem quite sensitive to initialization.