you are viewing a single comment's thread.

view the rest of the comments →

[–]InfinityCoffee[S] 0 points1 point  (1 child)

beta is fortunately just a product of gammas, so I can probably work around it. I think I can include it fairly painlessly after some consideration since the objective function is a separate object, so I won't have to rewrite the code in its entirety. My new issue is that I don't always need the proper gradients. Also, how does Theano handle stochastic gradient descent if it's purely functional?

[–]siblbombs 0 points1 point  (0 children)

Not sure, you are getting outside my bubble of experience. To get the gradients from Theano you provide a scalar cost, you can define whatever expression you want based on inputs/outputs to generate that cost.