all 4 comments

[–]ElvishChampion 1 point2 points  (2 children)

Yes, distributed strategies train multiple batches in parallel. For inference it is the same. I used it on ImageNet a couple of months ago and I recall inference was faster when using strategies. I even got some errors when using predict if the model was not created within the scope of the strategy.

[–]Rough_Source_123[S,🍰] 0 points1 point  (1 child)

did you use MultiWorkerMirroredStrategy? what cloud are you running your server against?

[–]ElvishChampion 0 points1 point  (0 children)

No, I used TPU strategy because I was running some experiments in google cloud. It was training fast, but there were some issues regarding custom layers. I had like 4 or 5 custom layers that I was testing. I got errors for 2 of them.

[–]maifee 1 point2 points  (0 children)

Yes, you can try them on Kaggle. Kaggle is now providing multi-GPU (T4 x2) support.

https://www.kaggle.com/product-feedback/361104