You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
{{ message }}
This repository has been archived by the owner on Sep 18, 2024. It is now read-only.
first I wanna say that it's so nice to see people sharing their thoughts and work like this.
I just wanted to ask, wrt. to the Keras distributed tests, are you scaling batch size with the number of gpus? (as Keras just splits the given batchsize, across the cards. so for a batchsize of 256 on 4 cards the real batchsize is 64 per card. (I honestly think this should be changed, but c'est la vie)
so this may be why you see less efficiency on the cards.
here's a plot from my tests, that shows quasilinear speedups on EC2 instances.
hope this helps!!
The text was updated successfully, but these errors were encountered:
Sign up for freeto subscribe to this conversation on GitHub.
Already have an account?
Sign in.
hey guys,
first I wanna say that it's so nice to see people sharing their thoughts and work like this.
I just wanted to ask, wrt. to the Keras distributed tests, are you scaling batch size with the number of gpus? (as Keras just splits the given batchsize, across the cards. so for a batchsize of 256 on 4 cards the real batchsize is 64 per card. (I honestly think this should be changed, but c'est la vie)
so this may be why you see less efficiency on the cards.
here's a plot from my tests, that shows quasilinear speedups on EC2 instances.
hope this helps!!
The text was updated successfully, but these errors were encountered: