You are viewing a plain text version of this content. The canonical link for it is here.
Posted to commits@mxnet.apache.org by GitBox <gi...@apache.org> on 2018/03/26 14:36:47 UTC

[GitHub] Jeffery4000 commented on issue #10247: Slow speed on distributed training

Jeffery4000 commented on issue #10247: Slow speed on distributed training 
URL: https://github.com/apache/incubator-mxnet/issues/10247#issuecomment-376188327
 
 
   To clarify, I'm able to achieve result below with CIFAR10 dataset with resnet-152 with the same configuration.
   
   Single node with 1 GPUs: 980 samples/sec
   Single node with 2 GPUs: 1910 samples/sec
   Single node with 3 GPUs: 2890 samples/sec
   Single node with 4 GPUs: 3850 samples/sec
   Two node with 1 GPUs each: 970 samples/sec on each machine = 1940 samples/sec
   Two node with 2 GPUs each: 1830 samples/sec on each machine = 3660 samples/sec
   Two node with 3 GPUs each: 2650 samples/sec on each machine = 5300 samples/sec
   Two node with 4 GPUs each: 3510 samples/sec on each machine = 7020 samples/sec

----------------------------------------------------------------
This is an automated message from the Apache Git Service.
To respond to the message, please log on GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
users@infra.apache.org


With regards,
Apache Git Services