Page 1 of 1

Training Speed on Multi-GPU

Posted: Mon Oct 12, 2020 10:44 pm
by dheinz70

Also I noticed this the other day.

Distributed with a batch of 14, and only gpu1 with a batch of 7.

Shouldn't the distributed batch of 14 have roughly 2x the EG/s of the single gpu with a batch of 7?

Screenshot from 2020-10-12 17-39-26.png
Screenshot from 2020-10-12 17-39-26.png (2.61 KiB) Viewed 2063 times

Re: Log and graph weirdness

Posted: Wed Oct 14, 2020 9:58 am
by torzdf

You get speed up by increasing your batchsize.

The same batchsize on a single GPU or multi GPUs is likely to run at about the same speed, or maybe slightly slower.

EDIT

Misread your message. I don't have a multi-gpu setup, so can't compare, and don't know where the "sweet-spot" is. Hopefully someone who does have a multi-gpu setup can offer some insight.